Abstract
The present study examined associations between presence/absence of a mask and facial emotion recognition, and how interoceptive awareness, i.e., the perception of internal bodily sensations, may influence associations between them. Eighty-two university students participated in an online behavioral experiment. Participants were required to evaluate categories of emotions as well as valence and arousal levels of facial stimuli that were either neutral or expressed one of Paul Ekman’s basic emotions, i.e., anger, disgust, fear, happiness, sadness, and surprise. Participants also completed a psychological scale on interoceptive awareness. Results showed that accuracy of categorization was significantly lower and levels of valence and arousal were significantly closer to neutral in masked than in unmasked faces for multiple emotions. In addition, individuals who showed higher, as compared to lower, emotional awareness reported significantly higher levels of valence for masked stimuli that expressed surprise. These results suggest that wearing a mask can impair accuracy of facial emotion recognition and sensitivity to emotions, whereas awareness of the association between interoception and emotion might mitigate impairments of sensitivity to emotions in masked faces.
1. Introduction
Since the outbreak of the COVID-19 pandemic in 2020, wearing face masks has become a key strategy for infection prevention. While masks offer benefits for physical health, they are also suggested to hinder accurate recognition of others’ emotional expressions. A growing body of evidence suggests that wearing a mask can impair one’s ability to categorize various emotional expressions correctly. In fact, multiple studies have reported that masked stimuli that expressed neutral emotion and five of the six basic emotions except for anger (i.e., disgust, fear, happiness, sadness, and surprise) were more frequently misidentified compared to unmasked stimuli (; ; ; ). Even before the pandemic, some studies examined the effect of limiting facial information by presenting only part of the face, such as the eye region (; ) and/or one half of the face (). For example, () found that expressions of disgust and happiness were less accurately recognized when only the eye region was visible compared to when the whole face was shown. In this context, studies have further investigated how accurately people can distinguish among different emotional expressions, especially when one expression is confused for another expression. Such studies reported that fear was frequently confused with surprise, and that disgust was evidently confused with anger when the lower half of the face was covered with a mask (; ).
Wearing a mask has also been shown to impair the ability to accurately infer emotional intensity, which corresponds to arousal levels (; ; ; ). () proposed the circular theory in which various types of emotions are mapped onto the coordinates on the axes of valence (i.e., pleasure–displeasure) and arousal (i.e., arousal–sleepiness; see also ; ). Based on the circular theory, () found that observers were prone to rate lower emotional intensity for masked facial stimuli expressing disgust, fear, happiness, sadness or surprise than for unmasked facial stimuli. Similarly, () reported that happy and sad faces were rated lower in intensity when those faces were masked than when they were unmasked. () also revealed that observers tended to report lower emotional intensity scores for positive emotions in both masked faces and the upper half of faces than in full faces. Furthermore, () showed that masked facial stimuli that expressed happiness, neutral expressions, anger, and sadness were rated as less arousing and closer to neutral than unmasked facial stimuli. () reported that masked faces expressing fear were less likely to be evaluated negatively, and masked faces expressing happiness were less likely to be evaluated positively, than unmasked facial stimuli. These findings suggest that wearing a mask can diminish emotional sensitivity, i.e., precision in the recognition of actual valence and arousal levels, to faces.
Given these impairments in accurate recognition of and sensitivity to facial expressions, it is essential to identify psychological and cognitive factors that may support accurate emotion recognition when external cues such as specific facial parts are restricted. Identifying such factors is crucial for facilitating smoother social communication under conditions such as widespread mask-wearing during a pandemic. One candidate is interoceptive awareness, the conscious perception of internal bodily sensations (). Predictive coding accounts provide a unifying framework that the brain uses to minimize prediction errors by integrating exteroceptive signals (e.g., facial expressions) with interoceptive signals (i.e., internal bodily states) to infer affective meaning (; ). Within this framework, heightened interoceptive awareness may contribute to generating more accurate representations of internal bodily states, thereby compensating for reduced exteroceptive input when facial cues are masked. Given that interoception and exteroception share cognitive resources (), sensitivity to internal signals may play a greater role in emotion recognition under such constraints than when exteroceptive information is fully available. At the psychological level, constructionist theories of emotion propose that emotions are constructed from core affect deriving from interoception (i.e., valence and arousal), and are shaped by knowledge and situational contexts (; ). From this perspective, interoceptive awareness may allow individuals to more accurately perceive the bodily feelings that underlie valence and arousal. Such sensitivity could, in turn, help them infer others’ emotions when facial cues are masked. Embodied emotion models further suggest that interoceptive brain regions such as the anterior insula and anterior cingulate cortex play a central role in monitoring bodily states and recognizing emotions in others (; ). Together, these perspectives converge on the idea that interoceptive awareness is a crucial factor for emotion recognition of others, particularly under conditions of reduced external emotional cues by a mask.
In line with this body of research highlighting the close relationship between interoception and emotion, several empirical studies have shown that interoceptive awareness is associated with the recognition of affective states of others (; ). () found that individuals with greater awareness of their own heartbeats, as indexed by higher accuracy in the heartbeat counting task, recognized anger, disgust, happiness, and sadness in facial expressions at lower emotional intensity levels than those with lower accuracy. Regarding interoceptive awareness assessed via self-report, () reported that individuals with higher interoceptive awareness in daily life showed lower thresholds for detecting emotions, which was reflected in shorter reaction times for emotional expressions at high intensity (see also ). () found that higher self-reported interoceptive awareness was significantly associated with more accurate categorization of negative emotions (i.e., fear, anger, disgust, and sadness). Importantly, () found that greater interoceptive awareness as reflected in heartbeat counting was associated with stronger facial mimicry when participants fixated on the eye region of facial stimuli. Although this study did not directly assess emotion recognition from eye region information alone, its findings suggest that interoceptive awareness may enhance sensitivity to emotional cues conveyed by the eyes. Considering these findings, it can be inferred that interoceptive awareness may facilitate the fine-grained recognition of emotional intensity and the accurate identification of discrete emotions when only the upper half of the face is visible.
Previous studies have suggested several factors that may influence emotion recognition in masked facial stimuli. () reported that contextual information (e.g., her cat was found or lost yesterday afternoon) mitigated difficulty in recognizing valence and arousal in masked facial stimuli. Several studies also showed that social sensitivity, which is defined as the ability to perceive others’ mental states, predicted accuracy in recognizing emotions from masked facial expressions (; ). Conversely, () reported that higher tendencies for alexithymia, a psychological trait characterized by difficulties in recognizing and describing one’s own emotions (), were associated with longer reaction times when categorizing emotions in facial stimuli with masked mouths or eyes. Despite these accumulating findings, the role of interoceptive awareness on recognition of and sensitivity to emotions in masked faces remains understudied.
Based on these underpinnings, the present study aimed to examine (1) associations between presence/absence of a mask and facial emotion recognition, and (2) potential effects of interoceptive awareness, on these associations. With respect to the first aim, we hypothesized that wearing a mask may hinder recognition of both categorization of emotions and levels of valence and arousal expressed in facial expressions. Specifically, we expected that accuracy in categorization of emotion would be lower for masked than for unmasked faces. Also, evaluation of valence and arousal levels were expected to be closer to neutral in masked than in unmasked faces. With respect to the second aim, We hypothesized that higher interoceptive awareness may be associated with more accurate and sensitive recognition of categories of emotion and arousal levels in masked faces. Specifically, we expected that, when presented with masked faces, individuals with higher self-reported interoceptive awareness in daily life, as measured by the Multidimensional Assessment of Interoceptive Awareness (MAIA; ), may categorize emotions more accurately and report higher levels of valence and arousal than those with lower interoceptive awareness.
2. Methods
2.1. Participants
Eighty-two undergraduate and graduate students living in Japan (37 males and 43 females; sex was unknown for 2 individuals; 18–25 years; mean age = 20.37; SD = 2.72) participated in this study. All participants were recruited through the university’s online platform soliciting for part-time jobs. To participate in the study, participants were asked to access the URL for the experimental platform that was shown within the job recruitment. Prior to participation, participants read the objective and content of the study which were displayed on the browser. Participants gave informed consent by clicking a checkbox on the platform. After completing the study, participants received a reward of JPY 500 through bank transfer. Data collection was conducted on 10 and 11 October 2022.
2.2. Stimuli
Facial stimuli used for the present study were selected from the National Institute of Advanced Industrial Science and Technology (AIST) facial expression database 2017 (). The database contains 12 types of emotional expressions from eight Japanese models (four men and four women), i.e., 10 types of expressions with closed mouths (i.e., anger, disgust, fear, happiness, sadness, surprise, sleepiness, excitement, relaxation, and neutral) and 2 types of expressions with open mouths (i.e., anger and disgust). The facial stimuli were provided free of charge, under the condition that they were used only for research purposes and were not disclosed to external parties. For the test sessions, we used a total of 56 stimuli with closed mouths, each of which expressed either neutral emotions or one of the six basic emotions designated by (), i.e., anger, disgust, fear, happiness, sadness, and surprise. For the practice session, we used eight stimuli with open mouths, all of which expressed anger, by selecting one stimulus from each model. All the selected stimuli were filmed from the front angle. Facial stimuli with a mask were virtually generated by superimposing an image of a nonwoven mask (Matsukiyo mimi ga itaku narinikui mask; Lib Laboratories Co., Ltd., Tokyo, Japan) over the lower half of the face by using Adobe Photoshop Elements 2019 (Adobe Inc., San Jose, CA, USA).
2.3. Psychological Scale
To assess sensibility of interoceptive awareness, we used a Japanese version of the Multidimensional Assessment of Interoceptive Awareness (MAIA) developed and validated by (). The MAIA, originally developed by (), is a widely used self-report scale that multidimensionally assesses awareness of interoceptive signals in daily life (). The Japanese version of the MAIA is composed of six subscales, each of which is designed to measure distinct facets of interoceptive awareness, i.e., Noticing (awareness of uncomfortable, comfortable, and neutral body sensations, 5 items: e.g., I notice how my body changes when I am angry.), Not-distracting (the tendency not to ignore or distract oneself from sensations of pain or discomfort, 3 items: e.g., I distract myself from sensations of discomfort.), Attention Regulation (the ability to sustain and control attention to bodily sensations, 7 items: e.g., I can refocus my attention from thinking to sensing my body), Emotional Awareness (awareness of a connection between body sensing and emotional states, 3 items: e.g., I notice that my breathing becomes free and easy when I feel comfortable.), Body-listening (the tendency to actively listen to the body for insight, 4 items: e.g., I listen to my body to inform me about what to do.), and Trusting (an experience of one’s body as safe and trustworthy, 3 items: e.g., I feel my body is a safe place.). Each item is rated on a six-point scale ranging from 0 (Never) to 5 (Always). Means of each subscale were used following the developers’ protocol, although hypotheses were formulated at the level of interoceptive awareness as a whole due to the lack of sufficient theoretical and empirical grounds and in accordance with conventions in previous studies (e.g., ; ). Internal consistency scores (Cronbach’s α) for each subscale in the present sample were 0.69, 0.78, 0.86, 0.86, 0.82, and 0.82, respectively.
2.4. Procedure
General Procedure. The present study was run entirely online using Gorilla Experiment Builder, a dedicated platform for online studies in psychology. Gorilla Experiment Builder is designed to minimize confounding factors stemming from variations in settings, equipment, and connection types. Its reliability as an experimental platform has been demonstrated by successful replications of experiments across diverse environments (). Participants were instructed to access the experimental platform from any location with internet access, by using either a laptop or a desktop computer equipped with a keyboard. They were not allowed to use a smartphone or a tablet. Participants were also instructed to use one of the following browsers: Firefox, Safari, Chrome, or Microsoft Edge, on their local computer. After the informed consent, the procedure for the experimental task was explained. After reading instructions of the procedure, participants underwent the practice session before proceeding to the test session. After the behavioral experiment, participants completed the MAIA scale by clicking checkboxes on the platform. Participation in the study took approximately 30 min in total.
Flow of each trial. At the beginning of each trial, a fixation cross appeared at the center of the screen for 1 s. Following an interstimulus interval of 500 ms, a facial stimulus with a mask (Figure 1a) or with no mask (Figure 1b) was displayed on the screen for 2 s. Then, the next display for selection of the category of emotions appeared (Figure 2a). Participants indicated categorization of the emotional expression in each facial stimulus by clicking a relevant button presented on the display. Selecting one of the six basic emotions (excluding neutral) resulted in the next display, in which participants were required to evaluate valence of the facial stimulus (Figure 2b). Participants reported the valence of each face on a 9-point Likert scale (1 = displeasure; 9 = pleasure) by clicking a relevant button with a number on it. Next, the display appeared (Figure 2c), in which participants were required to rate arousal of the facial stimulus on a 9-point Likert scale (1 = sleepiness; 9 = arousal). Selection of the neutral option immediately resulted in the next trial without requiring participants to evaluate valence or arousal levels. There was no upper limit in the response times for these ratings.
Figure 1.
Examples of facial stimuli (a) with and (b) without a mask. Note that images shown in this figure are the examples created by the authors and were not actually used in the experiment. The eyes in the example images are covered with a black bar to prevent personal identification; black bars were not present in the actual experimental stimuli.

Figure 2.
Examples of screens to require participants to evaluate (a) categories of emotions and levels of (b) valence and (c) arousal for each facial stimulus.
Practice and test sessions. The practice session involved 8 (models) × 1 (emotion: anger with open mouths) × 2 (with or without a mask) = 16 trials. In the practice session, we used facial stimuli expressing only a single emotion (anger), so that they were consistent with those used in the test session. To prevent an anchor effect, we used facial stimuli with open mouths in the practice session. Each of the 16 stimuli were presented once in a pseudo-randomized order. After the practice session was completed, participants took an approximately 5-min break before proceeding to the test session. The test session generally replicated the procedure applied in the practice session. Specifically, the test session consisted of 8 (models) × 7 (emotions) × 2 (with or without a mask) = 112 trials. Each facial stimulus was presented once in a pseudo-randomized order. An approximately 5-min break was inserted after the participant completed 56 trials, so that the session was separated into two blocks.
2.5. Statistical Analysis Methods
Statistical analyses were conducted using a download-free software HAD 18 () and R (version 4.4.2). Prior to the main analyses, Cronbach’s α was calculated using R to assess the internal consistency of trial-by-trial scores in categorization, valence, and arousal. Cronbach’s α indicated acceptable internal consistency for categorization (α = 0.84), valence (α = 0.93), and arousal (α = 0.98) scores, suggesting that our tasks reliably measured the intended constructs across stimuli. Subsequently, means and standard deviations of accuracy in categorization of emotions for the mask or unmasked facial stimuli, valence and arousal levels for them, and the subscale scores from the MAIA were calculated. For valence and arousal levels, values averaged across the eight models were used as representative values for each participant. For scores from the subscales of the MAIA, median values were also calculated to categorize participants into groups based on the degrees of interoceptive awareness. Participants whose scores were below the median value were classified into the lower group, and those whose scores were the median value or above were classified into the higher group. Next, for accuracy of categorization of emotions (% correct) and levels of valence and arousal, respectively, we conducted three-way analyses of variance (ANOVA) using HAD 18 with presence of mask (mask or no-mask) and emotion (accuracy of categorization: seven emotions; valence and arousal: six emotions) as within-participants factors and interoceptive awareness (higher or lower groups based on scores from each subscale of the MAIA) as a between-participants factor. When a three-way interaction (mask × emotion × interoceptive awareness) was statistically significant, tests of simple interaction effects (e.g., whether the mask effect differed across emotions within higher vs. lower interoceptive awareness groups) were conducted. If a simple interaction was significant, we then examined simple–simple main effects, such as comparing masked versus unmasked faces for a particular emotion within each interoceptive awareness group, followed by post-hoc multiple comparisons. When a two-way interaction between any two factors (e.g., mask × emotion, mask × interoceptive awareness) was significant in the three-way ANOVA, we examined the corresponding simple main effects (e.g., the effect of mask within each emotion), again followed by post-hoc multiple comparisons. For post-hoc comparisons, p-values were not corrected because both the mask and interoceptive awareness factors consisted of only two conditions. Finally, to further examine the pattern of misclassifications, we constructed confusion matrices separately for the mask and no-mask conditions. For each cell of the matrix (true emotion × response), we computed the proportion of trials across participants. To determine whether the observed proportions significantly deviated from chance level, we conducted one-sample t-tests against the theoretical chance level (1/7 = 14.29%) for each off-diagonal cell (false alarms), as well as for the diagonal cells (hits). Holm corrections were applied within each true-emotion row to control for multiple comparisons ().
3. Results
3.1. Descriptive Statistics
Means, SDs, and median scores for each subscale of the MAIA were as follows: 2.73 (SD = 0.89, median score = 2.60) for noticing, 2.55 (SD = 1.10, median score = 2.67) for not-distracting, 2.49 (SD = 0.91, median score = 2.43) for attention regulation, 2.53 (SD = 1.25, median score = 2.67) for emotional awareness, 2.15 (SD = 1.01, median score = 2.00) for body-listening, and 2.58 (SD = 1.11, median score = 2.67) for trusting. Table 1 displays the means and SDs for the accuracy of categorization, shown separately by mask condition (No-mask vs. Mask) and by groups split at the median on each MAIA subscale (Higher vs. Lower). Table 2 and Table 3 each show the rating scores for valence and arousal, presented for each mask condition (No-mask/Mask) and for each group based on MAIA subscale scores (Higher/Lower).
Table 1.
Mean accuracy (% correct) in categorization of emotional facial expressions.
Table 2.
Mean rated scores of valence for the emotional facial expressions.
Table 3.
Mean rated scores of arousal for the emotional facial expressions.
3.2. Results from Three-Way ANOVAs with Accuracy in Categorization of Emotions
- Three-way ANOVAs. Table 4 shows results from three-way ANOVAs with accuracy of categorization of emotions as a dependent variable. The three-way interaction (mask × emotion × interoceptive awareness) was statistically significant only for the analysis with the body-listening subscale from the MAIA as an independent variable. For the other MAIA subscales (i.e., noticing, not-distracting, attention regulation, emotional awareness and trusting), no significant three-way interaction was observed; however, two-way interactions between the presence of a mask and emotion were statistically significant. In addition, a two-way interaction between not-distracting and presence of mask was also significant. Regarding the main effects, the presence of a mask and emotion were significant, whereas interoceptive awareness was not significant across all MAIA subscales.
Table 4. Results of three-way ANOVAs for accuracy of categorization of emotions. - Post-hoc simple interaction analyses following a three-way interaction. Because the three-way interaction involving body-listening was significant, post-hoc simple interaction analyses were conducted. These analyses revealed that simple interactions between presence of mask and body-listening were significant for the anger and disgust stimuli (anger: F [1, 560] = 8.385, p = 0.004, ηp2 = 0.095; disgust: F [1, 560] = 4.637, p = 0.032, ηp2 = 0.055). For fear, sadness, and surprise, only the main effect of the mask was significant (fear: F [1, 560] = 27.239, p < 0.001, ηp2 = 0.254; sadness: F [1, 560] = 55.904, p < 0.001, ηp2 = 0.411; surprise: F [1, 80] = 69.153, p < 0.001, ηp2 = 0.464). By contrast, statistically significant main effects or interactions were not observed for the happiness and neutral conditions.
For the mask condition, the simple interaction between body-listening and emotion (F [6, 960] = 2.963, p = 0.014, ηp2 = 0.036) and the main effect of emotion (F [6, 960] = 172.853, p < 0.001, ηp2 = 0.684) were significant. For the no-mask condition, only the main effect of emotion was significant (F [6, 960] = 138.734, p < 0.001, ηp2 = 0.634). Furthermore, simple interactions between mask and emotion were significant for both the lower body-listening group (F [6, 480] = 8.378, p < 0.001, ηp2 = 0.098) and the higher body-listening group (F [6, 480] = 9.557, p < 0.001, ηp2 = 0.088).
- Post-hoc multiple comparisons following simple interactions. Given these significant simple interactions, post-hoc multiple comparisons were performed. Multiple comparisons further revealed that the accuracy for masked anger and sadness stimuli was significantly lower in individuals with higher body-listening than in those with lower body-listening (all ps < 0.05; Figure 3). Overall, these results show that individuals with a higher tendency to attend to interoception for guidance were less accurate in categorizing masked faces expressing negative emotions than those with a lower tendency to attend to interoception for insight.
Figure 3. Bar plots for the mean accuracy of categorization of emotions, with groups divided based on the higher/lower scores from the body-listening subscale from the MAIA. Error bars indicate standard errors (SEs) of the mean. * p < 0.05, ** p < 0.01, *** p < 0.001.
Moreover, multiple comparisons revealed that accuracy of categorization of emotions was significantly lower in the mask condition than in the no-mask condition for disgust, fear, sadness, and surprise stimuli in the group with lower body-listening and for anger, disgust, fear, sadness, or surprise stimuli in the group with higher body-listening. By contrast, accuracy of categorization was significantly higher in the mask than in the no-mask condition for neutral stimuli in the group with higher body-listening (all ps < 0.001; Figure 3). Taken together, these results show that individuals with both lower and higher tendency to attend to interoception for insight exhibited lower accuracy in categorizing facial stimuli expressing disgust, fear, sadness, and surprise in the mask than in the no-mask condition. Also, individuals with higher tendency to attend to interoception for insight were more accurate in categorizing masked neutral expressions than those with lower tendency to attend to interoception for insight.
- Post-hoc multiple comparisons following two-way interactions. For the analyses with the noticing, not-distracting, attention regulation, emotional awareness, and trusting subscales from the MAIA as independent variables, multiple comparisons revealed that accuracy of categorization for stimuli expressing disgust, fear, sadness, and surprise was significantly lower in the mask than in the no-mask condition for noticing or attention regulation (all ps < 0.001). Similarly, for the analyses with not-distracting, emotional awareness, or trusting, accuracy of categorization for stimuli expressing disgust, fear, sadness, and surprise was significantly lower in the mask than in the no-mask condition (all ps < 0.05). In these analyses, however, accuracy of categorization for neutral stimuli was significantly higher in the mask than in the no-mask condition (all ps < 0.05). Collectively, these results show that accuracy of categorization for disgust, fear, sadness, and surprise stimuli was lower in the no-mask than in the mask condition. In contrast, accuracy of categorization for neutral stimuli was higher in the no-mask than in the mask condition.
Finally, multiple comparisons showed that for both the low and high not-distracting groups, accuracy of categorization was significantly lower in the mask than in the no-mask condition (all ps < 0.001). No statistically significant differences in the accuracy of categorization were observed between the high and low not-distracting groups in either the mask or no-mask conditions. These results show that accuracy of categorization was consistently lower in the mask condition than in the no-mask condition, independent of levels in not-distracting.
3.3. Results from Three-Way ANOVAs with Valence
- Three-way ANOVAs. Results from three-way ANOVAs with valence as a dependent variable are summarized in Table 5. A statistically significant three-way interaction was observed only for the analysis with the emotional awareness subscale from the MAIA as an independent variable. For the analyses involving the noticing, not-distracting, attention regulation, body-listening or trusting subscales from the MAIA as independent variables, two-way interactions between presence of mask and emotion were significant. The main effects of mask and emotion were statistically significant, whereas the main effects of interoceptive awareness were not significant for the all MAIA subscales.
Table 5. Results of three-way ANOVAs for the rating scores of valence. - Post-hoc simple interaction analyses following a three-way interaction. Because the three-way interactions involving emotional awareness were significant, post-hoc interaction analyses were conducted. These analyses revealed significant simple interactions between presence of mask and emotional awareness for the happiness (F [1, 480] = 4.270, p = 0.039, ηp2 = 0.051) and surprise stimuli (F [1, 480] = 10.493, p = 0.0012, ηp2 = 0.116). For the disgust and fear stimuli, the main effect of presence of mask was significant in the simple interaction test that examined an interaction between presence of mask and emotional awareness (disgust: F [1, 480] = 54.215, p < 0.001, ηp2 = 0.404; fear: F [1, 480] = 62.670, p < 0.001, ηp2 = 0.439).
Simple interactions between emotional awareness and emotion were not significant for either the no-mask (F [5, 800] = 1.516, p = 0.220, ηp2 = 0.019) or mask (F [5, 800] = 0.755, p = 0.479, ηp2 = 0.009) conditions; however, the main effect of emotion was significant (no-mask: F [5, 800] = 478.700, p < 0.001, ηp2 = 0.857; mask: F [5, 800] = 347.672, p < 0.001, ηp2 = 0.813).
Furthermore, simple interactions between the presence of mask and emotion were significant for both the lower emotional awareness group (F [5, 400] = 18.790, p < 0.001, ηp2 = 0.155) and the higher emotional awareness group (F [5, 400] = 15.028, p < 0.001, ηp2 = 0.165).
- Post-hoc multiple comparisons following simple interactions. Multiple comparisons further revealed that scores for valence were significantly lower in the mask than in the no-mask condition for disgust and fear stimuli in the groups with higher and lower emotional awareness (all ps < 0.001). In addition, scores for valence were significantly lower in the mask than in the no-mask condition for the happiness or surprise stimuli in the lower emotional awareness group, and for the happiness stimuli in the higher emotional awareness group (all ps < 0.001; Figure 4). Taken together, these results show that individuals with both low and high awareness of the link between interoception and emotion reported scores on valence that were closer to neutral for disgust, fear, and happiness stimuli in the mask than in the no-mask condition.
Figure 4. Bar plots for the mean rating scores of valence, with groups divided based on the higher/lower scores from the emotional awareness subscale from the MAIA. Error bars indicate standard errors (SEs) of the mean. ** p < 0.01, *** p < 0.001.
Moreover, multiple comparisons further revealed that, for facial stimuli with no mask that expressed happiness and for stimuli with a mask that expressed surprise, valence was significantly higher in the higher emotional awareness group than in the lower emotional awareness group (all ps < 0.01; Figure 4). These results show that, compared with individuals low in awareness of the link between interoception and emotion, those high in emotional awareness reported higher scores on valence on masked stimuli expressing happiness and unmasked stimuli expressing surprise.
- Post-hoc multiple comparisons following two-way interactions. As two-way interactions between mask and emotion were significant for the analyses with noticing, not-distracting, attention regulation, body-listening or trusting, we subsequently performed post-hoc multiple comparisons. Multiple comparisons following two-way interactions revealed that scores on valence for disgust and fear stimuli were significantly higher and scores on valence for happiness and surprise stimuli were significantly lower in the mask than in the no-mask condition (all ps < 0.05). These results show that scores on valence in masked stimuli expressing disgust, fear, sadness, and surprise were closer to neutral compared with those in unmasked stimuli.
3.4. Results from Three-Way ANOVAs with Arousal
- Three-way ANOVAs. Table 6 summarizes results from three-way ANOVAs with arousal as a dependent variable. A three-way interaction between presence of mask, emotion, and interoceptive awareness was statistically significant only for the analysis with the emotional awareness subscale from the MAIA as an independent variable. For the analyses with the noticing, not-distracting, attention regulation, body-listening and trusting subscales from the MAIA as independent variables, two-way interactions between presence of mask and emotion were significant. Main effects of mask and emotion were statistically significant. The main effect of interoceptive awareness was also significant for the analysis with attention regulation, showing that arousal scores were higher in the higher attention regulation group compared with the lower group.
Table 6. Results from three-way ANOVAs for the rating scores of arousal. - Post-hoc simple interaction analyses following a three-way interaction. Because the three-way interaction involving emotional awareness was significant, post-hoc simple interaction analyses were conducted. These analyses revealed a significant simple interaction between presence of mask and emotional awareness for the surprise stimuli (F [1, 480] = 7.435, p = 0.007, ηp2 = 0.085). Regarding the happiness and sadness stimuli, simple interaction test that involved an interaction between presence of mask and emotional awareness revealed a significant main effect of presence of mask (happiness: F [1, 480] = 15.267, p < 0.001, ηp2 = 0.160; sadness: F [1, 480] = 4.937, p = 0.027, ηp2 = 0.058), although no other significant main effects nor interactions were observed.
In contrast, simple interaction between emotional awareness and emotion was significant for neither the no-mask (F [5, 800] = 0.207, p = 0.915, ηp2 = 0.003) nor the mask conditions (F [5, 800] = 1.340, p = 0.258, ηp2 = 0.016).
Furthermore, simple interaction between presence of mask and emotion was significant in the lower emotional awareness group (F [5, 400] = 4.755, p = 0.004, ηp2 = 0.044), but not in the higher emotional awareness group (F [5, 400] = 2.673, p = 0.052, ηp2 = 0.034). In the higher emotional awareness group, only the main effect of emotion was significant (F [5, 400] = 26.512, p < 0.001, ηp2 = 0.438).
- Post-hoc multiple comparisons following simple interactions. Given these significant simple interactions, post-hoc multiple comparisons were performed. These comparisons revealed that scores on arousal were significantly higher for the sadness stimuli in the mask than in the no-mask condition in individuals with lower emotional awareness. Conversely, scores on arousal were significantly lower in the mask than in the no-mask condition for happiness stimuli in the lower emotional awareness group and for surprise stimuli in the higher emotional awareness group (ps < 0.001). These data show that scores on arousal for masked facial stimuli were lower than those on unmasked stimuli, regardless of perceived connection between interoception and emotion.
Multiple comparisons revealed no statistically significant differences in the scores on arousal for masked and unmasked stimuli between the higher and lower emotional awareness groups (all ps > 0.05).
- Post-hoc multiple comparisons following two-way interactions. As significant two-way interactions between mask and emotion were found for the analyses with the noticing, not-distracting, attention regulation, body-listening and trusting, we further examined them using post-hoc multiple comparisons. These analyses revealed that arousal was significantly lower for happiness and surprise stimuli, and significantly higher for sadness stimuli, in the mask than in the no-mask condition (all ps < 0.05). These data show that participants reported masked stimuli expressing sadness as being more aroused and masked stimuli expressing happiness or surprise as being less aroused than unmasked stimuli expressing these emotions.
3.5. Error Patterns in Categorization of Emotions
Figure 5 illustrates the misclassification matrices for (a) masked and (b) unmasked facial expressions, with correct responses (diagonal cells) omitted. Cells in red indicate response proportions significantly greater than chance level, whereas blue cells indicate those significantly lower than chance level, after Holm correction for multiple comparisons.

Figure 5.
Misclassification matrices for (a) masked facial expressions and for (b) unmasked facial expressions. Correct responses (diagonal cells) are omitted. Cell colors indicate deviation from chance level (1/7 = 14.29%): red = significantly above chance level, blue = significantly below chance level, and gray = not significant after Holm correction for multiple comparisons. Numbers denote mean response proportions (%). The color scale is fixed to ±30 percentage points (pps) for comparability across panels.
For masked faces, fear stimuli were frequently misidentified as surprise (t(79) = 12.45, adjp < 0.001, d = 1.39), and surprise stimuli were often misidentified as fear (t(79) = 10.83, adjp < 0.001, d = 1.21). In addition, sadness stimuli tended to be categorized as disgust more often than expected (t(79) = 7.02, adjp < 0.001, d = 0.79). Conversely, some confusions (e.g., happiness with sadness) occurred significantly less frequently than chance level (t(79) = −6.88, adjp < 0.001, d = −0.77).
For unmasked faces, overall categorization was closer to chance expectations, and misclassification patterns were less systematic. Nevertheless, a consistent tendency remained for fear stimuli to be misidentified as surprise and vice versa, although the magnitude of these effects was smaller than in the masked condition.
These results show that the presence of a mask amplifies systematic misclassification patterns, particularly among negatively valenced emotions such as fear, surprise, and sadness.
4. Discussion
The present study examined accuracy in categorization of emotions and sensitivity of emotion recognition in masked faces, as well as potential impacts of interoceptive awareness on emotion recognition in masked facial stimuli. Overall, the results suggest that emotions expressed in the facial stimuli with masks were less accurately identified, and that they were perceived as closer to neutral for both valence and arousal dimensions compared with emotions expressed in faces with no masks. Interestingly, interoceptive awareness was found to have differential effects on emotion recognition, depending on the dependent variables. Specifically, individuals with a higher emotional awareness dimension of interoceptive awareness perceived masked faces that expressed surprise as higher in valence than those with lower emotional awareness did. On the other hand, individuals with lower body-listening dimension of interoceptive awareness were more accurate in categorizing masked faces expressing disgust and fear than those with higher body-listening.
4.1. Effects of a Mask on Recognition of and Sensitivity to Emotions
Consistent with preceding studies on mask-wearing (; ; ; ) and broader research on partial-face recognition (; ; ; ), results from the present study support the notion that wearing facial masks can impair the accurate recognition of multiple emotions, including anger, sadness, fear, and surprise. Exploratory confusion-matrix analyses further revealed that errors were systematic rather than random: fear was frequently misclassified as surprise, and disgust as anger. Such misidentification patterns converge with findings by (), who showed that partial occlusion of the face induces non-random confusions between expressions with overlapping visual features. From the perspective of the Facial Action Coding System (; ), these confusions are consistent with overlapping Action Units (AUs) in the upper face, such as the eyebrows and eyelids. Specifically, both anger and fear involve raising the upper lids (AU 5) and lowering the eyebrows (AU 4) components. Fear and surprise share raising the inner eyebrows (AU 1), raising the outer eyebrows (AU 2), and raising the upper lids (AU 5). Consistently, both sadness and fear involve raising the inner eyebrow (AU 1) and lowering the eyebrows (AU 4). The present study required participants to judge categories of emotions based on these emotional cues around the eyes, because cues in other facial parts were occluded by masks. Therefore, wearing a mask may have hindered the ability to distinguish these emotional expressions (). In contrast, no statistically significant difference was found in the accuracy to identify happiness between the mask and no-mask conditions. This may be because happiness uniquely involves the Duchenne smile, which refers to a genuine smile characterized by wrinkles around the eyes arising from raising the cheeks (AU 6), which are not occluded by a mask (; ). The present study also found that neutral stimuli were categorized more accurately in the mask than in the no-mask condition. This could be attributed to the fact that neutral faces with no masks appear to give slightly negative impressions (). Thus, partial occlusion of emotional cues induced by wearing a mask may have resulted in the neutral facial expressions being interpreted as more neutral in the present study.
The presence of a mask was found to make participants’ evaluation of valence for disgust, fear, surprise, and happiness closer to neutral. These findings suggest that a facial mask may introduce ambiguity in the perception of valence. Previous studies on emotion recognition of masked faces primarily focused on whether masked facial stimuli are judged to convey pleasant or unpleasant emotional valence (e.g., ). The present study extends this line of research by providing novel evidence of reduced perceived intensity of valence across multiple emotions. Consistent with previous studies (; ; ; ), we also found that the presence of a facial mask made participants’ evaluation of arousal for faces expressing sadness, happiness, and surprise closer to neutral. Evaluations of valence and arousal are thought to rely on salient intensity-related cues specific to each emotion: the smiling mouth conveys positive valence in happiness, the distorted mouth conveys negative valence in disgust, and the downturned mouth corners provide a low-arousal cue in sadness (, ; ; ; ). For fear, wide-open eyes provide a strong cue, but negative valence evaluations typically rely on the combined saliency of both the eyes and the open mouth (; ). Thus, when the mouth is occluded by a mask, the remaining eye information may be insufficient to support a clear negative valence judgment. In addition, wide-open eyes and an open mouth contribute both to valence and arousal judgments in surprise. Mask occlusion removes many of these diagnostic cues from the lower part of the face, thereby eliminating crucial sources of information for both valence and arousal. This likely explains why participants’ evaluations of valence and arousal were shifted toward neutral in the present study.
4.2. Effects of Interoceptive Awareness
Individuals with higher scores on body-listening, which is characterized by sustained attention to bodily sensations, exhibited lower accuracy in categorizing masked stimuli compared to those with lower scores on body-listening. Although our hypothesis concerned interoceptive awareness as a whole, this subscale-level effect emerged in the analyses and therefore should be interpreted with caution. A potential reason for these results concerns the fact that the present study design assumed multisensory processing of visual and interoceptive signals. () provided insights into allocation of attentional resources in multisensory perception, using an auditory oddball task synchronized with heartbeat by using heartbeat-evoked potentials (HEP) and event-related potentials (ERP). Amplitudes of both HEP (reflecting interoception) and ERP (reflecting auditory processing) were largest when pure tones were presented during diastole. Tanaka et al. suggest that interoceptive and auditory stimuli share common attentional resources when they were processed simultaneously. Assuming that these findings may extend to simultaneous processing of interoceptive and visual stimuli, individuals high in body-listening may have directed a relatively large amount of attentional resources to their own interoceptive signals when presented with the facial stimuli. Consequently, individuals high in body-listening may have allocated relatively fewer attentional resources to facial expressions, thus reducing their ability to recognize visual cues required to accurately categorize emotions.
Regarding evaluation of valence, individuals high in emotional awareness, i.e., awareness of the connection between bodily sensations and emotional states, perceived masked faces expressing surprise as more pleasant than those low in emotional awareness. This result partially supports the view that interoceptive awareness may help compensate for the evaluation of valence when facial emotional cues are restricted by a mask. From a predictive coding perspective, interoceptive awareness may refine internal bodily representations and thereby strengthen valence judgments when exteroceptive information is reduced (; ). Constructionist and embodied emotion models similarly emphasize the role of interoceptive processes in shaping core affective dimensions and supporting recognition of others’ emotions (; ; ; ). Importantly, this finding provides novel evidence that a specific aspect of interoceptive awareness can enhance the evaluation of valence in masked facial expressions. Although this effect was limited to one subscale and one emotional category, it suggests that interoceptive awareness may serve as a psychological resource supporting emotional sensitivity when facial cues are partially obscured. More broadly, this finding contributes to understanding how interoceptive awareness may facilitate social cognition under constrained communicative contexts, such as mask-wearing during pandemics or in clinical settings.
On the other hand, interoceptive awareness did not significantly influence participants’ evaluation of arousal for masked faces. This might be due to the characteristics of the psychological scale used in the present study. Other than the MAIA, several psychological scales exist for measuring interoceptive awareness. For example, both the Autonomic Perception Questionnaire () and the Body Perception Questionnaire () measure frequency of experience of specific sensations that arise from autonomic nervous activities and/or internal organs. Considering the fact that both interoception and emotion are suggested to arise from perception of specific autonomic and somatic activities (; ), subjective experience of specific interoceptive sensations may be more closely related to perception of arousal in facial expressions than to interoceptive awareness in daily life. Future research may examine whether and how subjective evaluation of interoceptive awareness might influence perception of arousal in masked facial expressions.
Taken together, the present findings provide preliminary but meaningful evidence that aspects of interoceptive awareness can differentially shape emotion recognition when facial cues are restricted by a mask. Theoretically, these results extend predictive coding, constructionist, and embodied accounts by showing that interoceptive awareness not only refines internal affective representations but also interacts with attentional allocation to influence sensitivity to others’ emotions. Practically, the present study highlights interoceptive awareness as a potential psychological resource that could be targeted to support social communication under constrained conditions, such as widespread mask-wearing during pandemics or in clinical populations where facial expressivity is reduced. Although the observed effects were limited in scope and require replication, the present study contributes to bridging the gap between interoception research and applied challenges in everyday social interaction. In particular, interoceptive interventions that enhance awareness of internal bodily signals (e.g., meditation and yoga; ) may offer promising avenues for fostering sensitivity to others’ emotions under conditions where facial cues are constrained, such as during mask-wearing.
4.3. Limitations and Future Perspectives
Despite these novel findings and perspectives, the present study has apparent limitations. Firstly, control over the experimental environment was inevitably limited because the present study was entirely conducted online. Although the Gorilla platform has been validated to provide reliable stimulus timing across browsers and operating systems (; ), online experiments are inevitably accompanied by methodological constraints. A major issue is the lack of experimental control (). For example, it is impossible to regulate participants’ viewing distance, posture, location, or ambient environment, making precise control of stimulus perception difficult. Moreover, participants’ concentration and motivation cannot be directly monitored, and they may be distracted by other sources (e.g., mobile phones, conversations, background noise), which introduces uncontrolled variance in performance (; ). Although these uncontrolled factors are unlikely to have systematically distorted the observed effects, they may have added random noise to the data and thus reduced the statistical sensitivity of the present study. Future research conducted under laboratory settings, where the testing environment can be rigidly controlled, would help to confirm the robustness of the present findings.
The second limitation relates to the relatively low ecological validity of the facial stimuli. The present study used images with digitally overlaid masks to maintain strict control of facial features. Although this method ensured consistency across conditions, it may have reduced ecological validity. Future research should therefore employ photographs of individuals actually wearing masks to enhance the generalizability of the findings to real-world social interactions.
The third limitation concerns potential individual difference variables that were not directly measured in this study. Previous studies have shown that alexithymia is moderately associated with difficulties in emotion recognition (e.g., ; ) and is closely related to low interoceptive awareness (e.g., ; ). Therefore, it remains possible that the effects observed here partly reflect the influence of alexithymia rather than interoceptive awareness alone. Future research should measure both constructs simultaneously to disentangle their unique and overlapping contributions to emotion recognition. In addition, familiarity with masked faces represents a distinct individual difference factor. Previous studies have suggested that prior experience with masked faces may shape performance in emotion recognition (; ). In Japan during 2022, the use of masks in daily life was widespread, which likely minimized individual variability in familiarity with masked faces within our sample. Nevertheless, in other cultural or temporal contexts, such familiarity may emerge as an important factor influencing performance in emotion categorization and ratings of valence and arousal. Future research should therefore control for individual differences in familiarity with masked faces to avoid confounding effects on these outcome measures.
The fourth limitation pertains to the methodology of measuring interoceptive awareness. In the present study, interoceptive awareness was assessed using the MAIA, which primarily measures tendencies in daily contexts. However, such daily life tendencies do not necessarily reflect awareness triggered by salient bodily events. In our paradigm, facial stimuli elicited transient and salient external inputs, suggesting a potential mismatch between the context of interoceptive awareness and the demands of the task. Recent studies suggest that interoceptive awareness assessed in relation to salient events, such as through momentary experience sampling or experimental manipulations of bodily states, reflects aspects of awareness more directly tied to affective processing (; ). In light of these considerations, the impact of emotional sensitivity on masked facial expressions might have been more apparent if interoceptive awareness was measured by questionnaires asking frequency of experiencing specific sensations rather than by assessing interoceptive awareness in daily life (; ). Furthermore, interoceptive awareness can also be assessed using objective and experimental indices targeting specific sensory modalities, beyond the self-reported tendencies measured by questionnaires. For example, cardiac interoceptive awareness has often been measured by the heartbeat detection tasks, and gastric interoceptive awareness has been measured by water-drinking or rectal distension (; ). Future research may therefore benefit from incorporating multiple methods to clarify which operationalization may predominantly enhance accuracy in emotion recognition of and/or emotional sensitivity to masked faces.
The fifth limitation is that the present study did not examine unconscious levels of emotional processing. Our methods focused exclusively on explicit behavioral measures, namely emotion categorization and ratings of valence and arousal, which do not reveal the processing stage at which interoceptive awareness exerts its influence. For example, it remains unclear whether higher scores on MAIA subscales such as emotional awareness and body-listening reflect stronger sensitivity to salient bodily signals themselves or differences in higher-level cognitive strategies for evaluating sensations and emotions. To address this issue, future research could combine behavioral tasks with physiological recordings of ongoing interoceptive signals during stimulus presentation, using techniques with high temporal resolution such as electroencephalography (EEG), electrocardiography (ECG), or electrogastrography (EGG). Previous studies have shown that heartbeat-evoked potentials modulate neural responses to emotional stimuli (e.g., ; ), suggesting that such approaches may clarify the stage at which interoceptive and emotional processes interact, beyond what can be inferred from behavioral outcomes alone.
The sixth limitation concerns generalizability of the present findings to different populations. The present sample consisted exclusively of young adults from Japan, which constrains the extent to which the results can be generalized. Previous research suggests that impairments in recognizing masked facial emotions vary with age; for example, older adults exhibit lower accuracy than younger adults for anger, sadness, and fear, suggesting differential reliance on facial regions for emotion recognition (). With respect to interoception, developmental studies show that interoceptive awareness improves during childhood and approaches adult-like levels by late childhood or early adolescence (; ). It appears relatively stable across early and middle adulthood, although evidence summarized by () indicates greater variability and potential decline in older age. Regarding cultural variation, studies of gaze allocation have demonstrated systematic differences in the use of facial cues during emotion recognition (e.g., Americans tend to rely more on the mouth region, whereas Japanese observers rely more on the eye region; ; ; ). In addition, cross-cultural studies report differences in interoceptive awareness as measured by both self-report and experimental tasks across diverse cultural groups, including Western, East Asian, and African samples (; ; ). Taken together, these findings suggest that the association between interoceptive awareness and performance on emotion recognition of masked faces in the present study may not generalize to populations differing in age or cultural background. Future studies should therefore extend investigations to participants across a broader age range and from diverse cultural contexts.
5. Conclusions
Results from the present study suggests that wearing a mask can reduce both accuracy and sensitivity of facial emotion recognition across multiple categories of emotions in the studied sample. Furthermore, results suggest that individuals with higher emotional awareness may perceive masked faces that expressed surprise as more positive than those with lower emotional awareness. These findings indicate that interoceptive awareness may play a compensatory role in emotion recognition when exteroceptive cues are limited, highlighting its potential importance for maintaining social communication under restricted facial visibility. Future research should replicate these findings under more controlled laboratory conditions with naturalistic stimuli, examine potential confounding variables, refine/reconsider approaches to measuring interoceptive awareness, and explore the generalizability of the present findings to broader populations.
Author Contributions
Conceptualization, K.Y. and H.M.; methodology, K.Y. and H.M.; investigation, K.Y. and H.M.; formal analysis, K.Y.; data curation, K.Y.; writing―original draft, K.Y.; writing―review & editing, H.M.; supervision, H.M. funding acquisition, H.M. All authors have read and agreed to the published version of the manuscript.
Funding
This research was supported by Waseda University Individual Research Allowance to the second author, Waseda University Grant for Special Research Projects (Project number: 2025C-329), JSPS KAKENHI Grant Number JP24K14588, and a personal crowd-funding project conducted in 2022 by the second author. The funders had no role in the study design, data collection, analysis, and interpretation, the decision to publish, or in the writing of the manuscript.
Institutional Review Board Statement
The study was conducted in accordance with the guidelines of the Declaration of Helsinki, and approved by the Ethics Review Committee on Research with Human Subjects of Waseda University (Approved No. 2022-257; Date of approval: 15 September 2022).
Informed Consent Statement
Informed consent was obtained from all subjects involved in the study.
Data Availability Statement
The datasets used and analyzed during the study are available from the corresponding author upon reasonable request.
Acknowledgments
The authors thank Haruyuki Ishikawa, Faculty of Culture, Education and Communication in Hyogo University of Teacher Education, and Honoka Ito, Shiori Nakazawa, and Kazutoshi Fukui, School of Culture, Media and Society in Waseda University, for their cooperation in drafting the previous version of the manuscript. The authors utilized ChatGPT-4o to modify grammatical and lexical errors in English when editing this manuscript.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Anwyl-Irvine, A. L., Massonnié, J., Flitton, A., Kirkham, N., & Evershed, J. K. (2020). Gorilla in our midst: An online behavioral experiment builder. Behavior Research Methods, 52, 388–407. [Google Scholar] [CrossRef]
- Anzellotti, S., & Caramazza, A. (2016). From parts to identity: Invariance and sensitivity of face representations to different face halves. Cerebral Cortex, 26(5), 1900–1909. [Google Scholar] [CrossRef]
- Atanasova, K., Lotter, T., Reindl, W., & Lis, S. (2021). Multidimensional assessment of interoceptive abilities, emotion processing and the role of early life stress in inflammatory bowel diseases. Frontiers in Psychiatry, 12, 680878. [Google Scholar] [CrossRef]
- Barrett, L. F. (2006). Solving the emotion paradox: Categorization and the experience of emotion. Personality and Social Psychology Review, 10(1), 20–46. [Google Scholar] [CrossRef]
- Barrett, L. F., Quigley, K. S., Bliss-Moreau, E., & Aronson, K. R. (2004). Interoceptive sensitivity and self-reports of emotional experience. Journal of Personality and Social Psychology, 87(5), 684–697. [Google Scholar] [CrossRef]
- Barrett, L. F., & Simmons, W. K. (2015). Interoceptive predictions in the brain. Nature Reviews Neuroscience, 16(7), 419–429. [Google Scholar] [CrossRef] [PubMed]
- Beaudry, O., Roy-Charland, A., Perron, M., Cormier, I., & Tapp, R. (2014). Featural processing in recognition of emotional facial expressions. Cognition & Emotion, 28(3), 416–432. [Google Scholar]
- Bird, G., & Cook, R. (2013). Mixed emotions: The contribution of alexithymia to the emotional symptoms of autism. Translational Psychiatry, 3(7), e285. [Google Scholar] [CrossRef] [PubMed]
- Blais, C., Linnell, k. J., Caparos, S., & Estéphan, A. (2021). Cultural differences in face recognition and potential underlying mechanisms. Frontiers in Psychology, 12, 627026. [Google Scholar] [CrossRef]
- Brewer, R., Cook, R., & Bird, G. (2016). Alexithymia: A general deficit of interoception. Royal Society Open Science, 3(10), 150664. [Google Scholar] [CrossRef]
- Bridges, D., Pitiot, A., MacAskill, M. R., & Peirce, J. W. (2020). The timing mega-study: Comparing a range of experiment generators, both lab-based and online. PeerJ, 8, e9414. [Google Scholar] [CrossRef] [PubMed]
- Calì, G., Ambrosini, E., Picconi, L., Mehling, W. E., & Committeri, G. (2015). Investigating the relationship between interoceptive accuracy, interoceptive awareness, and emotional susceptibility. Frontiers in Psychology, 6, 1202. [Google Scholar] [CrossRef]
- Calvo, M. G., Fernández-Martín, A., & Nummenmaa, L. (2012). Perceptual, categorical, and affective processing of ambiguous smiling facial expressions. Cognition, 125(3), 373–393. [Google Scholar] [CrossRef]
- Calvo, M. G., & Nummenmaa, L. (2008). Detection of emotional faces: Salient physical features guide effective visual search. Journal of Experimental Psychology: General, 137(3), 471–494. [Google Scholar] [CrossRef]
- Calvo, M. G., & Nummenmaa, L. (2016). Perceptual and affective mechanisms in facial expression recognition: An integrative review. Cognition and Emotion, 30(6), 1081–1106. [Google Scholar] [CrossRef]
- Carbon, C.-C. (2020). Wearing face masks strongly confuses counterparts in reading emotions. Frontiers in Psychology, 11, 566886. [Google Scholar] [CrossRef] [PubMed]
- Chentsova-Dutton, Y. E., & Dzokoto, V. (2014). Listen to your heart: The cultural shaping of interoceptive awareness and accuracy. Emotion, 14(4), 666–678. [Google Scholar] [CrossRef]
- Craig, A. D. (2009). How do you feel—Now? The anterior insula and human awareness. Nature Reviews Neuroscience, 10(1), 59–70. [Google Scholar] [CrossRef]
- Critchley, H. D., & Garfinkel, S. N. (2017). Interoception and emotion. Current Opinion in Psychology, 17, 7–14. [Google Scholar] [CrossRef] [PubMed]
- de Vignemont, F., & Singer, T. (2006). The empathic brain: How, when and why? Trends in Cognitive Sciences, 10, 435–441. [Google Scholar] [CrossRef]
- Ekman, P., & Friesen, W. V. (1971). Constants across cultures in the face and emotion. Journal of Personality and Social Psychology, 17(2), 124–129. [Google Scholar] [CrossRef] [PubMed]
- Ekman, P., & Friesen, W. V. (1978). Facial action coding system (FACS): A technique for the measurement of facial action. Consulting Psychologists Press. [Google Scholar]
- Fujimura, T., & Umemura, H. (2018). Development and validation of a facial expression database based on the dimensional and categorical model of emotions. Cognition and Emotion, 32(8), 1663–1670. [Google Scholar] [CrossRef]
- Fukushima, H., Terasawa, Y., & Umeda, S. (2011). Association between interoception and empathy: Evidence from heartbeat-evoked brain potential. International Journal of Psychophysiology, 79, 259–265. [Google Scholar] [CrossRef]
- Garfinkel, S. N., Seth, A. K., Barrett, A. B., Suzuki, K., & Critchley, H. D. (2015). Knowing your own heart: Distinguishing interoceptive accuracy from interoceptive awareness. Biological Psychology, 104, 65–74. [Google Scholar] [CrossRef]
- Greening, S. G., Mitchell, D. G. V., & Smith, F. W. (2018). Spatially generalizable representations of facial expressions: Decoding across partial face samples. Cortex, 101, 31–43. [Google Scholar] [CrossRef] [PubMed]
- Grootswagers, T. (2020). A primer on running human behavioural experiments online. Behavior Research Methods, 52(6), 2283–2286. [Google Scholar] [CrossRef]
- Grynberg, D., Chang, B., Corneille, O., Maurage, P., Vermeulen, N., Berthoz, S., & Luminet, O. (2012). Alexithymia and the processing of emotional facial expressions (EFEs): Systematic review, unanswered questions and further perspectives. PLoS ONE, 7(8), e42429. [Google Scholar] [CrossRef]
- Gunnery, S. D., & Ruben, M. A. (2015). Perceptions of Duchenne and non-Duchenne smiles: A meta-analysis. Cognition and Emotion, 30(3), 501–515. [Google Scholar] [CrossRef]
- Han, T., Alders, G. L., Greening, S. G., Neufeld, R. W. J., & Mitchell, D. G. V. (2012). Do fearful eyes activate empathy-related brain regions in individuals with callous traits? Social Cognitive and Affective Neuroscience, 7, 958–968. [Google Scholar] [CrossRef]
- Harp, N. R., Langbehn, A., Larsen, J. T., Niedentahl, P. M., & Neta, M. (2023). Face coverings differentially alter valence judgments of emotional expressions. Basic Applied Social Psychology, 45(4), 91–106. [Google Scholar] [CrossRef] [PubMed]
- Haruki, Y., Miyahara, K., Ogawa, K., & Suzuki, K. (2025). Attentional bias towards smartphone stimuli is associated with decreased interoceptive awareness and increased physiological reactivity. Communications Psychology, 3, 42. [Google Scholar] [CrossRef]
- Hessels, R. S., Iwabuchi, T., Niehorster, D. C., Funawatari, R., Benjamins, J. S., Kawakami, S., Nyström, M., Suda, M., Hooge, I. T., Sumiya, M., Heijnen, J. I., & Senju, A. (2025). Gaze behavior in face-to-face interaction: A cross-cultural investigation between Japan and The Netherlands. Cognition, 263, 106174. [Google Scholar] [CrossRef]
- Hübner, A. M., Trempler, I., Gietmann, C., & Schubotz, R. I. (2021). Interoceptive sensibility predicts the ability to infer others’ emotional states. PLoS ONE, 16(10), e0258089. [Google Scholar] [CrossRef]
- Hübner, A. M., Trempler, I., & Schubotz, R. I. (2022). Interindividual differences in interoception modulate behavior and brain responses in emotional inference. Neuroimage, 261, 119524. [Google Scholar] [CrossRef] [PubMed]
- Ikeda, S. (2024). Social sensitivity predicts accurate emotion inference from facial expressions in a face mask: A study in Japan. Current Psychology, 43, 848–857. [Google Scholar] [CrossRef]
- Imahuku, M., Fukushima, H., Nakamura, Y., Myowa, M., & Koike, S. (2020). Interoception is associated with the impact of eye contact on spontaneous facial mimicry. Scientific Reports, 10(1), 19866. [Google Scholar] [CrossRef] [PubMed]
- Karmakar, S., & Das, K. (2023). Investigating the role of visual experience with face-masks in face recognition during COVID-19. Journal of Vision, 22(14), 3243. [Google Scholar] [CrossRef]
- Kastendieck, T., Zilmer, S., & Hess, U. (2022). (Un) mask yourself! Effects of face masks on facial mimicry and emotion perception during the COVID-19 pandemic. Cognition and Emotion, 36(1), 59–69. [Google Scholar] [CrossRef] [PubMed]
- Keltner, D., Sauter, D., Tracy, J., & Cowen, A. (2019). Emotional expression: Advances in basic emotion theory. Journal of Nonverbal Behavior, 43(2), 133–160. [Google Scholar] [CrossRef]
- Koch, A., & Pollatos, O. (2014). Interoceptive sensitivity, body weight and eating behavior in children: A prospective study. Frontiers in Psychology, 5, 1003. [Google Scholar] [CrossRef]
- Lee, E., Kang, J. I., Park, I. H., Kim, J.-J., & An, S. K. (2008). Is a neutral face really evaluated as being emotionally neutral? Psychiatry Research, 157, 77–85. [Google Scholar] [CrossRef] [PubMed]
- Levitan, C. A., Rusk, I., Jonas-Delson, D., Lou, H., Kuzniar, L., Davidson, G., & Sherman, A. (2022). Mask wearing affects emotion perception. i-Perception, 13(3), 1–8. [Google Scholar] [CrossRef]
- Maiorana, N., Dini, M., Poletti, B., Tagini, S., Rita Reitano, M., Pravettoni, G., Priori, A., & Ferrucci, R. (2022). The effect of surgical masks on the featural and configural processing of emotions. International Journal of Environmental Research and Public Health, 19, 2420. [Google Scholar] [CrossRef]
- Maister, L., Tang, T., & Tsakiris, M. (2017). Neurobehavioral evidence of interoceptive sensitivity in early infancy. eLife, 6, e25318. [Google Scholar] [CrossRef]
- Ma-Kellams, C. (2014). Cross-cultural differences in somatic awareness and interoceptive accuracy: A review of the literature and directions for future research. Frontiers in Psychology, 5, 1379. [Google Scholar] [CrossRef]
- Mandler, G., Mandler, J. M., & Uviller, E. T. (1958). Autonomic perception questionnaire (APQ) [Database record]. APA PsycTests. [Google Scholar]
- Marshall, A. C., Gentsch, A., & Schütz-Bosbach, S. (2017). Interoceptive awareness and sensitivity differentially predict neural responses to emotional faces. Social Cognitive and Affective Neuroscience, 12(5), 718–727. [Google Scholar]
- McCrackin, S. D., & Ristic, J. (2022). Emotional context can reduce the negative impact of face masks on inferring emotions. Frontiers in Psychology, 13, 928524. [Google Scholar] [CrossRef]
- Mehling, W. E., Acree, M., Stewart, A., Silas, J., & Jones, A. (2018). The multidimensional assessment of interoceptive awareness, version 2 (MAIA-2). PLoS ONE, 13(12), e0208034. [Google Scholar] [CrossRef]
- Mehling, W. E., Price, C., Daubenmier, J. J., Acree, M., Bartmess, E., & Stewart, A. (2012). The multidimensional assessment of interoceptive awareness (MAIA). PLoS ONE, 7(11), e48230. [Google Scholar] [CrossRef] [PubMed]
- Murphy, J., & Brewer, R. (2024). Interoception: A comprehensive guide. Springer. [Google Scholar]
- Murphy, J., Catmur, C., & Bird, G. (2018). Alexithymia is associated with a multidimensional failure of interoception: Evidence from novel tests. Journal of Experimental Psychology: General, 147(3), 398–408. [Google Scholar] [CrossRef]
- Oliver, L. D., Virani, K., Finger, E. C., & Mitchell, D. G. V. (2014). Is the emotion recognition deficit associated with frontotemporal dementia caused by selective inattention to diagnostic facial features? Neuropsychologia, 60, 84–92. [Google Scholar] [CrossRef]
- Poerio, G. L., Hagan, C. C., Chilton, S., Lush, P., & Critchley, H. D. (2024). Interoceptive attention and mood in daily life. Nature Human Behaviour, 8(6), 1060–1071. [Google Scholar]
- Porges, S. W. (1993). Body perception questionnaire. Laboratory of Developmental Assessment, University of Maryland. [Google Scholar]
- Price, C. J., & Hooven, C. (2018). Interoceptive awareness skills for emotion regulation: Theory and approach of mindful awareness in body-oriented therapy (MABT). Frontiers in Psychology, 9, 798. [Google Scholar] [CrossRef]
- Ren, Q., Marshall, A. C., Liu, J., & Schütz-Bosbach, S. (2024). Listen to your heart: Trade-off between cardiac interoceptive processing and visual exteroceptive processing. NeuroImage, 299, 120808. [Google Scholar] [CrossRef] [PubMed]
- Rinck, M., Primbs, M. A., Verpaalen, I. A. M., & Bijlstra, G. (2022). Face masks impair facial emotion recognition and induce specific emotion confusions. Cognitive Research: Principles and Implications, 7, 83. [Google Scholar] [CrossRef] [PubMed]
- Ross, P., & George, E. (2022). Are face masks a problem for emotion recognition? Not when the whole body is visible. Frontiers in Neuroscience, 16, 915927. [Google Scholar] [CrossRef] [PubMed]
- Russell, J. A. (2003). Core affect and the psychological construction of emotion. Psychological Review, 110, 145–172. [Google Scholar] [CrossRef]
- Russell, J. A., & Barrett, L. F. (1999). Core affect, prototypical emotional episodes, and other things called emotion: Dissecting the elephant. Journal of Personality and Social Psychology, 76, 805–819. [Google Scholar] [CrossRef]
- Russell, J. A., & Bullock, M. (1986). Fuzzy concepts and the perception of emotion in facial expressions. Social Cognition, 4(3), 309–341. [Google Scholar] [CrossRef]
- Russell, J. A., & Lemay, G. (2000). A dimensional contextual perspective on facial expressions. Japanese Psychological Review, 43, 161–176. (In Japanese). [Google Scholar]
- Russell, J. A., Weiss, A., & Mendelson, G. A. (1989). Affect grid: A single-item scale of pleasure and arousal. Journal of Personality and Social Psychology, 57(3), 493–502. [Google Scholar] [CrossRef]
- Sauter, M., Draschkow, D., & Mack, W. (2020). Building, hosting and recruiting: A brief introduction to running behavioral experiments online. Brain Sciences, 10(4), 251. [Google Scholar] [CrossRef]
- Scherer, K. R., & Ellgring, H. (2007). Multimodal expression of emotion: Affect programs or componential appraisal patterns? Emotion, 7(1), 158–171. [Google Scholar] [CrossRef]
- Seth, A. K., & Friston, K. J. (2016). Active interoceptive inference and the emotional brain. Philosophical Transactions of the Royal Society B: Biological Sciences, 371(1708), 20160007. [Google Scholar] [CrossRef]
- Shepherd, J. L., & Rippon, D. (2022). The impact of briefly observing faces in opaque facial masks on emotion recognition and empathic concern. Quarterly Journal of Experimental Psychology, 76(2), 404–418. [Google Scholar] [CrossRef]
- Shimizu, H. (2016). An introduction to the statistical free software HAD: Suggestions to improve teaching, learning and practice data analysis. Journal of Media, Information and Communication, 1, 59–73. (In Japanese). [Google Scholar]
- Shoji, M., Mehling, W. E., Hautzinger, M., & Herbert, B. M. (2018). Investigating multidimensional interoceptive awareness in a Japanese population: Validation of the Japanese MAIA-J. Frontiers in Psychology, 9, 1855. [Google Scholar] [CrossRef] [PubMed]
- Sifneos, P. E. (1973). The prevalence of “alexithymic” characteristics in psychosomatic Patients. Psychotherapy and Psychosomatics, 22, 255−262. [Google Scholar] [CrossRef] [PubMed]
- Slessor, G., Insch, P., Donaldson, I., Sciaponaite, V., Adamowicz, M., & Phillips, L. H. (2022). Adult age differences in using information from the eyes and mouth to make decisions about others’ emotions. Journals of Gerontology: Psychological Sciences, 77(12), 2241–2251. [Google Scholar] [CrossRef]
- Swain, R. H., O’Hare, A. J., Brandley, K., & Tye Gardner, A. (2022). Individual differences in social intelligence and perception of emotion expression of masked and unmasked faces. Cognitive Research: Principles and Implications, 7, 54. [Google Scholar] [CrossRef]
- Tanaka, J. W., Kaiser, M. D., Butler, S., & Le Grand, R. (2012). Mixed emotions: Holistic and analytic perception of facial expressions. Cognition & Emotion, 26(6), 961–977. [Google Scholar]
- Tanaka, Y., Ito, Y., Terasawa, Y., & Umeda, S. (2023). Modulation of heartbeat-evoked potential and cardiac cycle effect by auditory stimuli. Biological Psychology, 182, 108637. [Google Scholar] [CrossRef]
- Terasawa, Y., Moriguchi, Y., Tochizawa, S., & Umeda, S. (2014). Interoceptive sensitivity predicts sensitivity to the emotions of others. Cognition and Emotion, 28(8), 1435–1448. [Google Scholar] [CrossRef] [PubMed]
- Tsantani, M., Podgajecka, V., Gray, K. L. H., & Cook, R. (2022). How does the presence of a surgical face mask impair the perceived intensity of facial emotions? PLoS ONE, 17(1), e0262344. [Google Scholar] [CrossRef] [PubMed]
- van Dyck, Z., VoÈgele, C., Blechert, J., Lutz, A. P. C., Schulz, A., & Herbert, B. M. (2016). The water load test as a measure of gastric interoception: Development of a two-stage protocol and application to a healthy female population. PLoS ONE, 11(9), e0163574. [Google Scholar] [CrossRef]
- Ventura-Bort, C., Wendt, J., & Weymar, M. (2021). The role of interoceptive sensibility and emotional conceptualization for the experience of emotions. Frontiers in Psychology, 12, 712418. [Google Scholar] [CrossRef]
- Wallman-Jones, A., Lewis, J., Mackintosh, K., & Eaves, D. L. (2022). Acute physical-activity related increases in interoceptive ability are not enhanced with simultaneous interoceptive attention. Scientific Reports, 12(1), 15261. [Google Scholar] [CrossRef] [PubMed]
- Weng, H. Y., Feldman, J. L., Leggio, L., Napadow, V., Park, J., & Price, C. J. (2021). Interventions and manipulations of interoception. Trends in Neurosciences, 44, 52–62. [Google Scholar] [CrossRef]
- Wong, H. K., & Estudillo, A. J. (2022). Face masks affect emotion categorisation, age estimation, recognition, and gender classification from faces. Cognitive Research: Principles and Implications, 7, 91. [Google Scholar] [CrossRef]
- Yuki, M., Maddux, W. W., & Masuda, T. (2007). Are the windows to the soul the same in the East and West? Cultural differences in using the eyes and mouth as cues to recognize emotions in Japan and the United States. Journal of Experimentyukal Social Psychology, 43(2), 303–311. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).