You are currently viewing a new version of our website. To view the old version click .
Journal of Personalized Medicine
  • Article
  • Open Access

2 November 2025

Dynamic Facial Health Predicts Psychological First Impressions with Applications to Tailored Treatments for Facial Paralysis

,
,
,
and
1
Department of Psychology, University of Minnesota, Minneapolis, MN 55455, USA
2
School of Statistics, University of Minnesota, Minneapolis, MN 55455, USA
3
Department of Statistics, Grand Valley State University, Allendale, MI 49401, USA
4
Department of Otolaryngology, Harvard Medical School, Boston, MA 02114, USA
This article belongs to the Section Personalized Therapy in Clinical Medicine

Abstract

Background: Past studies demonstrate that certain facial features systematically affect first impressions of psychological traits. However, no previous studies have examined how individual differences in facial health affect first impressions of psychological traits. Methods: In this study, we asked a large sample of fairgoers to give their first impressions of psychological traits in response to viewing videos of unilateral facial paralysis patients with varying degrees of facial functioning. Then, we used linear mixed-effects regression models to understand how individual differences in facial health predict first impressions. Results: Our results replicate previous findings regarding first impressions of faces, such as the attractiveness halo effect, as well as age (maturity) and gender (masculinity) effects. More importantly, our results reveal that facial health, as measured by a clinician-graded scale, is a significant predictor of first impressions. Specifically, we found that individuals with better dynamic facial health (as assessed by clinicians) were perceived to be more competent and more affiliative, but not more dominant, than individuals with lower levels of dynamic facial functioning. Conclusions: Our results have important implications for personalized medicine via the development and refinement of individually tailored therapies to improve facial reanimation surgery outcomes.

1. Introduction

1.1. Motivating Background

The questions of how and why facial features influence first impressions have sparked the interest of many [,,,,,]. Past studies have revealed robust findings regarding psychological impressions of facial characteristics. One well-established finding is the “attractiveness halo effect”, which is the finding that attractive people are perceived to have better psychological traits [,,]. Specifically, people who are perceived to be more attractive are perceived to be more intelligent, sociable, and dominant—even though such perceptions may be inaccurate []—and such effects have been shown to exist across cultures []. These findings have motivated researchers to study the general and specific mechanisms that associate facial features with perceived psychological characteristics.
Theories of social perception offer general mechanisms by which people may associate facial features with psychological traits [,,]. The ecological theory of social perception [] suggests that the attractiveness halo effect is an example of overgeneralization. Specifically, the ecological approach predicts that people may overgeneralize judgements based on characteristics that are important for survival and/or functioning []. As an example, the ecological theory suggests that physical attractiveness may be an indicator of reproductive fitness, which is an important characteristic for species survival []. Such evolutionary adaptations can account for the overgeneralization of facial attractiveness to positive psychological traits and of facial anomalies to negative psychological traits [].

1.2. Limitations of Past Work

The idea that people overgeneralize facial features important for survival and/or functioning can elucidate other well-known findings about first impressions []. For example, past research suggests that facial maturity is positively correlated with perceived competence []. Facial maturity may be an indicator of experience and wisdom, and perceived competence is important for electing leaders [,]. As another example, research suggests that facial masculinity is positively correlated with perceived dominance []. Facial masculinity may be an indicator of physical strength, and the quick identification of threatening faces is vital for survival [].
Although past studies have provided general insights into how facial features affect first impressions, they have largely ignored specific details regarding how individual differences in facial function affect first impressions (e.g., [,,]). This is due to two reasons. First, many past studies have focused on impressions drawn from photos of faces, which makes it impossible to examine the influence of dynamic aspects (e.g., [,]). There exist studies exploring how the medium of the presented information (e.g., photo vs. video) affects the first impressions of faces (e.g., [,,,,]); however, this body of work does not explore how individual differences in facial health and function affect first impressions. Second, until recently, researchers have lacked a systematic scale for the quantification of individual differences in facial function. As a result, much of our current knowledge about psychological impressions of facial characteristics lacks information about the influence of facial health and function.

1.3. Present Research

We address this gap in the literature by exploring how individual differences in facial health relate to psychological first impressions. Specifically, the goals of this study are to determine (i) whether overall facial health is a relevant factor in forming first impressions of psychological traits and (ii) which aspects of facial health are most relevant for determining those first impressions. Given the well-known “attractiveness halo effect”, we explore the relationships between facial health and first impressions after statistically controlling for perceived attractiveness. We also control for age and gender effects, given that past research has demonstrated associations between first impressions and both facial maturity [] and facial masculinity []. Controlling for covariates that are known to affect first impressions of faces allows us to explore how facial health affects first impressions above and beyond well-known effects.
We measure facial health using the eFACE scale [], which is a clinician-graded measure of facial health and function. The eFACE scale grades facial health in terms of three subscores, Static, Dynamic, and Synkinesis, along with an overall facial health score. We hypothesize that, after controlling for the covariates of interest, (A) overall facial health will be significantly positively associated with the outcome variables, and (B) each subdomain of facial health will significantly contribute to first impressions, with the effect sizes ordered such that Dynamic > Static > Synkinesis. Note that hypothesis A is suggested by past research that demonstrates overall facial health is positively associated with perceived disfigurement [,]. Hypothesis B is proposed because the same past research has revealed that each subdomain of facial health affects first impressions of disfigurement, with resulting effect sizes in the aforementioned order.
We expect our results to be useful for advancing personalized medicine within the field of facial reanimation surgery and rehabilitation. In particular, our experimental design and analysis makes it possible to examine precisely how individual differences in facial health and function, as measured by the eFACE scale, influence first impressions of psychological traits. By studying the relative influence of each component of facial health, our approach makes it possible to directly study the influence of different facial function dimensions on psycho-social perceptions, which has the potential to enhance pre- and post-surgical outcomes. In particular, we anticipate our findings to be informative for facial reanimation surgeons and rehabilitative therapists, as they design facial movement exercises to help patients (re)learn to express psychologically and socially appropriate facial expressions.

2. Materials and Methods

2.1. Raters

Participants were fairgoers at the 2016 Minnesota State Fair who were recruited as they passed by the Driven to Discover building, a University of Minnesota facility on the fairgrounds. Participants were asked to report the number of alcoholic drinks they had consumed that day, so that inebriated fairgoers could be excluded from the analyses. We predetermined that those who reported consuming three or more drinks should be excluded from the analyses, which resulted in the exclusion of data from nine fairgoers. After removing these individuals, our sample consisted of 550 raters with the following demographics: 383 females (mean age, 40.01 years; range, 18 to 80), 164 males (mean age, 39.25 years; range, 18 to 86), and 3 who identified with the gender “other” (ages 18, 21, and 63 years). The three individuals who identified with a gender of “other” were excluded from the analyses, because three individuals is not enough to provide reliable effect estimates. Due to small amounts of missing data on the different response and/or predictor variables, our effective sample size differed slightly for each analysis: 538 raters for Competence, 533 raters for Affiliation, and 536 raters for Dominance.

2.2. Materials

2.2.1. Patient Videos

Videos consisted of 61 patients with varying degrees of unilateral facial paralysis, ranging from low to high facial functionality. Each video lasted 3–5 s and was framed around the patient’s face and upper body with a blue backdrop. The videos showed the patient forming a volitional smile (from neutral expression to full smile), which is part of a larger clinical assessment of the patients’ facial function. Patients consisted of 42 females (mean age, 47.74 years; range, 18 to 80) and 19 males (mean age, 50.79 years; range, 24 to 84). The 61 patient videos were selected from 500 patient videos obtained from the Massachusetts Eye and Ear Infirmary. The videos were selected using a multi-step procedure that involved (i) an initial screening to define the patient population, (ii) stratified random sampling to obtain a sample of videos with diverse combinations of facial health, and (iii) quality control checks to remove videos with various issues, as described below.
First, the 500 videos were screened to remove children (<18 years old), post-surgical patients, and patients with severe facial deformities. There were 270 videos (168 female) that met these initial screening criteria. Second, a stratified sampling procedure was used to select 64 videos that spanned a wide variety of facial function. This was done by partitioning the three-dimensional domain of eFACE subscores into 27 mutually exclusive bins (representing all combinations of low, medium, and high on the three eFACE subscores) and then randomly sampling up to three patients from each bin. Third, the videos were subjected to quality control checks, which included an inspection for completeness of recording, image quality, and inclusion of the entire smile function (i.e., all phases of the smile). These quality control checks resulted in the exclusion of three patients (two due to video quality and one due to incomplete smile), which produced our sample of sixty-one patients.

2.2.2. eFACE Scale

Each patient’s degree of facial function was determined by the eFACE scale, which is a clinician-graded scale for assessing facial function in patients with unilateral facial paralysis []. The eFACE scale consists of 15 items that combine to form three subscores measuring clinically relevant factors of facial health: Static, Dynamic, and Synkinesis. The Static subscore measures the health of the still (i.e., resting) face, the Dynamic subscore measures the health of the face during voluntary movement tasks (e.g., raise eyebrows, close eyes, etc.), and the Synkinesis subscore measures the health of the face in terms of lack of synkinetic activity (i.e., involuntary movements). The eFACE scale was found to have high inter- and intra-rater reliability [,,], as well as high predictive validity for both expert and layperson perceptions of disfigurement [,,,].
All of the subscores range from 0 to 100 such that larger values correspond to better facial health. Specifically, “a value of 0 indicated the most extreme malposition (static items), total absence of movement (dynamic items), or the most severe synkinesis (synkinesis items). A value of 100 indicated balanced position (static items), normal movement (dynamic items), or absence of synkinesis (synkinesis items).” (p. 224e of []). The eFACE subscores combine to form an eFACE Total score, which ranges from 0 to 100, such that higher values indicate better overall facial function. See Figure 1 for a visualization of the distributions of and associations between the eFACE Total scores.
Figure 1. Kernel density estimates showing the sample distribution of the scores (diagonal), scatterplots showing the bivariate associations between the scores (lower triangle), and Pearson correlation between scores (upper triangle).

2.2.3. Interpersonal Adjective Scales-Revised

The Interpersonal Adjective Scales-Revised [] measure aspects of personality that deal with interacting with others. The IAS-R consists of 64 adjectives that could be used to describe a person, e.g., friendly, cunning, etc. Each item is rated on a Likert-type scale from 1 (extremely inaccurate) to 8 (extremely accurate), indicating how well the given adjective describes a particular individual. The adjectives can be categorized into eight groups (or octants), each consisting of eight adjectives. The octants are theorized to take on the structure of a circumplex, which is embedded within two orthogonal dimensions: Affiliation (i.e., warmth/agreeableness) and Dominance [,]. Scores on these two dimensions are obtained by taking a linear combination of the octant scores, which are linear combinations of the item scores []. To reduce the burden on study participants, we created a brief version of the scale by selecting two items from each octant, i.e., sixteen items in total, where the two items were selected to have high communalities [].

2.3. Procedure

Raters were provided a tablet (Apple iPad3, 16GB, iOS 8) with a custom-built application to display the patient videos, and a rating sheet to record their responses. The iPad application (i) obtained informed consent from the rater, (ii) provided basic instructions on how to use the app, and then (iii) randomly sampled 1 of the 61 patient videos to display. The rater was asked to both evaluate the patient using 16 adjectives from the IAS-R (see Appendix B) and judge the patient’s Disfigurement, Attractiveness, and Competence (using a 10-point Likert scale). A glossary providing the adjective definitions was available to the raters (see Appendix A and Appendix B). The videos could be replayed as many times as desired while completing the rating sheet, but participants were instructed to rate the videos based on their “gut reaction” to the person in the video. After completing their evaluation of a video, raters could choose to rate more videos (46 participants rated 1 video, 53 participants rated 2 videos, 145 participants rated 3 videos, 294 participants rated 4 videos, and 9 participants rated 5 videos). If the rater chose to rate more than one video, subsequent videos were randomly sampled without replacement such that a rater would not rate the same video more than once.

2.4. Analysis

To facilitate comparisons between the results, the Competence rating variable was transformed to the interval [ 15 , 15 ] , which is the approximate range of the Affiliation and Dominance scores. Relating the response variables of interest (i.e., Competence, Affiliation, and Dominance) to the eFACE scores requires a method that can quantify the effects of facial health on the response variables, while accounting for (i) the influence of the covariates, (ii) the correlation between the repeated responses for different raters, and (iii) the correlation between the repeated ratings for different stimuli. We used linear mixed-effects regression (LMER), which allows us to explore the (fixed) effects of facial health on the outcome variables while controlling for the (fixed) effects of the covariates and the (random) effects of the raters and patients. For each response variable, we fit two models: (i) a model including the eFACE total score as the fixed effect predictor and (ii) a model including the additive effects of the three eFACE subscores as fixed effects predictors. For all models, we controlled for the following fixed effects: the rater’s Age and Gender, the patient’s Age and Gender, and the rater’s perceived Attractiveness of the patient. Furthermore, for all models we included random intercepts for both the raters and the patients. The LMER models were fit via the lme4 [] and lmerTest [] packages in the R software environment []. The R code and data to reproduce our analyses are provided in the Supplementary Materials.

3. Results

3.1. eFACE Total Score Models

Table 1, Table 2 and Table 3 display the results for the three models that used the eFACE Total score as a predictor. The results in Table 1, Table 2 and Table 3 reveal several interesting effects, and only one effect is consistent across all three response variables: Attractiveness is a significant predictor, such that people who are perceived to be more attractive are also perceived to have more Competence, Affiliation, and Dominance. The other effects differ depending on the response variable. We discuss the results separately for each response variable.

3.1.1. Competence

The results for the Competence ratings are displayed in Table 1. We find that both the patient’s Age and the rater’s Age are significant predictors of the Competence ratings. Older patients are perceived to be more competent, whereas older raters perceive patients to be less competent, which suggests that the relative age of the perceiver to the target may be driving these relationships. The patient’s Gender and the rater’s Gender are not significantly related to the Competence ratings. The rater’s perceived Attractiveness of the patient has the largest effect on the rater’s perceived Competence of the patient, such that patients who are rated to be more attractive are also rated to be more competent. Controlling for the covariates, the patient’s degree of facial health (as measured by eFACE Total) was significantly related to Competence ratings, such that patients with better facial health were rated to be more competent.
Table 1. Competence from eFACE Total score: coefficient and variance component information.
Table 1. Competence from eFACE Total score: coefficient and variance component information.
EffectEstimateSE95% CIp
LL UL
Fixed effects
  (Intercept)−7.5600.792−9.120−5.9990.000
  Patient Age0.0510.0100.0310.0700.000
  Patient Gender a−0.1490.338−0.8250.5270.661
  Rater Age−0.0450.011−0.066−0.0230.000
  Rater Gender a−0.1800.389−0.9440.5830.643
  Attractiveness1.6980.0701.5621.8350.000
  eFACE Total b0.0320.0120.0080.0570.009
Random effects
  Rater9.2590.0437.19211.3920.000
  Patient0.4950.0410.0131.0260.017
  Residual22.3330.09920.60924.273NA
Note. Number of raters = 538; number of patients = 61; total N = 1757 ; SE = standard error; CI = confidence interval; LL = lower limit; UL = upper limit. R-squared fixed effects only = 0.316; R-squared fixed and random effects = 0.623. a 0 = Female; 1 = Male. b Centered across patients.

3.1.2. Affiliation

Table 2 presents the results for the Affiliation rating data. The Affiliation fixed effects results reveal a similar pattern as the Competence results, with one noteworthy exception: we find that male patients are rated to have less Affiliation than female patients. The other fixed effects produce the same interpretation as the Competence results, i.e., (i) older patients are perceived to be more Affiliative, (ii) older raters perceive patients to be less Affiliative, (iii) more Attractive patients are rated to be more Affiliative, and (iv) patients with better facial health were rated to be more Affiliative.
Table 2. Affiliation from eFACE Total score: coefficient and variance component information.
Table 2. Affiliation from eFACE Total score: coefficient and variance component information.
EffectEstimateSE95% CIp
LL UL
Fixed effects
  (Intercept)−4.1370.813−5.747−2.5260.000
  Patient Age0.0560.0130.0300.0830.000
  Patient Gender a−1.4860.450−2.387−0.5850.002
  Rater Age−0.0210.008−0.035−0.0060.006
  Rater Gender a−0.2590.271−0.7920.2740.340
  Attractiveness1.0600.0550.9511.1690.000
  eFACE Total b0.0380.0160.0060.0690.021
Random effects
  Rater3.1220.0432.1414.1530.000
  Patient1.9940.0461.1352.9380.000
  Residual14.3990.07913.27915.643NA
Note. Number of raters = 533; number of patients = 61; total N = 1718 ; SE = standard error; CI = confidence interval; LL = lower limit; UL = upper limit. R-squared fixed effects only = 0.253; R-squared fixed and random effects = 0.543. a 0 = Female; 1 = Male. b Centered across patients.

3.1.3. Dominance

The results for the Dominance ratings are displayed in Table 3. The patient’s Age and the rater’s Age are not significant predictors. In contrast, the patient’s Gender and the rater’s Gender are significant predictors. Male patients are perceived as more dominant, and male raters perceive patients as more dominant. The rater’s perceived Attractiveness of the patient still has a large effect: more attractive patients are rated to be more dominant. Controlling for the covariates, the patient’s degree of facial health was not significantly related to the perceived Dominance of the patient.
Table 3. Dominance from eFACE Total score: coefficient and variance component information.
Table 3. Dominance from eFACE Total score: coefficient and variance component information.
EffectEstimateSE95% CIp
LL UL
Fixed effects
  (Intercept)−7.1610.965−9.078−5.2440.000
  Patient Age0.0250.017−0.0080.0590.133
  Patient Gender a1.1390.5670.0032.2750.049
  Rater Age0.0040.007−0.0100.0180.565
  Rater Gender a0.6950.2630.1791.2110.008
  Attractiveness0.8250.0560.7150.9350.000
  eFACE Total b0.0000.020−0.0400.0400.997
Random effects
  Rater2.4260.0461.4643.4290.000
  Patient3.5220.0542.1555.0190.000
  Residual15.1080.08213.92516.425NA
Note. Number of raters = 536; number of patients = 61; total N = 1725 ; SE = standard error; CI = confidence interval; LL = lower limit; UL = upper limit. R-squared fixed effects only = 0.126; R-squared fixed and random effects = 0.469. a 0 = Female; 1 = Male. b Centered across patients.

3.2. eFACE Subscore Models

The results for the eFACE subscores reveal similar patterns as the Total score models with respect to fit information and parameter estimates (see Table A1, Table A2 and Table A3 in Appendix A). As a result, we only comment on aspects of the results which differ. Focusing on the contributions of the eFACE components, the Dynamic subscore is significantly related to Competence and Affiliation ratings, such that patients with better Dynamic facial function are rated to be more competent (see Table A1) and affiliative (see Table A2). In contrast, the patient’s Dynamic subscore is not significantly related to the Dominance ratings (see Table A3). Given the other effects, the eFACE Static and Synkinesis subscores are not significantly related to the patient’s perceived Competence, Affiliation, or Dominance (see Table A1 and Table A2). Thus, the Dynamic subscores seem to be driving the relationships observed in the Total score analysis (see Figure 2).
Figure 2. Relationship between first impressions on psychological traits (rows) and overall facial health (left), static facial health (left center), dynamic facial health (right center), and synkinesis facial health (right). Points denote the average (across raters) of the model residuals with the plotted effect excluded from the prediction, whereas the blue line denotes the estimated effect. The shadings denote a 95% confidence interval, and the dotted line denotes zero (i.e., no effect).

4. Discussion

4.1. Summary of Findings

Controlling for well-known effects, our results clearly demonstrate that differences in facial health influence first impressions drawn from faces. First, we established that overall facial health can predict first impressions of Competence and Affiliation, but not Dominance. Then we demonstrated that dynamic facial health is the driving component behind these associations. In other words, dynamic facial health alone predicts both Competence and Affiliation—but not Dominance. Conditioned on dynamic facial health, an individual’s level of static (resting) facial health and synkinesis (involuntary movement) facial health had no significant effects on first impressions of Competence, Affiliation, or Dominance. These findings complement past studies, which have noted that facial dynamics can influence psychological first impressions drawn from healthy faces [,,,,].
Consistent with previous work [,], we found that patients who were perceived as more attractive were also perceived to have more favorable psychological traits, i.e., more Competence, Affiliation, and Dominance. We also observed that male patients were perceived to be more dominant and less affiliative, which agrees with past findings regarding the effects of facial masculinity on first impressions []. We found that older patients were perceived to have higher levels of Competence and Affiliation. These results agree with past findings regarding the effects of facial maturity on Competence first impressions [] but somewhat disagree with previous results on Affiliation first impressions: Oosterhof and Todorov [] found that more mature faces were perceived as less trustworthy. However, it should be noted that our result is not directly comparable to the result of Oosterhof and Todorov [] for two reasons: (a) we used Affiliation as the response, whereas they used Trustworthiness, and (b) we used real videos of patients with different ages, whereas they used computer-generated faces with different (simulated) degrees of facial maturity.

4.2. Ecological Interpretation

The idea that facial health can influence first impressions is well-supported by the ecological approach of facial perception. As predicted by ecological theory [,], we found evidence that people may overgeneralize deficits in facial health, causing them to perceive negative psychological attributes from unhealthy faces. Note that raters’ perceptions of Disfigurement tend to agree with clinicians’ ratings on the eFACE scale [], evidencing that raters are, in fact, perceptive of the deficits in patients’ facial health. The importance of dynamic facial health, as compared to static facial health, is also well-supported by the ecological approach, which posits that dynamic functioning provides more information relevant to survival and functioning. This highlights the importance of utilizing dynamic stimuli (e.g., videos or virtual reality) instead of static images when trying to study the affects of facial health on psychological first impressions.
Interestingly, our results suggest that facial health indicators are important for forming judgements about one’s Competence and Affiliation, but not one’s Dominance. Interpreting these findings from the ecological perspective offers some intriguing insights regarding the sorts of first impressions that are obtained from different facial cues. Note that perceptions of Competence and Affiliation are both relevant to choosing reliable leaders and companions [], making them important for long-term survival and social functioning. In contrast, perceptions of Dominance are relevant for determining whether a person has the potential to be a threat [], making this trait important for short-term (or immediate) survival and functioning. Consequently, our results suggest that facial health plays a role in the overgeneralization of psychological traits that are relevant to long-term, but not necessarily short-term, survival and functioning.

4.3. Limitations and Future Directions

Our research establishes new directions regarding the influence of dynamic facial health on first impressions of psychological traits. The primary limitation of our study is that the fairgoers’ first impressions were based solely on a 3–5 s video of the patient smiling. Future work is needed to determine whether our results generalize to other presentation mediums (e.g., face-to-face interaction) and/or other facial tasks (e.g., having a conversation). Furthermore, although the sample of fairgoers was diverse compared to the WEIRD samples typically used in research [], future work should validate the results in more diverse samples. Finally, additional research is required to determine whether improving one’s dynamic facial health can improve first impressions, e.g., by obtaining ratings of patients before and after facial surgery. Such extensions could (i) elucidate which aspects of dynamic facial health are most important for forming first impressions, (ii) help guide therapeutic efforts for individuals experiencing issues with facial expression, and (iii) inform both patients and clinicians about the progress of facial rehabilitation efforts.

5. Conclusions

This study’s findings, demonstrating that dynamic facial health significantly influences first impressions of competence and affiliation, offer crucial insights for advancing personalized medicine in facial reanimation surgery and rehabilitation. By establishing dynamic facial health as a primary driver of these social perceptions, the research provides a clear target for therapeutic interventions. For instance, understanding that improving dynamic facial health, rather than static features or synkinesis, is key to enhancing positive first impressions can guide surgeons and rehabilitation specialists in prioritizing specific functional outcomes. Future research exploring the impact of facial surgery on first impressions (e.g., pre- and post-operative ratings) can further refine personalized treatment plans, elucidating which aspects of dynamic facial health are most critical for social integration and well-being. This personalized approach, focusing on the dynamic aspects of facial expression, holds the potential to significantly improve the holistic outcomes for individuals undergoing facial reanimation, extending beyond physical restoration to encompass improved social interactions and quality of life.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/jpm15110530/s1, Dataset: the file “eface-ias-data.csv” is a comma separated values (CSV) file containing the raw data; Report: the file “eface-ias-code.html” is an Rmarkdown generative HTML file that contains all of the R code and output from our analysis.

Author Contributions

Conceptualization, N.E.H., S.J.G. and S.L.-P.; methodology, N.E.H., S.J.G. and S.L.-P.; software, N.E.H. and L.N.B.; validation, S.L.-P. and T.A.H.; formal analysis, N.E.H. and L.N.B.; investigation, N.E.H., S.J.G., T.A.H. and S.L.-P.; resources, T.A.H. and S.L.-P.; data curation, S.J.G., T.A.H. and S.L.-P.; writing—original draft preparation, N.E.H. and L.N.B.; writing—review and editing, S.J.G., T.A.H. and S.L.-P.; visualization, N.E.H. and S.L.-P.; supervision, T.A.H. and S.L.-P.; project administration, N.E.H., S.J.G. and S.L.-P.; funding acquisition, N.E.H., T.A.H., S.J.G. and S.L.-P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by faculty start-up funds from the University of Minnesota and a Lions Club International grant.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board (or Ethics Committee) of the University of Minnesota (proposal: 1605S87229; date of approval: 12 July 2016).

Data Availability Statement

The data and R code necessary to reproduce our results are included with the Supplementary Materials that accompany this article.

Acknowledgments

Data were collected at the University of Minnesota’s Driven to Discover facility. Patient videos were recorded at Massachusetts Eye and Ear Infirmary at Harvard Medical School. During the preparation of this manuscript/study, the author(s) used generative AI [Gemini, 2.5 Flash] for the purposes of [formatting tables, producing text, and typesetting references]. The authors have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
IASInterpersonal Adjectives Scale
LMERLinear Mixed-Effects Regression

Appendix A. Supplementary Tables of Results

Table A1. Competence from eFACE subscores: coefficient and variance component information.
Table A1. Competence from eFACE subscores: coefficient and variance component information.
EffectEstimateSE95% CIp
LL UL
Fixed effects
  (Intercept)−7.0270.826−8.655−5.4000.000
  Patient Age0.0400.0110.0180.0620.001
  Patient Gender a−0.0480.339−0.7270.6310.888
  Rater Age−0.0440.011−0.065−0.0230.000
  Rater Gender a−0.1930.389−0.9580.5710.620
  Attractiveness1.6860.0701.5481.8230.000
  eFACE Static b−0.0110.010−0.0320.0100.309
  eFACE Dynamic b0.0260.0080.0110.0410.001
  eFACE Synkinesis b0.0030.007−0.0100.0170.601
Random effects
  Rater9.2990.0437.21711.4270.000
  Patient0.4280.0430.0000.8700.040
  Residual22.3160.09920.59824.261NA
Note. Number of raters = 538; number of patients = 61; total N = 1757 ; SE = standard error; CI = confidence interval; LL = lower limit; UL = upper limit. R-squared fixed effects only = 0.317; R-squared fixed and random effects = 0.623. a 0 = Female; 1 = Male. b Centered across patients.
Table A2. Affiliation from eFACE subscores: coefficient and variance component information.
Table A2. Affiliation from eFACE subscores: coefficient and variance component information.
EffectEstimateSE95% CIp
LL UL
Fixed effects
  (Intercept)−3.3970.820−5.023−1.7720.000
  Patient Age0.0400.0140.0130.0670.005
  Patient Gender a−1.1330.418−1.971−0.2950.009
Rater Age−0.0200.008−0.035−0.0060.007
  Rater Gender a−0.2730.271−0.8050.2600.315
  Attractiveness1.0500.0550.9411.1580.000
  eFACE Static b−0.0200.013−0.0460.0060.135
  eFACE Dynamic b0.0360.0090.0170.0550.000
  eFACE Synkinesis b−0.0100.008−0.0270.0060.201
Random effects
  Rater3.1160.0432.1344.1430.000
  Patient1.5120.0440.7622.1810.000
  Residual14.4040.07913.28515.650NA
Note. Number of raters = 533; number of patients = 61; total N = 1718 ; SE = standard error; CI = confidence interval; LL = lower limit; UL = upper limit. R-squared fixed effects only = 0.273; R-squared fixed and random effects = 0.542. a 0 = Female; 1 = Male. b Centered across patients.
Table A3. Dominance from eFACE subscores: coefficient and variance component information.
Table A3. Dominance from eFACE subscores: coefficient and variance component information.
EffectEstimateSE95% CIp
LL UL
Fixed effects
  (Intercept)−6.7801.069−8.907−4.6520.000
  Patient Age0.0170.019−0.0210.0560.375
  Patient Gender a1.2200.5910.0372.4040.043
  Rater Age0.0040.007−0.0100.0190.560
  Rater Gender a0.6930.2630.1771.2100.009
  Attractiveness0.8240.0560.7130.9340.000
  eFACE Static b−0.0150.018−0.0520.0210.409
  eFACE Dynamic b0.0080.013−0.0180.0340.533
  eFACE Synkinesis b−0.0030.011−0.0260.0200.772
Random effects
  Rater2.4340.0461.4693.4360.000
  Patient3.6050.0562.1134.9310.000
  Residual15.1020.08213.92116.420NA
Note. Number of raters = 536; number of patients = 61; total N = 1725 ; SE = standard error; CI = confidence interval; LL = lower limit; UL = upper limit. R-squared fixed effects only = 0.128; R-squared fixed and random effects = 0.469. a 0 = Female; 1 = Male. b Centered across patients.

Appendix B. Interpersonal Adjectives Glossary

  • Assertive: tends to be aggressive and outspoken with others.
  • Coldhearted: have little warmth or feelings for others; unfeeling; harsh.
  • Cunning: crafty, skillful at manipulating others, devious.
  • Friendly: open, accepting, warm around others.
  • Gentle-hearted: warm or kind to others.
  • Introverted: enjoys being alone, reserved around others.
  • Outgoing: enjoy meeting other people.
  • Self-confident: sure of oneself around others, devious.
  • Shy: lacking in self-confidence; tends to be uncomfortable around others.
  • Sly: crafty, secretive, or cunning in dealing with others.
  • Tender-hearted: having a kind, gentle or sentimental nature.
  • Timid: tends to be fearful or uncomfortable around others.
  • Uncrafty: not tricky or sly when dealing with others.
  • Uncunning: not crafty or sly, tends to be straightforward with others.
  • Unsociable: doesn’t enjoy meeting people or being in the company of others.
  • Unsympathetic: not interested or concerned about others’ feelings or problems.

References

  1. Berry, D.S.; McArthur, L.Z. Perceiving character in faces: The impact of age-related craniofacial changes on social perception. Psychol. Bull. 1986, 100, 3–18. [Google Scholar] [CrossRef] [PubMed]
  2. Berry, D.S. Accuracy in social perception: Contributions of facial and vocal information. J. Personal. Soc. Psychol. 1991, 61, 298–307. [Google Scholar] [CrossRef]
  3. Secord, P.F. Facial features and inference processes in interpersonal perception. In Person Perception and Interpersonal Behavior; Tagiuri, R., Luigi, P., Eds.; Stanford University Press: Stanford, CA, USA, 1958; pp. 300–315. [Google Scholar]
  4. Zebrowitz, L.A. First impressions from faces. Curr. Dir. Psychol. Sci. 2017, 26, 237–242. [Google Scholar] [CrossRef] [PubMed]
  5. Hu, Y.; O’Toole, A.J. First impressions: Integrating faces and bodies in personality trait perception. Cognition 2023, 231, 105309. [Google Scholar] [CrossRef]
  6. McCulloch, K.; Steele, Y.; Gheorghiu, A.I. The Relationship Between Face-Based First Impressions and Perceptions of Purity and Compared to Other Moral Violations. Behav. Sci. 2024, 14, 1205. [Google Scholar] [CrossRef]
  7. Eagly, A.H.; Ashmore, R.D.; Makhijani, M.G.; Longo, L.C. What is beautiful is good, but?: A meta-analytic review of research on the physical attractiveness stereotype. Psychol. Bull. 1991, 110, 109–128. [Google Scholar] [CrossRef]
  8. Feingold, A. Good-looking people are not what we think. Psychol. Bull. 1992, 111, 304–341. [Google Scholar] [CrossRef]
  9. Batres, C.; Shiramizu, V. Examining the “attractiveness halo effect” across cultures. Curr. Psychol. 2023, 42, 25515–25519. [Google Scholar] [CrossRef]
  10. Sutherland, C.A.M.; Rhodes, G.; Burton, N.S.; Young, A.W. Do facial first impressions reflect a shared social reality? Br. J. Psychol. 2020, 111, 215–232. [Google Scholar] [CrossRef]
  11. Sutherland, C.A.M.; Young, A.W. Understanding trait impressions from faces. Br. J. Psychol. 2022, 113, 1056–1078. [Google Scholar] [CrossRef]
  12. Bjornsdottir, R.T.; Connor, P.; Rule, N.O. Social judgments from faces and bodies. J. Personal. Soc. Psychol. 2024, 127, 455–468. [Google Scholar] [CrossRef]
  13. McArthur, L.Z.; Baron, R.M. Toward an ecological theory of social perception. Psychol. Rev. 1983, 90, 215–238. [Google Scholar] [CrossRef]
  14. Zebrowitz, L.A.; Fellous, J.M.; Mignault, A.; Andreoletti, C. Trait impressions as overgeneralized responses to adaptively significant facial qualities: Evidence from connectionist modeling. Personal. Soc. Psychol. Rev. 2003, 7, 194–215. [Google Scholar] [CrossRef]
  15. Zebrowitz, L.A.; Rhodes, G. Sensitivity to “bad genes” and the anomalous face overgeneralization effect: Cue validity, cue utilization, and accuracy in judging intelligence and health. J. Nonverbal Behav. 2004, 28, 167–185. [Google Scholar] [CrossRef]
  16. Walker, M.; Wänke, M. Caring or daring? Exploring the impact of facial masculinity/femininity and gender category information on first impressions. PLoS ONE 2017, 12, e0181306. [Google Scholar] [CrossRef]
  17. Olivola, C.Y.; Todorov, A. Elected in 100 milliseconds: Appearance-based trait inferences and voting. J. Nonverbal Behav. 2010, 34, 83–110. [Google Scholar] [CrossRef]
  18. Todorov, A.; Mandisodza, A.N.; Goren, A.; Hall, C.C. Inferences of competence from faces predict election outcomes. Science 2005, 308, 1623–1626. [Google Scholar] [CrossRef]
  19. Ballew, C.C.I.; Todorov, A. Predicting political elections from rapid and unreflective face judgments. Proc. Natl. Acad. Sci. USA 2007, 104, 17948–17953. [Google Scholar] [CrossRef] [PubMed]
  20. Oosterhof, N.; Todorov, A. The functional basis of face evaluation. Proc. Natl. Acad. Sci. USA 2008, 105, 11087–11092. [Google Scholar] [CrossRef]
  21. Bar, M.; Neta, M.; Linz, H. Very first impressions. Emotion 2006, 6, 269–278. [Google Scholar] [CrossRef] [PubMed]
  22. Kleider-Offutt, H.; Meacham, A.; Branum-Martin, L.; Capodanno, L. The role of facial features in ratings of dominance, threat, and stereotypicality. Cogn. Res. Princ. Implic. 2021, 6, 53. [Google Scholar] [CrossRef]
  23. Jaeger, B.; Jones, A.L. Which Facial Features Are Central in Impression Formation? Soc. Psychol. Personal. Sci. 2022, 13, 553–561. [Google Scholar] [CrossRef]
  24. Paruzel-Czachura, M.; Workman, C.I.; El Toukhy, N.; Chatterjee, A. First impressions: Do faces with scars and palsies influence warmth, competence and humanization? Br. J. Psychol. 2024, 115, 706–722. [Google Scholar] [CrossRef]
  25. Ekman, P.; Friesen, W.V.; Ancoli, S. Facial signs of emotional experience. J. Personal. Soc. Psychol. 1980, 39, 1125–1134. [Google Scholar] [CrossRef]
  26. Chetrit, V.; Kannan, S.; Cohen-Levy, J. First Impression Formation: Impact of Lower Anterior Facial Height in a Gender- and Ethnicity-Diverse Photographic Model. AJO-DO Clin. Companion 2025, in press. [Google Scholar] [CrossRef]
  27. Gunaydin, G.; Selcuk, E.; Zayas, V. Impressions Based on a Portrait Predict, 1-Month Later, Impressions Following a Live Interaction. Soc. Psychol. Personal. Sci. 2017, 8, 36–44. [Google Scholar] [CrossRef]
  28. Lander, K. Relating visual and vocal attractiveness for moving and static faces. Anim. Behav. 2008, 75, 817–822. [Google Scholar] [CrossRef]
  29. Roberts, S.C.; Saxton, T.K.; Murray, A.K.; Burriss, R.P.; Rowland, H.M.; Little, A.C. Static and dynamic facial images cue similar attractiveness judgements. Ethology 2009, 115, 588–595. [Google Scholar] [CrossRef]
  30. Rubenstein, A.J. Variation in perceived attractiveness: Differences between dynamic and static faces. Psychol. Sci. 2005, 16, 759–762. [Google Scholar] [CrossRef] [PubMed]
  31. Jiang, Z.; Li, D.; Li, Z.; Yang, Y.; Liu, Y.; Yue, X.; Wu, Q.; Yang, H.; Cui, X.; Xue, P. Comparison of face-based and voice-based first impressions in a Chinese sample. Br. J. Psychol. 2024, 115, 20–39. [Google Scholar] [CrossRef] [PubMed]
  32. Banks, C.A.; Bhama, P.K.; Park, J.; Hadlock, C.R.; Hadlock, T.A. Clinician-graded electronic facial paralysis assessment: The eFACE. Plast. Reconstr. Surg. 2015, 136, 223e–230e. [Google Scholar] [CrossRef]
  33. Banks, C.A.; Jowett, N.; Hadlock, C.R.; Hadlock, T.A. Weighting of facial grading variables to disfigurement in facial palsy. JAMA Facial Plast. Surg. 2016, 18, 292–299. [Google Scholar] [CrossRef]
  34. Lyford-Pike, S.; Helwig, N.E.; Sohre, N.E.; Guy, S.J.; Hadlock, T.A. Predicting perceived disfigurement from facial function in patients with unilateral paralysis. Plast. Reconstr. Surg. 2018, 142, 722E–728E. [Google Scholar] [CrossRef]
  35. Banks, C.A.; Jowett, N.; Hadlock, T. Test-retest reliability and agreement between in-person and video assessment of facial mimetic function using the eFACE facial grading system. JAMA Facial Plast. Surg. 2017, 19, 206–211. [Google Scholar] [CrossRef]
  36. Malka, R.; Miller, M.; Guarin, D.; Fullerton, Z.; Hadlock, T.; Banks, C. Reliability Between In-Person and Still Photograph Assessment of Facial Function in Facial Paralysis Using the eFACE Facial Grading System. Facial Plast. Surg. Aesthetic Med. 2021, 23, 344–349. [Google Scholar] [CrossRef]
  37. Banks, C.A.; Jowett, N.; Azizzadeh, B.; Beurskens, C.; Bhama, P.; Borschel, G.; Coombs, C.; Coulson, S.; Croxon, G.; Diels, J.; et al. Worldwide testing of the eFACE facial nerve clinician-graded scale. Plast. Reconstr. Surg. 2017, 139, 491e–498e. [Google Scholar] [CrossRef] [PubMed]
  38. Mato-Patino, T.; Sanchez-Cuadrado, I.; Penarrocha, J.; Morales-Puebla, J.M.; Diez-Sebastián, J.; Gavilan, J.; Lassaletta, L. Validation of the Spanish version of the Electronic Facial Palsy Assessment (eFACE). Eur. Arch. Oto-Rhino-Laryngol. 2024, 281, 673–682. [Google Scholar] [CrossRef] [PubMed]
  39. Wiggins, J.S. A psychological taxonomy of trait-descriptive terms: The interpersonal domain. J. Personal. Soc. Psychol. 1979, 37, 395–412. [Google Scholar] [CrossRef]
  40. Wiggins, J.S.; Steiger, J.H.; Gaelick, L. Evaluating circumplexity in personality data. Multivar. Behav. Res. 1981, 16, 263–289. [Google Scholar]
  41. Wiggins, J.S.; Trapnell, P.; Phillips, N. Psychometric and geometric characteristics of the Revised Interpersonal Adjective Scales (IAS-R). Multivar. Behav. Res. 1988, 23, 517–530. [Google Scholar] [CrossRef]
  42. Knutson, B. Facial expressions of emotion influence interpersonal trait inferences. J. Nonverbal Behav. 1996, 20, 165–182. [Google Scholar] [CrossRef]
  43. Bates, D.; Mächler, M.; Bolker, B.; Walker, S. Fitting linear mixed-effects models using lme4. J. Stat. Softw. 2015, 67, 1–48. [Google Scholar] [CrossRef]
  44. Kuznetsova, A.; Brockhoff, P.B.; Christensen, R.H.B. lmerTest package: Tests in linear mixed effects models. J. Stat. Softw. 2017, 82, 1–26. [Google Scholar] [CrossRef]
  45. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2025. [Google Scholar]
  46. Helwig, N.E.; Sohre, N.E.; Ruprecht, M.R.; Guy, S.J.; Lyford-Pike, S. Dynamic properties of successful smiles. PLoS ONE 2017, 12, e0179708. [Google Scholar] [CrossRef] [PubMed]
  47. Krumhuber, E.G.; Manstead, A.S.; Cosker, D.; Marshall, D.; Rosin, P.L.; Kappas, A. Facial dynamics as indicators of trustworthiness and cooperative behavior. Emotion 2007, 7, 730–735. [Google Scholar] [CrossRef]
  48. Krumhuber, E.; Manstead, A.S.R.; Cosker, D.; Marshall, D.; Rosin, P.L. Effects of dynamic attributes of smiles in human and synthetic faces: A simulated job interview setting. J. Nonverbal Behav. 2009, 33, 1–15. [Google Scholar] [CrossRef]
  49. Krumhuber, E.G.; Kappas, A.; Manstead, A.S.R. Effects of dynamic aspects of facial expressions: A review. Emot. Rev. 2013, 5, 41–46. [Google Scholar] [CrossRef]
  50. Krumhuber, E.G.; Skora, L.I.; Hill, H.C.H.; Lander, K. The role of facial movements in emotion recognition. Nat. Rev. Psychol. 2023, 2, 283–296. [Google Scholar] [CrossRef]
  51. Henrich, J.; Heine, S.J.; Norenzayan, A. The weirdest people in the world? Behav. Brain Sci. 2010, 33, 61–83. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.