Next Article in Journal
License Plate Detection Based on Improved YOLOv8n Network
Previous Article in Journal
Robotic Ultrasound Diagnostic System for Non-Destructive Testing in Highly Variable Production
Previous Article in Special Issue
Simultaneous Localization of Two Talkers Placed in an Area Surrounded by Asynchronous Six-Microphone Arrays
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The Role of Facial Action Units in Investigating Facial Movements During Speech

Biomedical Engineering Program, Center for Biomechanical & Rehabilitation Engineering, School of Engineering and Applied Sciences, University of the District of Columbia, 4200 Connecticut Ave. NW, Washington, DC 20008, USA
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(10), 2066; https://doi.org/10.3390/electronics14102066
Submission received: 25 March 2025 / Revised: 12 May 2025 / Accepted: 12 May 2025 / Published: 20 May 2025

Abstract

:
Investigating how facial movements can be used to characterize and quantify speech is important, in particular, to aid those suffering from motor control speech disorders. Here, we sought to investigate how facial action units (AUs), previously used to classify human expressions and emotion, could be used to quantify and understand unimpaired human speech. Fourteen (14) adult participants (30.1 ± 7.9 years old), fluent in English, with no speech impairments, were examined. Within each data collection session, 6 video trials per participant per phoneme were acquired (i.e., 102 trials total/phoneme). The participants were asked to vocalize the vowels /æ/, /ɛ/, /ɪ/, /ɒ/, and /ʊ/; the consonants /b/, /n/, /m/, /p/, /h/, /w/, and /d/; and the diphthongs /eI/, /ʌɪ/, /i/, /a:/, and /u:/. Through the use of Python Py-Feat, our analysis displayed the AU contributions for each phoneme. The important implication of our methodological findings is that AUs could be used to quantify speech in populations with no speech disability; this has the potential to be broadened toward providing feedback and characterization of speech changes and improvements in impaired populations. This would be of interest to persons with speech disorders, speech language pathologists, engineers, and physicians.

1. Introduction

In everyday life, the ability to speak and be understood is something many of us take for granted. However, motor control impairments tied to speech can be debilitating. Disorders wherein the muscles used for speaking are dysfunctional due to neurological damage lead to difficulties with articulation, pronunciation, and clarity of speech [1]. Such impairments are commonly categorized as either dysarthria (muscle weakness) or apraxia of speech (difficulty planning and coordinating muscle movements for speech production) [2]. In adults, these disorders can arise from a variety of causes (e.g., stroke, brain tumors, multiple sclerosis, Parkinson’s disease, amyotrophic lateral sclerosis (ALS), cerebral palsy, head trauma, or neurodegenerative diseases) [3]. In children, speech motor control impairments can arise from conditions such as childhood apraxia of speech (CAS) [4]. Childhood apraxia of speech and developmental dysarthria may co-occur with organic syndromes, such as Down Syndrome, Williams–Beuren Syndrome, and Cri-du-Chat Syndrome that also present auxiliary difficulties [5,6], or could occur independent of these.
Currently, how speech language pathologists (SLPs) assess speech disorders is through subjective yet comprehensive assessment, case history of the subject and their family, performing an oral mechanism examination, hearing screening, and speech sound assessment. During the evaluation of speech, the accuracy of speech production, speech sound errors, and error patterns are observed by the SLP [7]. Intelligibility assessments are administered to help pathologists determine the severity of speech difficulty using subjective rating scales [8]. Speech is also evaluated with the use of objective measures and indices, such as the percentage of consonants correct and whole-word matches [9]. However, this approach results in several limitations: metrics such as the type and frequency of speech sound errors are usually quantified using rating scales; the speaker’s articulation approach and linguistic factors are challenging to interpret by the SLP; the listener’s (SLP’s) familiarity with the speaker’s (patient’s) speech pattern may vary; communication cues for the person being assessed may be misinterpreted; and the presence of background noise may affect the results of an intelligibility assessment [10]. Stimulability tests are administered to determine how a child articulates sounds in various contexts [11]. However, a major limiting factor includes variations in the way the test is administered (by different clinicians), leading to inconsistencies in the results. Speech perception tests allow the SLP/clinician to determine the patient’s ability to perceive the difference between the standard production. Limitations of this method include speech presentations that may not reflect how the subject speaks in everyday situations and the difficulty of use in younger children (i.e., misleading results due to the child’s misunderstanding and/or misinterpretation of the procedural requests).
For all of the above, a qualitative judgment, as opposed to a quantitative indication, is made by the SLP on how the speech sound disorder impacts functional communication. While the above clinical tests administered by the SLPs have proven to be valuable, there are several other limitations to consider [12]. These limitations include subjective bias and interpretation of the SLP, cultural and linguistic biases, lack of contextual relevance, and an overreliance on and confidence in traditional assessments, thus leading to potential for misdiagnosis [13]. Further, assessments must be tailored for children who are too young to be evaluated using the above, adults or children who are either reluctant or unable, and adults and children with less intelligible speech (i.e., more severe speech deficits) [14]. Furthermore, it is known that children have relatively smaller attention spans while undergoing instruction and training [15]. Attention is considered very critical in cognitive learning [16]. Some individuals with speech disorders may also have other exacerbating conditions and may therefore have difficulty either performing or comprehending traditionally used, standardized tests; thus, they would be unable to provide the requested responses. Alternative tests that are objective, as well as applicable across a wider range of individuals with speech deficits, require further inspection and development [17].
The studies below use vastly different approaches, and different goals than our study, which utilized facial action units (AUs) (linked to facial muscle groups), to examine the production of speech (phonemes). An audio–visual speech recognition method previously utilized side-face images of the lips, which are captured while the audio is recorded using a microphone. Lip contour geometric features (LCGFs), used for discriminating phonemes, and lip movement velocity features (LMVFs), used for detecting voice activity, were used to extract both audio and visual features. These features were used to assess speech using Hidden Markov Models (HMMs) [18]. Another study designed and compared two HMM classifiers for four emotions which have varying effects on the properties of different phonemes. The emotional states during speech were classified using phoneme-level modeling [19]. Phonetic feature-based prediction models were used to predict and understand variations in the pronunciation of phonemes. This allowed for better modeling of spontaneous speech by focusing on individual articulatory features [20]. Another approach focused on identifying the relationship between speech sounds and facial movements using a fine-grained statistical correlation analysis between various phonemes and facial anthropometric measurements (AMs) [21]. Another alternative to phoneme analysis using facial movements is the viseme model. Viseme maps are particularly useful in supporting accurate visual feature extraction for the visual representation of speech [22,23].
On the other hand, techniques like electropalatography (EPG) or palatography involve embedding an electrode to the palate to monitor tongue movements. Though noninvasive and shown to be effective in children and adults [24], these have not been proven to be reliable for vowels and glides where vocal tract constrictions have greater contribution [25]. Also, EPG therapy may not be as advantageous, especially for younger children who have not shown a resistance to traditional methods. It is important to note that many facial expressions are innate. In other words, facial expressions do not vary between a person blind from birth and an individual with progressive loss of sight [26,27]. Intuitively, the perception of speech, especially in a noisy environment, is better due to visual details [28]. Furthermore, the speech voice features independent of vocal tract characteristics will need other reliable estimators to develop speech–face mapping [29,30]. This is shown through some unique facial traits identifiers with acoustic models. This establishes a substantial use of facial expression for improved learning ability by visual and sound cues without discarding that Visemes are a subset of phonemes, and work related to Visemes will further enhance the presented investigation [31]. Thus, these findings establish the important impact of the presented AU acoustic-based assessment feedback approach [32,33,34].
In our research, we explored the potential for objective quantification of speech sounds involving measurement of facial muscle groups’ movements during speech via facial action units (AUs). We hypothesized that phonemes could be assessed and quantified via AU mapping. The facial action coding system (FACS) previously was developed primarily for the analysis and distinction of facial expressions toward human emotions [35]; however, here, we aimed to utilize AUs to quantify facial movements in relation to speech. Automated facial analysis includes facial detection and tracking, feature extraction, and classification, where the facial regions categorized by action units are detected from image frames extracted from videos. This approach has potential to allow for an objective, customized assessment for individual speech based on personal features. Our goal was to determine AU responses to define a baseline for normal speech (for unimpaired participants). This information can later be contrasted with impaired speech and used to provide feedback for speech training.

2. Materials and Methods

All study activities were conducted within the Center for Biomechanical and Rehabilitation Engineering (CBRE) laboratory at the University of the District of Columbia (Washington, DC, USA). The protocol was approved by the Institutional Review Board (2248324-1), and all participants gave their informed consent prior to participating in the study.

2.1. Participants

Participants learned about the study through flyers posted around the university and word of mouth. The population targeted were adults (18–45 years old) who were fluent, American English speakers without speech impairments; here, we did not examine non-English-speaking participants. Participants provided their informed consent prior to taking part in the research study. Fourteen participants were enrolled in the study and their demographics are shown in Table 1. After participants gave their informed consent, their data collection session was scheduled.

2.2. Data Collection

The experimental setup involved the use of the following equipment: a desktop computer, connected to an external microphone and speakers, as well as an additional desktop computer for the researcher to control the camera application while recording the participants as they articulate the sounds. Prior to the session, the research assistants ensured that the experimental setup was fully functional. The setup process involved turning on the desktop computers, recording a test video to confirm the functionality of the desktop’s camera application, and ensuring the video quality was sufficient. The desktop’s inbuilt camera was used to facilitate the position and circumvent the need of installing drivers for portable cameras in different systems. The objective was to develop a portable audio- and video-capturing system. Figure 1 shows the experimental setup used for recording the data.
A solid-colored background was placed behind the participant’s chair and adjusted to ensure it was the only visible background in the videos (Figure 1). The participant’s chair was positioned so that only the participant’s head, neck, and upper shoulders were visible in the recording on the computer. Priority was given to the visibility of the participant’s face; therefore, participants were asked to remove hats, jewelry, glasses, bulky clothing, or any other items that could interfere with facial visibility. The distance between the chair and the background was adjusted to accommodate variations in participant height. The recording desktop screen was controlled by the desktop computer manned by the researcher. To minimize background noise, additional movements in the lab were restricted during the recording sessions. The desktop’s camera and microphone and an additional microphone (for improving the sound quality of the participants’ voice) were used to record each participant while vocalizing phonemes (vowels, phonemes, and diphthongs). The recorded video format is MP4, with a bit depth of 8, frame rate of 29.8 frames per second, and resolution of 1920 × 1080. The audio format is MP3 with 2 channels, 1024 samples/frame, and a sampling rate of 44,100 Hz.
During the session, the researcher sat at the opposite side of the table, facing a desktop that allowed full control of the recording process. This setup ensured that participants only needed to speak when cued. Participants were prompted to look directly into the desktop’s camera lens for the duration of the recording. They were informed that a beep would indicate the start of the recording, after which the participant would produce the designated phoneme sound six times. Before recordings began, the experimenter provided instructions and asked the participant if any steps were unclear. With the test setup in place, video recording was ready to proceed.
Once the video recording was started, the participant proceeded with vocalizing each phoneme. The following phonemes were repeated 6 times per video recording, with breaks between recordings: /æ/, /ɛ/, /ɪ/, /ɑ/, /ɒ/, /b/, /n/, /m/, /p/, /h/, /w/, /d/, /eɪ/, /ʌɪ/, /i/, /a:/, and /u:/. The participant produced the sound six times with an estimated one-second interval between repetitions. After the recording for each phoneme ended, the participant was given a brief break before proceeding with the next recording.
At the end of the session, the participants’ videos (17 videos per participant, with 6 repetitions per phoneme) were stored in a separate folder on the computer. Participants were informed that their speech analysis results would be provided near the end of the study and were encouraged to reach out with any questions. In the presented data collection, ensuring sufficient natural or white light exposure was strictly considered and guaranteeing maximum coverage of the head shot in the field of view (FOV) was emphasized. For acoustic sound recording, any ambient noise was restricted by enclosing the recording room. A Logitech Snowball Microphone was used for its better directive features. However, no detailed study on the negative impact of reverberation and background noise has been presented here. Some apparent noises were edited out using Audacity (https://www.audacityteam.org/) and other speech processing software.

2.3. Data Processing

Once the videos were preprocessed to ensure each trial was separated, the Py-Feat: Python Facial Expression Analysis Toolbox (Python, software version 3.9) was used to determine action units (AUs), or facial muscle groups, used for each phoneme and word presentation. AUs are numbered muscle groups of the face, used to produce particular facial expressions—in this case, facial movements (Figure 2). A total of 20 AUs were generated with 11 specifically involved in speech production (e.g., facial muscle groups surrounding the mouth, cheeks, and jaws) (Table 2).
The videos were separated into individual trials and processed as separate videos; each video trial was converted into frames. All video preparations were carried out using the scripting language, Python. This facilitated time-intensive steps of converting recorded videos into a programmable format. The platform for analyzing and processing the vocalization videos was IDLE (integrated development environment for Python), using the PyFeat Facial Expression Recognition (FER) module [36]. The outputs generated from processing the videos are frames saved as jpeg. files with the corresponding contour maps for each frame. This allowed for visualization of the AUs representing the activated facial muscles. The summary of AUs and averages across frames were saved as csv. (comma-separated value) files for each of the recorded phonemes.
The results were pooled across participants for each phoneme to generate a boxplot for each AU. Within the boxplots, median, upper quartile range, lower quartile range, and outliers were determined for each phoneme as a function of AU. Boxplots were created to show the distribution in mean AU amplitude across all trials for each sound. Action units with the least variation in mean AU amplitude and a higher median indicated the most activated facial muscles specific to these sound types across all participants. A benchmark was developed to establish the range of mean AU amplitude for the AUs representing facial muscles involved in unimpaired speech production. This range varies depending on the phoneme.

3. Results

Boxplots show the distribution of mean AU amplitude for the different action units across all the trials produced by all the participants. The results for the vowels (Figure 3), consonants (Figure 4), and diphthongs (Figure 5) are shown.
Table 3 shows the median values of AU amplitude for each phoneme.

4. Discussion

Within the results, movement of the articulators, as reflected in the activity of specific facial muscle groups, was observed. More specifically, our study analyzed the median amplitude and distribution of action units (AUs) associated with facial muscle groups during the articulation of various speech sounds, including vowels, phonemes, and diphthongs. The articulators (tongue, lips, and jaw) play a critical role in speech production and are primarily controlled by muscles in the face. The findings are presented through boxplots displaying the median for the AU amplitude values of each articulated sound, providing a visual representation of the variation in AU amplitudes for different sounds and allowing for an objective assessment of the muscle groups most engaged during speech production. The median values across the different AUs indicate the central tendency of muscle activation in speech articulation across participants. Table 3 shows the range of AU activation, where the most activated AUs are represented by the darkest shades of blue and the least activated AUs are represented by the lightest shades of blue.
The analysis of vowel sounds reveals distinct activation patterns across the 11 facial AUs contributing to speech production. Notably, the AU12 (Nasolabial deepener) showed the highest activation in vowels /æ/, /ɛ/, and /ʊ/, with the exception of AU17 (Lip corner depressor) which showed the highest activation for vowel /ɪ/, and vowel /ɒ/ shows the highest activations of muscle groups represented by AUs 09 (Nose wrinkler), 12 (Nasolabial deepener), 14 (Lip corner puller), 15 (Dimpler), 17 (Lip corner depressor), 23 (Lip stretcher), 24 (Lip tightener), and 25 (Lip pressor). It was more challenging to isolate the most activated action unit based on the median values for consonants displayed in Table 3. The most activated AU for consonants /b/ and /p/ was AU12 (Nasolabial deepener). The remaining consonants, /n/, /m/, /h/, /w/, and /d/ showed the highest activation in muscle groups represented by AUs 09 (Nose wrinkler), 12 (Nasolabial deepener), 14 (Lip corner puller), 15 (Dimpler), 17 (Lip corner depressor), 23 (Lip stretcher), 24 (Lip tightener), and 25 (Lip pressor). Given that these consonants are not characterized by a single most activated AU, it is then recommended to pay close attention to these muscle groups during the articulation of these sounds. It is also important to note that the boxplots shown for AUs of particular interest in speech are mostly skewed, indicating the outliers and scope for increasing the sample size when developing AI and machine learning models for feedback interfaces. The latter has not been discussed in this paper and is mentioned here for its potential application [38]. The presented work is an exposition for developing AU-based modeling of normal speech of sounds and words assuming the sample size, gender, age, and other variability are secondary as opposed to primary factors in facial muscle activation. Hence, the accuracy and confidence of generalizing the model has not been stated and will be of greater importance in studies following the presented work.
AU12 (Nasolabial deepener) also showed the highest activation during the articulation of diphthongs /ʌɪ/, /i/, and /a:/. On the other hand, the AUs with the highest activation were 09, 12, 14, 15, 17, 24, and 25 for diphthong /eI/ and AUs 09, 12, 14, 15, 17, 23, 24, and 25. Identifying the AUs that are most activated during unimpaired speech could potentially help speech therapists focus on the muscle groups represented by these AUs, allowing them to customize their treatment approaches to meet individual patient needs.
While preliminary, these results highlight aspects that could potentially be used to develop robust measures for improving and training individuals with impaired speech. The observed differences in AU activation patterns across different sound types can inform speech therapy and rehabilitation techniques for individuals with speech impairments. These insights can also be applied to the development of facial recognition and speech synthesis technologies.
A similar approach was taken in a previous study for the application of facial action units in the analysis and prediction of stuttering speech [39]. Facial AUs were then extracted from video recordings of participants and calculated from participants’ faces using a model trained with various datasets. The AUs were then categorized based on the upper and lower facial regions. The facial movements represented by the different AUs were recorded and processed as inputs for a deep learning model used to distinguish between fluent and stuttered speech. A distinction was made between the AUs that contributed to a higher likelihood of stuttering and those that contributed the least or not at all to the prediction of stuttering. It was shown that facial AUs contributed significantly to distinguishing between fluent and stuttered speech while also providing insights into the non-verbal cues related to stuttering. Limitations of this study included variations in facial structures and expressions among individuals, which may have made it challenging to generalize the results across different individuals; the limited resolution of the recordings used to capture changes in facial muscle movements; the differences in the study setup, which could cause variations in results when applied to real-time stuttering; and the possibility that some AUs were miscategorized or excluded from the model, potentially leading to misunderstandings of the contributions of AUs in this study.
Future research could further categorize the data for a more detailed analysis of individual variations in speech production, as well as the impact of external factors such as speech rate and emotional expression. Our study provides insight into how facial muscle activation varies across speech sounds, emphasizing the role of the articulators in speech production. This serves as a baseline toward the application of this method to characterize the speech of individuals (children and adults) with speech and language disorders. We aim to increase the sample size of participants and the time allocated for data collection in future studies.

Author Contributions

Conceptualization, A.A.N., L.A.T., A.B., and N.A.; funding acquisition, L.A.T.; methodology, A.A.N., A.B., and N.A.; project administration, L.A.T.; software, A.A.N., A.B., N.A., and C.K.III; supervision, L.A.T.; visualization, A.A.N. and L.A.T.; writing—original draft, A.A.N. and L.A.T.; writing—review and editing, A.A.N., L.A.T., A.B., N.A., and C.K.III. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the NSF Waterman Award Grant (2229575). The NIH NIA MSTEM (1R25AG067896) provided funding support for some of the scholars on this project.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board for studies involving humans (Protocol #2248324-1).

Informed Consent Statement

Informed consent was obtained from all participants involved in the present study.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

The first author would like to express her sincere gratitude to her research advisor, Lara A. Thompson, for her invaluable guidance and support throughout this study. We also wish to thank the other contributing authors for their essential roles in the planning, execution, and development of this paper. Special thanks to the study participants for their valuable contributions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Menon, E.B.; Ravichandran, S.; Tan, E.S. Speech disorders in closed head injury patients. Singap. Med. J. 1993, 34, 45–48. [Google Scholar]
  2. A Tool for Differential Diagnosis of Childhood Apraxia of Speech and Dysarthria in Children: A Tutorial. Available online: https://pubs.asha.org/doi/epdf/10.1044/2022_LSHSS-21-00164 (accessed on 24 March 2025).
  3. Speech Disorders: Types, Symptoms, Causes, and Treatment. Available online: https://www.medicalnewstoday.com/articles/324764 (accessed on 20 March 2025).
  4. What Is Apraxia of Speech?|NIDCD. Available online: https://www.nidcd.nih.gov/health/apraxia-speech (accessed on 20 March 2025).
  5. Björelius, H.; Tükel, Ş.; Björelius, H.; Tükel, Ş. Comorbidity of Motor and Sensory Functions in Childhood Motor Speech Disorders. In Advances in Speech-Language Pathology; IntechOpen: London, UK, 2017. [Google Scholar] [CrossRef]
  6. Steingass, K.J.; Chicoine, B.; McGuire, D.; Roizen, N.J. Developmental Disabilities Grown Up: Down Syndrome. J. Dev. Behav. Pediatr. 2011, 32, 548–558. [Google Scholar] [CrossRef] [PubMed]
  7. American Speech-Language-Hearing Association. Speech Sound Disorders: Articulation and Phonology. Available online: https://www.asha.org/practice-portal/clinical-topics/articulation-and-phonology/ (accessed on 12 March 2025).
  8. Berisha, V.; Utianski, R.; Liss, J. Towards a Clinical Tool for Automatic Intelligibility Assessment. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 2825–2828. [Google Scholar] [CrossRef]
  9. van Doornik, A.; Welbie, M.; McLeod, S.; Gerrits, E.; Terband, H. Speech and language therapists’ insights into severity of speech sound disorders in children for developing the speech sound disorder severity construct. Int. J. Lang. Commun. Disord. 2025, 60, e70022. [Google Scholar] [CrossRef]
  10. Tyler, A.A.; Tolbert, L.C. Speech-Language Assessment in the Clinical Setting. Am. J. Speech-Lang. Pathol. 2002, 11, 215–220. [Google Scholar] [CrossRef]
  11. Rvachew, S. Stimulability and Treatment Success. Top. Lang. Disord. 2005, 25, 207–219. [Google Scholar] [CrossRef]
  12. Usha, G.P.; Alex, J.S.R. Speech assessment tool methods for speech impaired children: A systematic literature review on the state-of-the-art in Speech impairment analysis. Multimed. Tools Appl. 2023, 82, 35021–35058. [Google Scholar] [CrossRef]
  13. Selin, C.M.; Rice, M.L.; Girolamo, T.; Wang, C.J. Speech-Language Pathologists’ Clinical Decision Making for Children with Specific Language Impairment. Lang. Speech Hear. Serv. Sch. 2019, 50, 283–307. [Google Scholar] [CrossRef]
  14. Speech and Language Assessment—Pediatric (Standardized Tests). Available online: https://www.voxlingue.com/post/speech-and-language-assessment-pediatric-standardized-tests (accessed on 24 March 2025).
  15. Kannass, K.N.; Colombo, J.; Wyss, N. Now, Pay Attention! The Effects of Instruction on Children’s Attention. J. Cogn. Dev. 2010, 11, 509–532. [Google Scholar] [CrossRef]
  16. Draheim, C.; Pak, R.; Draheim, A.A.; Engle, R.W. The role of attention control in complex real-world tasks. Psychon. Bull. Rev. 2022, 29, 1143–1197. [Google Scholar] [CrossRef]
  17. Hamm, J.; Kohler, C.G.; Gur, R.C.; Verma, R. Automated Facial Action Coding System for Dynamic Analysis of Facial Expressions in Neuropsychiatric Disorders. J. Neurosci. Methods 2011, 200, 237–256. [Google Scholar] [CrossRef]
  18. Iwano, K.; Yoshinaga, T.; Tamura, S.; Furui, S. Audio-Visual Speech Recognition Using Lip Information Extracted from Side-Face Images. EURASIP J. Audio Speech Music Process. 2007, 2007, 64506. [Google Scholar] [CrossRef]
  19. Lee, C.M.; Yildirim, S.; Bulut, M.; Kazemzadeh, A.; Busso, C.; Deng, Z.; Lee, S.; Narayanan, S.S. Emotion recognition based on phoneme classes. In Proceedings of the Interspeech 2004, ISCA, Jeju, Republic of Korea, 4–8 October 2004; pp. 889–892. [Google Scholar] [CrossRef]
  20. Bates, R.A.; Ostendorf, M.; Wright, R.A. Symbolic phonetic features for modeling of pronunciation variation. Speech Commun. 2007, 49, 83–97. [Google Scholar] [CrossRef]
  21. Qu, L.; Zou, X.; Li, X.; Wen, Y.; Singh, R.; Raj, B. The Hidden Dance of Phonemes and Visage: Unveiling the Enigmatic Link between Phonemes and Facial Features. arXiv 2023, arXiv:2307.13953. [Google Scholar] [CrossRef]
  22. Yau, W.C.; Kumar, D.K.; Arjunan, S.P. Visual recognition of speech consonants using facial movement features. Integr. Comput. Eng. 2007, 14, 49–61. [Google Scholar] [CrossRef]
  23. Cappelletta, L.; Harte, N. Phoneme-To-Viseme Mapping for Visual Speech Recognition. In Proceedings of the International Conference on Pattern Recognition Applications and Methods, SCITEPRESS, Vilamoura, Portugal, 6–8 February 2012; pp. 322–329. [Google Scholar] [CrossRef]
  24. Pickett, K.L. The Effectiveness of Using Electropalatography to Remediate a Developmental Speech Sound Disorder in a School-Aged Child with Hearing Impairment; Brigham Young University: Provo, Utah, 2013. [Google Scholar]
  25. McAuliffe, M.J.; Ward, E.C. The use of electropalatography in the assessment and treatment of acquired motor speech disorders in adults: Current knowledge and future directions. NeuroRehabilitation 2006, 21, 189–203. [Google Scholar] [CrossRef]
  26. Matsumoto, D.; Willingham, B. Spontaneous Facial Expressions of Emotion of Congenitally and Noncongenitally Blind Individuals. J. Personal. Soc. Psychol. 2009, 96, 1. [Google Scholar] [CrossRef]
  27. Valente, D.; Theurel, A.; Gentaz, E. The role of visual experience in the production of emotional facial expressions by blind people: A review. Psychon. Bull. Rev. 2018, 25, 483–497. [Google Scholar] [CrossRef]
  28. Yu, Y.; Lado, A.; Zhang, Y.; Magnotti, J.F.; Beauchamp, M.S. The Effect on Speech-in-Noise Perception of Real Faces and Synthetic Faces Generated with either Deep Neural Networks or the Facial Action Coding System. bioRxiv 2024. [Google Scholar] [CrossRef]
  29. Garg, S.; Hamarneh, G.; Jongman, A.; Sereno, J.A.; Wang, Y. Computer-vision analysis reveals facial movements made during Mandarin tone production align with pitch trajectories. Speech Commun. 2019, 113, 47–62. [Google Scholar] [CrossRef]
  30. Garg, S.; Hamarneh, G.; Jongman, A.; Sereno, J.A.; Wang, Y. ADFAC: Automatic detection of facial articulatory features. MethodsX 2020, 7, 101006. [Google Scholar] [CrossRef]
  31. Ma, J.; Cole, R. Animating visible speech and facial expressions. Vis. Comput. 2004, 20, 86–105. [Google Scholar] [CrossRef]
  32. Fasel, B.; Luettin, J. Automatic facial expression analysis: A survey. Pattern Recognit. 2003, 36, 259–275. [Google Scholar] [CrossRef]
  33. Esfandbod, A.; Rokhi, Z.; Meghdari, A.F.; Taheri, A.; Alemi, M.; Karimi, M. Utilizing an Emotional Robot Capable of Lip-Syncing in Robot-Assisted Speech Therapy Sessions for Children with Language Disorders. Int. J. Soc. Robot. 2023, 15, 165–183. [Google Scholar] [CrossRef]
  34. Schipor, O.-A.; Pentiuc, S.-G.; Schipor, M.-D. Towards a multimodal emotion recognition framework to be integrated in a Computer Based Speech Therapy System. In Proceedings of the 2011 6th Conference on Speech Technology and Human-Computer Dialogue (SpeD), Brasov, Romania, 18–21 May 2011; pp. 1–6. [Google Scholar] [CrossRef]
  35. Clark, E.A.; Kessinger, J.; Duncan, S.E.; Bell, M.A.; Lahne, J.; Gallagher, D.L.; O’Keefe, S.F. The Facial Action Coding System for Characterization of Human Affective Response to Consumer Product-Based Stimuli: A Systematic Review. Front. Psychol. 2020, 11, 920. [Google Scholar] [CrossRef]
  36. Cheong, J.H.; Jolly, E.; Xie, T.; Byrne, S.; Kenney, M.; Chang, L.J. Py-Feat: Python Facial Expression Analysis Toolbox. Affect. Sci. 2023, 4, 781–796. [Google Scholar] [CrossRef]
  37. Parent, R.; King, S.; Fujimura, O. Issues with lip sync animation: Can you read my lips? In Proceedings of the Computer Animation 2002 (CA 2002), Geneva, Switzerland, 21 June 2002; pp. 3–10. [Google Scholar] [CrossRef]
  38. Wynn, A.T.; Wang, J.; Umezawa, K.; Cristea, A.I. An AI-Based Feedback Visualisation System for Speech Training. In Artificial Intelligence in Education. Posters and Late Breaking Results, Workshops and Tutorials, Industry and Innovation Tracks, Practitioners’ and Doctoral Consortium; Rodrigo, M.M., Matsuda, N., Cristea, A.I., Dimitrova, V., Eds.; Springer International Publishing: Cham, Switzerland, 2022; pp. 510–514. [Google Scholar] [CrossRef]
  39. Das, A.; Mock, J.; Chacon, H.; Irani, F.; Golob, E.; Najafirad, P. Stuttering Speech Disfluency Prediction using Explainable Attribution Vectors of Facial Muscle Movements. arXiv 2020, arXiv:2010.01231. [Google Scholar] [CrossRef]
Figure 1. Example of AU data collection session with a participant: (a) experimental setup for data collection (desktop computer with researcher; participant facing camera and microphone); (b) researcher controlling camera application and recording speech audio; (c) participant vocalizing during the trial.
Figure 1. Example of AU data collection session with a participant: (a) experimental setup for data collection (desktop computer with researcher; participant facing camera and microphone); (b) researcher controlling camera application and recording speech audio; (c) participant vocalizing during the trial.
Electronics 14 02066 g001
Figure 2. Mapping and numbering of the facial action units (AUs) [36].
Figure 2. Mapping and numbering of the facial action units (AUs) [36].
Electronics 14 02066 g002
Figure 3. Vowel AU amplitude boxplots (median, upper and lower interquartile range (IQR), and standard deviations pooled across participants) for (a) /æ/, (b) /ε/, (c) /ɪ/, (d) /ɒ/, and (e) /ʊ/). Blue boxes indicate AUs specifically related to speech, while gray boxes represent AUs not associated with speech.
Figure 3. Vowel AU amplitude boxplots (median, upper and lower interquartile range (IQR), and standard deviations pooled across participants) for (a) /æ/, (b) /ε/, (c) /ɪ/, (d) /ɒ/, and (e) /ʊ/). Blue boxes indicate AUs specifically related to speech, while gray boxes represent AUs not associated with speech.
Electronics 14 02066 g003
Figure 4. Consonant AU amplitude boxplots (median, upper and lower interquartile range (IQR), and standard deviations pooled across participants) for (a) /b/, (b) /n/, (c) /m/, (d) /p/, (e) /h/, (f) /w/, and (g) /d/. Blue boxes indicate AUs specifically related to speech, while gray boxes represent AUs not associated with speech.
Figure 4. Consonant AU amplitude boxplots (median, upper and lower interquartile range (IQR), and standard deviations pooled across participants) for (a) /b/, (b) /n/, (c) /m/, (d) /p/, (e) /h/, (f) /w/, and (g) /d/. Blue boxes indicate AUs specifically related to speech, while gray boxes represent AUs not associated with speech.
Electronics 14 02066 g004
Figure 5. Diphthong AU amplitude boxplots (median, upper and lower interquartile range (IQR), and standard deviations pooled across participants) for (a) /eI/, (b) /ʌɪ/, (c) /i/, (d) /a:/, and (e) /u:/. Blue boxes indicate AUs specifically related to speech, while gray boxes represent AUs not associated with speech.
Figure 5. Diphthong AU amplitude boxplots (median, upper and lower interquartile range (IQR), and standard deviations pooled across participants) for (a) /eI/, (b) /ʌɪ/, (c) /i/, (d) /a:/, and (e) /u:/. Blue boxes indicate AUs specifically related to speech, while gray boxes represent AUs not associated with speech.
Electronics 14 02066 g005
Table 1. Overview of participant demographics.
Table 1. Overview of participant demographics.
ParticipantGenderAgeEthnicityTraumatic Brain/Stroke InjuryCognitive ImpairmentsDevelopment DelaysLearning Disability and/or Speech Impairment
S1F26Black/African AmericanNoNoNoNo
S2F46Black/African AmericanNoNoNoNo
S3F38Black/African AmericanNoNoNoNo
S4M26Asian/Pacific IslanderNoNoNoNo
S5M21Black/African AmericanNoNoNoNo
S6F41Asian/Pacific IslanderNoNoNoNo
S7M25Black/African AmericanNoNoNoNo
S8F29Black/African AmericanNoNoNoNo
S9M25Black/African AmericanNoNoNoADHD, Graphomotor Disorder
S10M23Black/African AmericanNoNoNoNo
S11F26Black/African AmericanNoNoNoNo
S12M28Hispanic/LatinoNoNoNoNo
S13F43Black/African AmericanNoNoNoNo
S14F24Black/African AmericanNoNoNoNo
Table 2. Action units (AUs) specifically used for speech production [37].
Table 2. Action units (AUs) specifically used for speech production [37].
Action UnitsMuscle GroupsDescription of Muscle Group Function
AU09Levator Labii Superioris Alaeque NasiNose wrinkler
AU11Zygomaticus MinorNasolabial deepener
AU12Zygomaticus MajorLip corner puller
AU14BuccinatorDimpler
AU15Depressor Anguli OrisLip corner depressor
AU17MentalisChin raiser
AU20Risorius, PlatysmaLip stretcher
AU23Orbicularis OrisLip tightener
AU24Orbicularis OrisLip pressor
AU25Depressor Labii InferiorisLip part
AU26Masseter, Temporalis, Medial PterygoidJaw drop [36,37]
Table 3. Median values of AU amplitude for facial action units specific to articulated sounds; red text indicates those AUs specific to speech.
Table 3. Median values of AU amplitude for facial action units specific to articulated sounds; red text indicates those AUs specific to speech.
VowelsConsonantsDiphthongs
Range of AUs/æ//ε//ɪ//ɒ//ʊ//b//n//m//p//h//w//d//eI//ʌɪ//i//a://u:/
AU010.3750.3710.3680.350.350.350.380.350.390.340.370.420.370.370.370.370.37
AU020.2500.2390.250.260.250.250.290.270.260.260.250.260.250.260.270.250.27
AU040.240.2640.270.250.230.260.240.240.250.250.250.30.250.280.250.240.25
AU050.30.2970.630.300.300.300.310.310.30.300.310.310.300.300.300.310.30
AU060.590.5920.630.540.60.590.550.610.590.530.560.530.530.580.570.550.49
AU070.450.520.420.280.380.460.40.50.470.50.340.350.430.450.550.450.34
AU090.4250.4250.120.400.40.390.430.430.40.40.440.450.440.430.440.410.41
AU100.1750.130.660.140.090.120.110.110.130.130.080.130.10.110.10.130.1
AU110.0750.1650.50.1250.210.150.10.210.30.210.240.130.230.240.210.280.23
AU120.6950.6910.470.520.610.620.540.580.660.580.50.470.460.620.610.630.5
AU140.4750.4770.540.430.460.440.460.480.460.470.430.460.450.480.460.440.43
AU150.450.450.180.450.450.510.490.510.460.50.450.510.420.450.390.450.48
AU170.550.530.660.540.540.540.520.580.540.550.530.550.530.520.520.530.53
AU200.210.250.490.220.170.090.110.120.180.180.180.160.180.20.190.210.17
AU230.390.380.460.410.390.430.430.410.410.420.460.420.380.390.390.40.42
AU240.50.490.530.510.490.530.490.560.530.530.540.550.530.530.460.480.44
AU250.4750.480.510.450.370.40.580.390.480.50.420.410.50.540.480.560.44
AU260.2750.2430.270.250.250.250.220.230.270.260.250.220.260.240.290.290.27
AU280.190.240.190.180.160.140.130.210.20.190.20.20.140.170.160.130.14
AU430.180.4170.390.380.380.390.40.410.320.360.390.350.40.390.390.40.38
Legend for Median Ranges
0–0.20.21–0.390.4–0.590.6–0.790.8–1.0
For the table displayed above, the red text indicates the AUs that are specifically related to speech while the black and white texts indicate the AUs that are not associated with speech production.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Newby, A.A.; Bhatta, A.; Kirkland, C., III; Arnold, N.; Thompson, L.A. The Role of Facial Action Units in Investigating Facial Movements During Speech. Electronics 2025, 14, 2066. https://doi.org/10.3390/electronics14102066

AMA Style

Newby AA, Bhatta A, Kirkland C III, Arnold N, Thompson LA. The Role of Facial Action Units in Investigating Facial Movements During Speech. Electronics. 2025; 14(10):2066. https://doi.org/10.3390/electronics14102066

Chicago/Turabian Style

Newby, Aliya A., Ambika Bhatta, Charles Kirkland, III, Nicole Arnold, and Lara A. Thompson. 2025. "The Role of Facial Action Units in Investigating Facial Movements During Speech" Electronics 14, no. 10: 2066. https://doi.org/10.3390/electronics14102066

APA Style

Newby, A. A., Bhatta, A., Kirkland, C., III, Arnold, N., & Thompson, L. A. (2025). The Role of Facial Action Units in Investigating Facial Movements During Speech. Electronics, 14(10), 2066. https://doi.org/10.3390/electronics14102066

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop