Next Article in Journal
A Review of Wearable Solutions for Physiological and Emotional Monitoring for Use by People with Autism Spectrum Disorder and Their Caregivers
Next Article in Special Issue
Lane Endpoint Detection and Position Accuracy Evaluation for Sensor Fusion-Based Vehicle Localization on Highways
Previous Article in Journal
Foreground Detection with Deeply Learned Multi-Scale Spatial-Temporal Features
Previous Article in Special Issue
Cloud Update of Tiled Evidential Occupancy Grid Maps for the Multi-Vehicle Mapping
Article Menu

Article Versions

Export Article

Open AccessArticle
Sensors 2018, 18(12), 4270;

Driver’s Facial Expression Recognition in Real-Time for Safe Driving

Department of Computer Engineering, Keimyung University, Daegu 42601, Korea
Author to whom correspondence should be addressed.
Received: 6 November 2018 / Revised: 27 November 2018 / Accepted: 3 December 2018 / Published: 4 December 2018
(This article belongs to the Special Issue Sensors Applications in Intelligent Vehicle)
PDF [1383 KB, uploaded 4 December 2018]


In recent years, researchers of deep neural networks (DNNs)-based facial expression recognition (FER) have reported results showing that these approaches overcome the limitations of conventional machine learning-based FER approaches. However, as DNN-based FER approaches require an excessive amount of memory and incur high processing costs, their application in various fields is very limited and depends on the hardware specifications. In this paper, we propose a fast FER algorithm for monitoring a driver’s emotions that is capable of operating in low specification devices installed in vehicles. For this purpose, a hierarchical weighted random forest (WRF) classifier that is trained based on the similarity of sample data, in order to improve its accuracy, is employed. In the first step, facial landmarks are detected from input images and geometric features are extracted, considering the spatial position between landmarks. These feature vectors are then implemented in the proposed hierarchical WRF classifier to classify facial expressions. Our method was evaluated experimentally using three databases, extended Cohn-Kanade database (CK+), MMI and the Keimyung University Facial Expression of Drivers (KMU-FED) database, and its performance was compared with that of state-of-the-art methods. The results show that our proposed method yields a performance similar to that of deep learning FER methods as 92.6% for CK+ and 76.7% for MMI, with a significantly reduced processing cost approximately 3731 times less than that of the DNN method. These results confirm that the proposed method is optimized for real-time embedded applications having limited computing resources.
Keywords: facial expression recognition; deep neural networks; embedded application; ADAS; weighted random forest facial expression recognition; deep neural networks; embedded application; ADAS; weighted random forest
This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited (CC BY 4.0).

Share & Cite This Article

MDPI and ACS Style

Jeong, M.; Ko, B.C. Driver’s Facial Expression Recognition in Real-Time for Safe Driving. Sensors 2018, 18, 4270.

Show more citation formats Show less citations formats

Note that from the first issue of 2016, MDPI journals use article numbers instead of page numbers. See further details here.

Related Articles

Article Metrics

Article Access Statistics



[Return to top]
Sensors EISSN 1424-8220 Published by MDPI AG, Basel, Switzerland RSS E-Mail Table of Contents Alert
Back to Top