Next Article in Journal
Fast Linde–Buzo–Gray (FLBG) Algorithm for Image Compression through Rescaling Using Bilinear Interpolation
Previous Article in Journal
Imaging-Based Deep Learning for Predicting Desmoid Tumor Progression
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Head Gesture Recognition Combining Activity Detection and Dynamic Time Warping

1
College of Building Environmental Engineering, Zhengzhou University of Light Industry, Zhengzhou 450001, China
2
Engineering Training Center, Zhengzhou University of Light Industry, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
J. Imaging 2024, 10(5), 123; https://doi.org/10.3390/jimaging10050123
Submission received: 26 March 2024 / Revised: 11 May 2024 / Accepted: 13 May 2024 / Published: 19 May 2024
(This article belongs to the Section Computer Vision and Pattern Recognition)

Abstract

:
The recognition of head movements plays an important role in human–computer interface domains. The data collected with image sensors or inertial measurement unit (IMU) sensors are often used for identifying these types of actions. Compared with image processing methods, a recognition system using an IMU sensor has obvious advantages in terms of complexity, processing speed, and cost. In this paper, an IMU sensor is used to collect head movement data on the legs of glasses, and a new approach for recognizing head movements is proposed by combining activity detection and dynamic time warping (DTW). The activity detection of the time series of head movements is essentially based on the different characteristics exhibited by actions and noises. The DTW method estimates the warp path distances between the time series of the actions and the templates by warping under the time axis. Then, the types of head movements are determined by the minimum of these distances. The results show that a 100% accuracy was achieved in the task of classifying six types of head movements. This method provides a new option for head gesture recognition in current human–computer interfaces.

1. Introduction

Enabling machines to understand the meaning behind human behavior has become one of the research topics that has aroused great interest among researchers. In particular, head gesture recognition plays an important role in human–computer interfaces. Currently, the main methods for detecting head movements include image processing, depth sensing, radar technology, and inertial sensors [1]. Head movement detection has a wide range of applications in various fields, such as human–computer interactions [2,3,4], assisted driving [5,6,7], virtual reality [8,9,10], augmented reality [11,12], and more.
Compared to image processing methods based on video and depth maps, an inertial measurement unit (IMU) sensor-based detection system has lower complexity, a faster processing speed, and a lower cost. Many researchers have studied algorithms for identifying head movement types using IMU sensors. To address the human–computer interaction problems of patients with quadriplegia, Rudigkeit et al. [4] achieved a maximum action recognition rate of 85.8% with an IMU sensor-based behavior-assisted robot controlled by head movements. Wang et al. [3] proposed using a simple nod of the head as a method for unlocking, in order to address the inconvenience caused by the reliance on external input devices for password authentication during the use of AR/VR devices. They selected initial-end postures as well as gesture dynamics, including the angular velocity magnitude, trajectory properties, and the power spectral density of the linear acceleration magnitude, as features to train a random forest classifier, achieving an accuracy rate of 97.1%. Severin et al. [13] conducted a series of studies on head movement classification using a 6DOF inertial sensor consisting of three-directional accelerations and angular velocities. Firstly, they compared various methods, such as random the forest classifier, the k-Neighbors Classifier (kNN), the Decision Tree Classifier, Gaussian Naive Bayes (NB), Logistic Regression, the Extra Trees Classifier, a Quadratic Discrimination Analysis, the Adaptive Boosting Classifier, and a Support Vector Machine (SVM), and achieved the best classification accuracy of 99.48%. Analyzing the impact of features on classification results, it was found that using 16 features represented by the dominant frequency, the max value of the time series, the min value of the time series, the time series length, the interquartile range, the median value, the mean value, autocorrelation, standard deviations, Kurtosis, autoregressive coefficients, energy, the Skew value, variance, Spectral Entropy, and Sample Entropy had better classification results than using six features. The former achieved a complete and correct classification except for the kNN algorithm [14]. When analyzing the impact of the number of sensors on classification results, it was found that a system with three sensors can achieve better classification results [15]. However, in an experiment, the three sensors were installed in completely different locations compared to the single sensor, which makes it difficult to draw a significant comparison between the two. Wong et al. [16] developed a smart helmet for motorcycle riders using an IMU sensor. They used a linear discriminant analysis and a neural network to recognize head movements with an accuracy of 99.1% and found that gyroscope sensor data are more suitable for head movement recognition than accelerometer data. It can be seen that significant progress has been made in the accuracy of head movement recognition, but high accuracy requires relatively more head movement time series features. The goal of this paper is to find a method that can still achieve completely correct classification results with fewer features.
Dynamic Time Warping (DTW) measures the similarity between two time sequences, which might be obtained by sampling a source with varying sampling rates or by recording the same phenomenon occurring with varying speeds. It has a wide range of applications in speech recognition and action recognition [17,18,19,20,21]. However, there are few studies on the application of DTW in head gesture recognition. Mavus et al. [22] were the first to study this and achieved an accuracy of 85.68% with only four types of head movement classifications, possibly due to the use of fewer features. Then, Hachaj et al. [23,24] trained a DTW barycenter averaging algorithm using the quaternion averaging obtained from IMU sensors and a bagged variation of the previous method that utilizes many DTW classifiers that perform voting. The classification results achieved an accuracy of 97.5%. This algorithm requires the average of a quaternion-based signal and the estimation of head movement trajectories for recognition, and it does not take into account the issue of endpoint detection for head movements.
Compared with previous studies, this paper achieves better classification performance by minimizing the use of feature numbers and reducing data processing. Furthermore, an activity detection algorithm for head movements is proposed to automatically extract time sequences of head motions instead of manual segmentations, which enables the adoption of this algorithm in devices such as VR/AR, wheelchairs, and assistive robots to help people improve their gaming experience or quality of life.

2. Related Work

2.1. Applications of Head Motion Recognition

Head motion recognition has been studied extensively over the years, and research results have been successfully applied in many domains. Rudigkeit et al. [4,25] developed a behavior-assistive robot that uses head gestures as a control interface. They found that the mapping of head motions onto robot motions is intuitive and that the given feedback is useful, enabling smooth, precise, and efficient robot control. Furthermore, Dey et al. [26] developed a smart wheelchair using IMU-based head motion recognition, enabling quadriplegic patients to easily navigate a wheelchair using only their head. Dobrea et al. [27] also successfully used a four-capacitive-sensor system to recognize head motions for wheelchair control. As AR/VR devices become increasingly prevalent in our lives, Wang et al. [3] developed a seamless user-authentication technique for AR/VR users based on biometric features extracted from IMU-sensor readings in the devices, using simple head motions. To identify the head movements of motorcycle riders, Wong et al. [16] also used IMU sensors and a neural network to recognize four types of actions (Looking Up, Looking Down, Turning Left, and Turning Right). They found that the head movements were concentrated on rotations, and the gyroscope data had greater importance than the accelerometer data. To facilitate smart watch users to accept or reject operations, Yang et al. [28] developed a head shaking and nodding recognition system based on millimeter-wave sensors. These research findings have provided possibilities to facilitate people’s work and life, especially for those with disabilities who cannot communicate through their limbs. This further demonstrates the important significance of in-depth research on head movement recognition.

2.2. Methods for Head Motion Recognition

Head motions carry a wealth of information that can effectively support and assist human communication. Ashish et al. [29] used an infrared-sensitive camera equipped with infrared LEDs to track aa pupil and used the pupil movements as observations in a discrete Hidden Markov Model (HMM)-based pattern analyzer to detect head nods and shakes. Their system achieved a real-time recognition accuracy of 78.46% on the test dataset. Fujie et al. [30] used the optical flow over the head region as a feature and modeled them using HMM for recognition. They used the optical flow of the body area to compensate for swayed images, as it was challenging to collect sufficient data to train the probabilistic models. They introduced Maximum Likelihood Linear Regression (MLLR) to adapt the original model to camera movements. Wei et al. [31] employed Microsoft Kinect and utilized discrete HMMs as the backbone for a machine learning-based classifier. They trained three HMMs (nodHMM, shakeHMM, and otherHMM) based on five head movement states (Up, Down, Left, Right, and Still), achieving an 86% accuracy on test datasets. Wang et al. [3] reconstructed head motion trajectories using IMU data from AR/VR devices and extracted features such as initial-end postures, skeletal proportions, and gesture dynamics for a random forest classifier in N-label classification tasks, achieving an average accuracy of 97.1%. Mavus et al. [22] proposed a genetic algorithm-based cost function minimization, but did not specify the exact IMU data used, achieving an accuracy of 85.68%. Hachaj et al. [24] trained a DTW barycenter averaging algorithm using the quaternion averaging obtained from IMU sensors, as well as a bagged variation of the previous method that utilizes multiple DTW classifiers with voting. Their classification results achieved an accuracy of 97.5%. Severin et al. [13,14,15] tested some classification methods based on IMU data and found that when using 16 features from the head motion time series, most classifiers can achieve a completely accurate classification. However, these studies did not cover the endpoint detection methods for head motion time series data, and endpoint detection is a necessary step for the real-time segmentation of head motion time series data.
This paper proposes a method for the automatic detection of head motion endpoints to extract time series data during head movements. Subsequently, using a small number of features, only involving the acceleration and angular velocity from an IMU, it analyzes the time series features of six common head motion types and creates templates for their motion time series. Then, based on the templates and the extracted head motion time series, the type of head motion is determined by calculating the shortest path using the DTW method, which has significant advantages in assessing the similarity of time series.

3. Material and Methods

3.1. Data Acquisition

The present study utilized inertial sensors to perceive changes in head posture and collected six types of feature data, including acceleration and angular velocity in the X, Y, and Z directions. The inertial sensor used was the MPU9250, with a sampling frequency of fs = 100 Hz, and was installed on the eyeglass temple close to the front, as shown in the middle view of Figure 1. During data collection, the participants wore glasses equipped with IMU sensors and sat quietly on a stool while naturally performing six different head movements: nodding (Rotation down around the z-axis and recovery to the normal position along the previous motion trajectory, as shown in the bottom view of Figure 1); tilting up (Rotation up around the z-axis and recovery to the normal position along the previous motion trajectory, as shown in the top view of Figure 1); shaking left (Rotation to the left around the x-axis and recovery to the normal position along the previous motion trajectory, as shown in the bottom left view of Figure 1); shaking right (Rotation to the right around the x-axis and recovery to the normal position along the previous motion trajectory, as shown in the bottom right view of Figure 1); tilting left (Rotation to the left around the y-axis and recovery to the normal position along the previous motion trajectory, as shown in the top left view of Figure 1); and tilting right (Rotation to the right around the y-axis and recovery to the normal position along the previous motion trajectory, as shown in the top right view of Figure 1), as shown in the outer view of Figure 1.
The data in this experiment was self-collected using an IMU sensor module with Bluetooth communication. All 25 participants were students from the Zhengzhou University of Light Industry. A total of 25 × 6 datasets were collected and automatically labeled using the head movement activity detection algorithm (detailed in Section 2.2). The accuracy of these labels was verified manually. A total of 2333 data points were labeled, and they were randomly divided into a training set of 1914 and a testing set of 419. The distribution of each type of head gesture is shown in Table 1.
The format of the dataset is [data, label], where data is a 6-dimensional matrix consisting of the acceleration and angular velocity data of the x, y, and z-axis collected from the IMU sensor, and the length of the data is variable under the different labels. Label is a categorical variable that corresponds to the 6 types of head movements.

3.2. Activity Detection for Head Gestures

Head pose endpoint detection is essentially distinguished by the different features exhibited during head movements. Based on the acquired data, as shown in Figure 2, it can be observed that the head exhibits similar total angular velocity waveforms regardless of the performed action. Therefore, this paper utilizes this characteristic for the endpoint detection of head actions. The specific detection method is as follows:
(1)
Data normalization
y ( t ) = arctan ( x ( t ) ) × 2 / π
where y ( t ) is the normalized data, and x ( t ) is the collected acceleration and angular velocity data from the sensor.
(2)
Sliding median filtering
y ( t ) = { median ( x ( t ( l 1 ) / 2 : t + ( l 1 ) / 2 ) ) , l = 2 n 1 , n N median ( x ( t l / 2 : t + l / 2 1 ) ) , l = 2 n , n N
where l is the length of the filter; l = 2 n 1 , n N represents an odd number, and y ( t ) is the median; l = 2 n , n N represents an even number, and y ( t ) is the average of the two middle values. The purpose of median filtering is to reduce the salt-and-pepper noise of the sensor and to reduce the possibility of misjudgment in subsequent action recognitions and endpoint detections.
(3)
Determining the start time of head movements
if a n g ( t ) > a n g m i n , t s t a r t = t
where a n g ( t ) = a n g x 2 ( t ) + a n g y 2 ( t ) + a n g z 2 ( t ) is the overall description of the angular velocity variation, reflecting the overall degree of change in the angle of head movements. a n g x ( t )   a n g y ( t )  and a n g z ( t ) represent the angular velocity components on the three-dimensional coordinate axes. a n g m i n is the threshold for detecting the start of head movements. t s t a r t is the moment when a head movement starts.
(4)
Determining the end time of head movements
if   ( a n g ( t ) < a n g m i n ) A N D ( a n g ( t t i n t e r v a l ) < a n g m i n ) , t e n d = t
where t i n t e r v a l is a threshold used to determine if a movement has truly ended. Please refer to Figure 2 for details. When the first peak of the head movement ends, the movement is halfway completed. Therefore, it is necessary for the threshold to cross this trough in order to capture the entire waveform of the head movement. Additionally, this threshold also serves as the minimum interval between two consecutive head movements, and t e n d represents the moment when the head movement ends.
(5)
Determining the validity of head movements
if   ( t e n d t s t a r t > t m i n ) A N D ( t e n d t s t a r t > t m a x ) ,   There   is   a   head   movement   present .
where t m i n is the small duration of the movement used to filter out sharp peak noise in the waveform, and t m a x is the maximum duration of the movement used to exclude movements with abnormal or incomplete durations.

3.3. Head Gesture Recognition Using the DTW Method

The DTW method is a dynamic programming-based algorithm widely used in the fields of speech and posture recognition. This algorithm can warp data on the time axis, stretching or shortening time series to achieve better alignment and improve the accuracy and robustness of the algorithm. Head movements, due to personal habits and different current states, may result in changes in the length of the movement, making it a typical recognition problem of unequal-length time series. Therefore, this paper uses the DTW method to evaluate the warping path distance between different movements and templates to recognize different head movements. The recognition algorithm is as follows:
(1)
Calculating time series templates for head movements
To perform action recognition based on the DTW method, a reference template is needed. In order to make this recognition method real-time, an endpoint detection algorithm proposed in this paper is used to extract the time series of the movement. The threshold values used in the algorithm are as follows: twin = 0.2 s, angmin = 0.2 rad/s, tmin = 0.5 s, and tmax = 2.5 s. The effect of other threshold values will be discussed in the experimental analysis. After obtaining the time series for each movement, to effectively evaluate the performance of the algorithm, only the training set is used to create the templates. Taking one movement as an example, suppose a set of data in its time series is S a = { s 1 , s 2 , , s a } , where a is the length of this set of data. Then, the total time series set of the training set is S = { S a , S b , , S k n } T , where n is the number of this movement in the training set; a, b,..., k represent the lengths of each sequence; and let the sequence length matrix be S l e n = { a ,   b , , k } . Therefore, the standard template for this movement is T = { t 1 , t 2 , t 3 , , t m } , where, t i = 1 l j = 1 l S j i , l represents the number of occurrences of S j i when i is fixed. If S j i does not exist, it is not counted. m represents the median of the S l e n set. By following this process, the standard time series T i j for other head movements can be obtained, where i = 1, 2,..., 6 represent six head movements; and j = 1, 2,..., 6 represent six types of time series, including acceleration and angular velocity.
(2)
Calculating Euclidean distance matrix
Due to the general inability to align standard templates with the time series of the head movements to be recognized, it is clearly unreliable to calculate similarity using linear alignment. In this case, it is necessary to warp a set of time series to align feature points, so that specific feature points are aligned with each other. The specific method is as follows: Let a time series of head movements in the test set be S = { s 1 , s 2 , , s n } and the time series of the template to be matched be T = { t 1 , t 2 , t 3 , , t m } . Then, the Euclidean distance between any two points is calculated as d ( s i , t j ) = k = 0 5 ( s i k t j k ) 2 , forming a n × m distance matrix D. Therefore, the similarity problem between these two time series is transformed into applying dynamic programming to solve the shortest path problem from the starting point D(1,1) to the ending point D(n,m), which is the warping path, denoted by W.
(3)
Finding the warping path
The form of the warping path can be represented as follows: W = { w 1 , w 2 , w 3 , , w k } , where w i represents the distance between points S and T, k is the length of the warping path, with a of range max ( m , n ) k m + n , and it satisfies the following constraints:
(1) Boundary condition: the path goes from the starting point w 1 = D ( 1 , 1 ) to the ending point w k = D ( n , m ) .
(2) Continuity: if w i - 1 = D ( a , b ) , then the next point w i = D ( a , b ) on the path needs to satisfy | a a | 1 , | b b | 1 , meaning it cannot skip over a certain point for matching.
(3) Monotonicity: if w i - 1 = D ( a , b ) , then the next point w i = D ( a , b ) on the path needs to satisfy a a 0 , b b 0 , meaning that for the points on W , the path must monotonically progress over time.
Based on this, we know that there are only three possible paths from w i - 1 = D ( a , b ) to the next point: D ( a + 1 , b ) , D ( a + 1 , b + 1 ) , and D ( a , b + 1 ) . Therefore, the optimal warping path is as follows:
DTW ( S , T ) = min { 1 k i = 1 k w i }
(4)
Solving the optima warping path
The dynamic programming idea of the cumulative distance is used to calculate Equation (6). The formula for the cumulative distance is defined as follows:
r ( i , j ) = d ( s i , t i ) + min { r ( i 1 , j ) , r ( i 1 , j 1 ) , r ( i , j 1 ) }
where i = 1 , 2 , 3 , , n and j = 1 , 2 , 3 , , m . The cumulative distance is actually a kind of recursive relationship, so the best warping path distance between the two time series S , T is DTW ( S , T ) = r ( n , m ) . This solves the problem of measuring the similarity of time series with different lengths and unaligned feature positions.
(5)
Determining the type of head movement
To separately calculate the DTW values between the head movement time series and the time series of the 6 template movements, repeat the above steps, and the minimum value will then represent the type of this movement.

4. Experimental Results and Discussion

4.1. Experimental Platform

This experiment’s hardware environment uses an Intel i7-8700 processor with 16 GB of memory, a P2000 graphics card with 5 GB of memory, Bluetooth 4.0 communication, and a sensor chip MPU9250 with a sampling frequency of 100 Hz.

4.2. Selection of Endpoint Detection Parameters

Based on the head movement template in Figure 2, it can be observed that head movements generally exhibit a back-and-forth motion, which is manifested as two peaks in the angular velocity angmin. In order to correctly identify these two peaks, it is necessary to determine the threshold of the angular velocity. At a minimum duration of head movement tmin = 0.5 s, the threshold of the angular velocity was tested in the range of [0.01, 0.42] with an increment of 0.01, and it was found that correct endpoint detection can be achieved within the range of [0.06, 0.32] of the angular velocity; hence, angmin = 0.2 was selected, as shown in Figure 3a. Subsequently, at angmin = 0.2 rad/s, the minimum duration of head movement was tested in the range of [0.05, 1] with an increment of 0.05, and it was found that correct endpoint detection can be achieved within the range of [0.4, 0.7]; hence, tmin = 0.6 s was selected, as shown in Figure 3b. Finally, the influence of the median filtering length on the results was examined, taking angmin = 0.2 rad/s and tmin = 0.6 s, resulting in twin = 0.3 s, as shown in Figure 3c.

4.3. An Analysis of the Recognition Effectiveness of the DTW Algorithm

To verify the accuracy of the DTW algorithm for head movement recognition, head movement templates were created using a training set, and the effectiveness of the algorithm was tested using a test set. The DTW algorithm was used to calculate the DTW values corresponding to the shortest path, reflecting the similarity of the two curves, i.e., the quality of the match. The better the match between the two curves, the higher the similarity and the shorter the shortest warping path, resulting in a smaller DTW value. Conversely, the worse the match between the two curves, the less similar the matching curves and the longer the shortest warping path, resulting in a larger DTW value. Ideally, when two matching curves are completely aligned, the shortest warping path is 0, meaning the DTW value is 0. In order to distinguish between these six types of head movement characteristics, this study, respectively, weighed the DTW values of the six curves, with the smallest DTW value corresponding to the specific head movement. The results show a recognition accuracy of 100%.
To demonstrate the effectiveness of this method, this paper uses an LSTM network as a control algorithm. Convolutional neural networks, as representatives of feedforward neural networks, have shown significant performance in classifying fixed-length spatial data, such as images and videos. However, for the processing of sequential data related to human pose detections, the effect is not too ideal due to the constraints of its own network structure. The LSTM network belongs to the recurrent neural network and has memory, which can model the temporal sequence of data and has a good fitting effect on sequential data. In the field of deep learning, traditional feedforward neural networks perform well in many aspects. Therefore, this paper adopts an LSTM neural network, with detailed parameters, as shown in Table 2. Specifically, the initialization method of batch size weights is very important for training neural network models. In this paper, the orthogonal initialization is chosen to ensure that the initial weights are orthogonal, in order to solve the problems of vanishing gradients and exploding gradients that can occur during the training process of an LSTM network. The training and test data are consistent with the DTW method, and after verification, the accuracy is only 86.31%.
A further analysis revealed that although the LSTM method has a significantly lower recognition rate compared to the DTW method, its recognition speed is higher than the DTW method. This is because the DTW method requires calculating the similarity of six sequences and comparing their magnitudes. However, the training time for an LSTM is much higher than that of the DTW method.

5. Conclusions

This paper presents a head movement detection method based on DTW. It collects feature data of head movement postures, including the acceleration and angular velocity in the X, Y, and Z directions, through an IMU fixed on the glasses. The method extracts the motion interval of head movements using endpoint detection and constructs a head movement template. Subsequently, it calculates the optimal path between the head movement data and the head movement template data, where the minimum value corresponds to the standard template of the head movement type, thus determining the head movement type of the test data. Compared to previous studies, the method proposed in this paper uses endpoint detection to automatically segment the head motion time series and accurately recognize the head motion types of the subjects using six features based on DTW. This provides a new option for human–computer interface technology in the field of head gesture recognition. Additionally, it is cost-effective, requires minimal data processing, provides quick responses, and achieves high recognition accuracy.

Author Contributions

Conceptualization, H.L. and H.H.; methodology, H.L; software, H.L.; validation, H.L. and H.H.; formal analysis, H.L.; investigation, H.L.; resources, H.L.; data curation, H.L.; writing—original draft preparation, H.L.; writing—review and editing, H.L.; visualization, H.L.; supervision, H.L.; project administration, H.L.; funding acquisition, H.L. and H.H.; All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Science and Technology Research Project of Henan Province (232102321021 and 232102211050) and the Research Fund of the Zhengzhou University of Light Industry (2018BSJJ053).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ionut-Cristian, S.; Dan-Marius, D. Using Inertial Sensors to Determine Head Motion—A Review. J. Imaging 2021, 7, 265. [Google Scholar] [CrossRef] [PubMed]
  2. Liu, H.; Li, D.; Wang, X.; Liu, L.; Zhang, Z.; Subramanian, S. Precise head pose estimation on HPD5A database for attention recognition based on convolutional neural network in human-computer interaction. Infrared Phys. Technol. 2021, 116, 103740. [Google Scholar] [CrossRef]
  3. Wang, X.; Zhang, Y. Nod to Auth: Fluent AR/VR Authentication with User Head-Neck Modeling. In Proceedings of the 2021 Conference on Human Factors in Computing Systems, Virtual, 8–13 May 2021. [Google Scholar]
  4. Rudigkeit, N.; Gebhard, M. AMiCUS—A Head Motion-Based Interface for Control of an Assistive Robot. Sensors 2019, 19, 2836. [Google Scholar] [CrossRef] [PubMed]
  5. Ju, J.; Zheng, H.; Li, C.; Li, X.; Liu, H.; Liu, T. AGCNNs: Attention-guided convolutional neural networks for infrared head pose estimation in assisted driving system. Infrared Phys. Technol. 2022, 123, 104146. [Google Scholar] [CrossRef]
  6. Li, Z.; Fu, Y.; Yuan, J.; Huang, T.S.; Wu, Y. Query Driven Localized Linear Discriminant Models for Head Pose Estimation. In Proceedings of the 2007 IEEE International Conference on Multimedia and Expo, Beijing, China, 2–5 July 2007. [Google Scholar]
  7. Mbouna, R.O.; Kong, S.G.; Chun, M.G. Visual Analysis of Eye State and Head Pose for Driver Alertness Monitoring. IEEE Trans. Intell. Transp. Syst. 2013, 14, 1462–1469. [Google Scholar] [CrossRef]
  8. Koniaris, B.; Huerta, I.; Kosek, M.; Darragh, K.; Malleson, C.; Jamrozy, J.; Swafford, N.; Guitian, J.; Moon, B.; Israr, A.; et al. IRIDiuM: Immersive rendered interactive deep media. In Proceedings of the 43rd Annual Conference on Computer Graphics and Interactive Techniques, SIGGRAPH 2016, Anaheim, CA, USA, 24–28 July 2016; ACM Special Interest Group on Computer Graphics and Interactive Techniques (SIGGRAPH): Los Angeles, CA, USA, 2016. [Google Scholar]
  9. Xu, Y.; Jung, C.; Chang, Y. Head Pose Estimation Using Deep Neural Networks and 3D Point Cloud. Pattern Recognit. 2021, 121, 108210. [Google Scholar] [CrossRef]
  10. Cordea, M.D.; Petriu, D.C.; Petriu, E.M.; Georganas, N.D.; Whalen, T.E. 3-D head pose recovery for interactive virtual reality avatars. IEEE Trans. Instrum. Meas. 2002, 51, 640–644. [Google Scholar] [CrossRef]
  11. Hoff, W.; Vincent, T. Analysis of head pose accuracy in augmented reality. IEEE Trans. Vis. Comput. Graph. 2000, 6, 319–334. [Google Scholar] [CrossRef]
  12. Murphy-Chutorian, E.; Trivedi, M.M. Head Pose Estimation and Augmented Reality Tracking: An Integrated System and Evaluation for Monitoring Driver Awareness. IEEE Trans. Intell. Transp. Syst. 2010, 11, 300–311. [Google Scholar] [CrossRef]
  13. Severin, I.C.; Dobrea, D.M.; Dobrea, M.C. Head Gesture Recognition using a 6DOF Inertial IMU. Int. J. Comput. Commun. Control. 2020, 15, 3856. [Google Scholar] [CrossRef]
  14. Severin, I.C. Time Series Feature Extraction For Head Gesture Recognition: Considerations Toward HCI Applications. In Proceedings of the 24th International Conference on System Theory, Control and Computing (ICSTCC), Sinaia, Romania, 8–10 October 2020; pp. 232–237. [Google Scholar]
  15. Severin, I.C. Head Gesture-Based on IMU Sensors: A Performance Comparison Between the Unimodal and Multimodal Approach. In Proceedings of the 2021 International Symposium on Signals, Circuits and Systems (ISSCS), Daegu, Republic of Korea, 15–16 July 2021; pp. 1–4. [Google Scholar]
  16. Wong, K.I.; Chen, Y.-C.; Lee, T.-C.; Wang, S.-M. Head Motion Recognition Using a Smart Helmet for Motorcycle Riders. In Proceedings of the 2019 International Conference on Machine Learning and Cybernetics (ICMLC), Kobe, Japan, 7–10 July 2019; pp. 1–7. [Google Scholar]
  17. Obaid, M.; Hodrob, R.; Abu Mwais, A.; Aldababsa, M. Small vocabulary isolated-word automatic speech recognition for single-word commands in Arabic spoken. Soft Comput. 2023, 1–14. [Google Scholar] [CrossRef]
  18. Shafie, N.; Azizan, A.; Adam, M.Z.; Abas, H.; Yusof, Y.M.; Ahmad, N.A. Dynamic Time Warping Features Extraction Design for Quranic Syllable-based Harakaat Assessment. Int. J. Adv. Comput. Sci. Appl. 2022, 13, 48–54. [Google Scholar] [CrossRef]
  19. Xu, J.; Wang, H.; Zhang, J.; Cai, L. Robust Hand Gesture Recognition Based on RGB-D Data for Natural Human-Computer Interaction. IEEE Access 2022, 10, 54549–54562. [Google Scholar] [CrossRef]
  20. Tuncer, E.; Unlu, M.Z. Handwriting recognition by derivative dynamic time warping methodology via sensor-based gesture recognition. Maejo Int. J. Sci. Technol. 2022, 16, 72–88. [Google Scholar]
  21. Peixoto, J.S.; Cukla, A.R.; Cuadros, M.A.; Welfer, D.; Gamarra, D.F. Gesture Recognition using FastDTW and Deep Learning Methods in the MSRC-12 and the NTU RGB plus D Databases. IEEE Lat. Am. Trans. 2022, 20, 2189–2195. [Google Scholar] [CrossRef]
  22. Mavuş, U.; Sezer, V. Head Gesture Recognition via Dynamic Time Warping and Threshold Optimization. In Proceedings of the 2017 IEEE Conference on Cognitive and Computational Aspects of Situation Management (CogSIMA), Savannah, GA, USA, 27–31 March 2017. [Google Scholar]
  23. Hachaj, T.; Piekarczyk, M. Evaluation of Pattern Recognition Methods for Head Gesture-Based Interface of a Virtual Reality Helmet Equipped with a Single IMU Sensor. Sensors 2019, 19, 5408. [Google Scholar] [CrossRef] [PubMed]
  24. Hachaj, T.; Ogiela, M.R. Head Motion Classification for Single-Accelerometer Virtual Reality Hardware. In Proceedings of the 2019 5th International Conference on Frontiers of Signal Processing (ICFSP), Marseille, France, 18–20 September 2019; pp. 45–49. [Google Scholar]
  25. Rudigkeit, N.; Gebhard, M. AMiCUS 2.0-System Presentation and Demonstration of Adaptability to Personal Needs by the Example of an Individual with Progressed Multiple Sclerosis. Sensors 2020, 20, 1194. [Google Scholar] [CrossRef] [PubMed]
  26. Dey, P.; Hasan, M.M.; Mostofa, S.; Rana, A.I. Smart wheelchair integrating head gesture navigation. In Proceedings of the 2019 International Conference on Robotics, Electrical and Signal Processing Techniques (ICREST), Dhaka, Bangladesh, 10–12 January 2019; pp. 329–334. [Google Scholar]
  27. Dobrea, M.-C.; Dobrea, D.-M.; Severin, I.-C. A new wearable system for head gesture recognition designed to control an intelligent wheelchair. In Proceedings of the 2019 E-Health and Bioengineering Conference (EHB), Iasi, Romania, 21–23 November 2019; pp. 1–5. [Google Scholar]
  28. Yang, X.; Wang, X.; Dong, G.; Yan, Z.; Srivastava, M.; Hayashi, E.; Zhang, Y. Headar: Sensing Head Gestures for Confirmation Dialogs on Smartwatches with Wearable Millimeter-Wave Radar. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2023, 7, 138. [Google Scholar] [CrossRef]
  29. Kapoor, A.; Rosalind, W.P. A real-time head nod and shake detector. In Proceedings of the 2001 Workshop on Perceptive User Interfaces, Orlando, FL, USA, 15–16 November 2001; pp. 1–5. [Google Scholar]
  30. Fujie, S.; Ejiri, Y.; Nakajima, K.; Matsusaka, Y.; Kobayashi, T. A conversation robot using head gesture recognition as para-linguistic information. In Proceedings of the RO-MAN 2004, 13th IEEE International Workshop on Robot and Human Interactive Communication (IEEE Catalog No.04TH8759), Kurashiki, Japan, 22 September 2004; pp. 159–164. [Google Scholar]
  31. Wei, H.; Scanlon, P.; Li, Y.; Monaghan, D.S.; O’Connor, N.E. Real-time head nod and shake detection for continuous human affect recognition. In Proceedings of the 2013 14th International Workshop on Image Analysis for Multimedia Interactive Services (WIAMIS), Paris, France, 3–5 July 2013; pp. 1–4. [Google Scholar]
Figure 1. A schematic diagram of the head movement types.
Figure 1. A schematic diagram of the head movement types.
Jimaging 10 00123 g001
Figure 2. The time series templates for the head movements.
Figure 2. The time series templates for the head movements.
Jimaging 10 00123 g002
Figure 3. The impact of parameters on endpoint detection. (a) The impact of angular velocity threshold on the number of head motion recognition. (b) The impact of minimum head motion duration on the number of head motion recognition. (c) The impact of filtering time window length on the number of head motion recognition. In these figures, the areas filled with red inside the circles represent the thresholds that can correctly recognize the number of head movements, otherwise it is inaccurate.
Figure 3. The impact of parameters on endpoint detection. (a) The impact of angular velocity threshold on the number of head motion recognition. (b) The impact of minimum head motion duration on the number of head motion recognition. (c) The impact of filtering time window length on the number of head motion recognition. In these figures, the areas filled with red inside the circles represent the thresholds that can correctly recognize the number of head movements, otherwise it is inaccurate.
Jimaging 10 00123 g003
Table 1. Head gesture classification dataset.
Table 1. Head gesture classification dataset.
Head Movement TypesNumber of Training Samples Number of Testing Samples
Nodding29367
Tilting up31769
Shaking left32664
Shaking right30271
Tilting left33780
Tilting right33968
Total1914419
Table 2. The parameters of the LSTM neural network.
Table 2. The parameters of the LSTM neural network.
ParametersValueParametersValue
Hidden layer2Loss functionCross entropy
Number of hidden layers, 1 unit200Learning rate0.0001
Number of hidden layers, 2 units100Batch size128
Maximum number of iterations1000The initialization method of batch size weightsOrthogonal initialization
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, H.; Hu, H. Head Gesture Recognition Combining Activity Detection and Dynamic Time Warping. J. Imaging 2024, 10, 123. https://doi.org/10.3390/jimaging10050123

AMA Style

Li H, Hu H. Head Gesture Recognition Combining Activity Detection and Dynamic Time Warping. Journal of Imaging. 2024; 10(5):123. https://doi.org/10.3390/jimaging10050123

Chicago/Turabian Style

Li, Huaizhou, and Haiyan Hu. 2024. "Head Gesture Recognition Combining Activity Detection and Dynamic Time Warping" Journal of Imaging 10, no. 5: 123. https://doi.org/10.3390/jimaging10050123

APA Style

Li, H., & Hu, H. (2024). Head Gesture Recognition Combining Activity Detection and Dynamic Time Warping. Journal of Imaging, 10(5), 123. https://doi.org/10.3390/jimaging10050123

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop