Next Article in Journal
Frequency Domain Panoramic Imaging Algorithm for Ground-Based ArcSAR
Previous Article in Journal
Using a Stochastic Agent Model to Optimize Performance in Divergent Interest Tacit Coordination Games
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Non-Contact Heart Rate Detection When Face Information Is Missing during Online Learning

1
Faculty of Information Technology, Beijing University of Technology, Beijing 100224, China
2
College of Continuing Education, Beijing University of Technology, Beijing 100224, China
*
Authors to whom correspondence should be addressed.
Sensors 2020, 20(24), 7021; https://doi.org/10.3390/s20247021
Submission received: 26 October 2020 / Revised: 4 December 2020 / Accepted: 5 December 2020 / Published: 8 December 2020
(This article belongs to the Section Biomedical Sensors)

Abstract

:
Research shows that physiological signals can provide objective data support for the analysis of human emotions. At present, non-contact heart rate data have been employed in the research of medicine, intelligent transportation, smart education, etc. However, it is hard to detect heart rate data using non-contact traditional methods during head rotation, especially when face information is missing in scenarios such as online teaching/learning. Traditional remote photoplethysmography (rPPG) methods require a static, full frontal face within a fixed distance for heart rate detection. These strict requirements make it impractical to measure heart rate data in real-world scenarios, as a lot of videos only partially record the subjects’ face information, such as profile, too small distance, and wearing a mask. The current algorithm aims to solve the problem of head deflections between 30 degrees and 45 degrees by employing a symmetry substitution method, which can replace the undetected region of interest (ROI) with the detectable one. When face information is partially missing, our algorithm uses face–eye location to determine ROI. The results show that the method in this paper can solve certain practical problems related to heart rate detection, with a root mean square error (RMSE) under 7.64 bpm.

1. Introduction

With increasing interest in non-intelligent factors in the learning process in the field of education research, people are beginning to pay attention to the emotional problems in education and teaching, especially those associated with students’ learning. As an important non-intelligence factor, emotion has a multi-dimensional and all-round influence on students’ growth and development [1]. Thus, expressing emotions through data plays an important role in intelligent education. At present, emotion recognition is mainly realized through facial expression recognition. Although the method is simple and does not need special equipment, it cannot guarantee accurate emotion recognition. On one hand, people can disguise their true feelings by affecting their facial expressions, which is often difficult to detect [2]. On the other, there are a lot of emotions that are not clearly conveyed through facial expressions. Furthermore, different personalities lead to different abilities in terms of expressing various emotions. Therefore, emotional states need to be represented in a higher dimensional space.
Research shows that physiological signals can provide objective data support for emotion recognition [3]. Thus, a relationship can be established between emotional recognition and heart rate. Bland et al. [4] evaluated the heart rate responses of students answering a series of questions related to the theory of evolution and found that students would produce perceptible emotional responses in the form of physiological changes when they were exposed to the theme of evolution. Stavroulia et al. [5] used heart rate data to evaluate the emotional changes of subjects in certain scenarios and virtual reality experiences. They found that the heart rate increased when the subjects’ moods changed significantly. In addition, when Nik et al. [6] designed an online learning guidance system that can detect and reflect learners’ emotional states, heart rate variability was selected to measure learners’ emotional changes.
Contact heart rate estimation equipment, such as electrocardiogram (ECG) or conventional photoplethysmography (PPG), requires electrodes, gel, and experienced nurses or doctors, and usually causes skin irritation and discomfort. Moreover, it is not convenient for students to wear special instruments to measure heart rate in online classes. However, detecting heart rate through non-contact methods does not affect the students’ daily life.
Camera-based remote photoplethysmography (rPPG) enables low-cost, non-contact cardiovascular activity testing. Online teaching is often carried out on computers or mobile phones. These electronic devices with cameras can be used to collect students’ physiological signals.
Video-based physiological signal detection has been in development for more than 10 years. The first attempt to remotely measure heart rate with ambient light was by Verkruysse et al. [7] in 2008, and measuring heart rate remains a challenge today. We applied a non-contact heart rate measurement method to an online teaching scenario; however, using the ECG and PPG methods to measure students’ heart rates in class has certain limitations in terms of scope of ease of use, cost, impact on subjects, etc. As a result of illumination variation, motion artifacts, and video compression, the cardiac pulses are small in comparison to the numerous other factors that affect skin appearance over time.
Heart rate detection based on face video largely depends on the illumination on the face. Both indoor and outdoor ambient lighting conditions are subject to constant change, and direct lighting sources can cast strong shadows that accentuate or diminish certain facial features [8]. There are two main methods to solve the illumination problems. The first method is to separate the illumination change signal from the pulsating signal using signal-processing methods such as the ensemble empirical mode decomposition (EEMD) algorithm [9,10]. The second method considers the facial region of interest (ROI) and background ROI to have similar illumination changes, and the background ROI is used as the noise reference to correct the interference of the illumination changes [11,12].
The distance (angle) from the light source to the skin tissue and to the camera is affected by motion artifacts. Poh et al. [13] proposed a blind source separation method to reduce the error caused by motion. In addition, there are some other methods to solve the problem of motion artifacts, such as Chrominance (CHROM) [14], wavelet transform [15], and bounded Kalman filter technology [16], etc.
Currently, most of the existing rPPG methods are effective based on uncompressed video data. McDuff et al. [17] pointed out that video compression has a great influence on video-based physiological signal extraction. The uncompressed videos occupy a large amount of storage space, which impedes the sharing of the data online. Zhao et al. [18] tried to extract the rPPG signal in the presence of video compression artifacts and proposed a video compression algorithm, named POSSC [19], so that the existing rPPG signal extraction methods can be directly applied to compressed video.
Although the aforementioned studies have made outstanding contributions to dealing with illumination variation, motion artifacts, and video compression, the video used for testing requires the whole frontal face to be included. However, situations in which the side of the face is directed to camera or the face is too close to the camera, are very common amongst students in class. As a result of the limitations of traditional non-contact heart rate detection, the above methods cannot work in the mentioned special cases. Thus, we improved it to apply to online teaching scenarios. In this paper, a method is presented to achieve heart rate detection under special conditions. In summary, our contributions are:
(1)
We propose a symmetry substitution method. When the head is rotated 30 degrees to 45 degrees, and the region of interest in the face is partially missing, the data detected in the left and right cheeks are symmetrically copied;
(2)
We designed a method to determine the effective facial region of interest (ROI) based on the face–eye location then calculate the physiological parameters;
(3)
We designed a video dataset in minutes.

2. Related Work

The current research methods focus on illumination [16,20], motion [20,21,22], video compression [19], and other aspects, but less attention is paid to the selection of regions of interest, the impact of face detection methods, and the datasets suitable for heart rate detection.
The ROI of reliable sites is the key to extracting physiological parameters based on the rPPG method and directly affects the accuracy of the measured values [23]. Marnix et al. [24] found that the use of video cameras to collect facial skin tissues is very accurate in calculating heart rate through rPPG, but the measurement of heart rate in the wrist and calf region is not reliable. Therefore, all video-based heart rate measurements need to be recorded from the face. There are three commonly used regions of interest: the full face (the whole face detected using a face detection algorithm); three rectangular regions (the forehead and left and right cheeks); and the band region (a rectangular area of the left and right cheek and nose). Table 1 summarizes the regions of interest used in related articles and compares their effects.
Face detection is a crucial step in heart rate measurement. Histogram of Oriented Gradient (HOG) [29] is a local descriptor that uses a gradient vector direction histogram. It uses the pixel gradient distribution to describe the appearance and shape of objects. The Viola–Jones algorithm [30] describes the gray distribution of human faces using Harr features, improves the detection speed using an integral graph, and then constructs a cascade classifier using the AdaBoost algorithm. The multi-task convolutional neural network (MTCNN) model [31] contains three cascaded multi-task convolutional neural networks, which can detect face and feature points simultaneously. This model outperforms the state-of-the-art methods, and a small part of the face can be blocked during measurement. Deep Alignment Network (DAN) [32] is a cascaded depth neural network that employs a thermal map to provide visual information on the position of key points on the face. When a small part of the face is occluded, the DAN face detection model can also be carried out. This paper compares the advantages and disadvantages of these four face detection methods. Information regarding the detection effects is shown in Table 2.
In addition, most published papers regarding the recovery of HR from facial videos were assessed on privately owned datasets. However, it is not fair to compare different detection methods on different datasets. Therefore, a new publicly available dataset, directly related to rPPG-suitable practical applications, is vital. The currently available datasets include MAHNOB-HCI [33], DEAP [34], MMSE-HR [35], PURE [36], OBF [37], and VIPL-HR [38] et al., and their specifications are listed in Table 3.
At present, rPPG is tentatively applied in Intensive Care Units (ICU), because the subjects are still and a frontal face video can be continuously collected. Under these ideal conditions, the root mean square error (RMSE) of the traditional non-contact heart rate measurement method is between 1.24 and 8.90 bpm [16,27,39,40]. The frontal face RMSE of our proposed method reaches 3.49 bpm, which is very close to state-of-the-art traditional methods. However, it is rare that the face information is partly missing in medical scenarios. Therefore, the proposed method is more applicable in teaching scenarios.

3. Methods

3.1. Process

Traditional non-contact heart rate measurement process is divided into three steps. Firstly, the video frame is processed, then the Blood Volume Pulse (BVP) signal is extracted, and finally the physiological parameters are calculated. It is necessary for the face to be detected in all three steps to calculate the heart rate, but this is sometimes impossible. The objective of the method is to enable a real-time measurement of HR across different face position conditions. If the face is not detected, the eye region is located to find the region of interest, and then the heart rate is calculated. Figure 1 is an outline of the method. The following sections provide a detailed explanation of the various processes involved in achieving the objective. Our contributions are highlighted in color.

3.1.1. Face Detection and Facial Feature Points Determination

In this step, face detection is carried out for each frame, and regions of interest are created based on the specific facial feature points on the detected face. With reference to Table 2, the MTCNN and DAN methods are accurate but time-consuming, and they are not suitable for real-time detection. The Viola–Jones face detection method is faster because it has no feature points marking function, and illumination intensity and direction do not easily change in online teaching conditions.

3.1.2. ROI Creation

After face detection and feature point marking, it is necessary to select the ROI of face and extract the BVP signal related to physiological signal from the ROI. Regions of interest identification is a crucial step in heart rate measurement. On the basis of most methods, this paper discusses the influence of different face regions and their combinations on heart rate measurement. In our experiments, the face was divided into eight regions: left cheek; right cheek; forehead; left cheek and forehead; right cheek and forehead; left cheek and right cheek; left cheek and right cheek and forehead; left cheek and right cheek and nose. The heart rate values were calculated through these regions and compared with the ground truth. Then, the mean absolute error (MAE) was calculated for comparison to find the optimal region. Table 4 compares the different ROI definition methods and their respective measurement MAE. The optimal situation is expressed in bold.
Because there are more capillaries in the cheek area, the ROI from the cheeks and nose region was found to achieve better measurement results and hence were chosen as the ROI.

3.1.3. Skin Segmentation

The principle of rPPG is to extract the signals related to the cardiac cycle from the subtle color changes in the skin. The background, clothes, teeth, hair, and other unrelated parts are useless for heart rate detection. Detecting such regions will not yield accurate HR measurements. Skin detection is performed on every frame to filter out non-skin pixels. This paper uses the skin color model established by Pitas et al. in H, S, V (Hue, Saturation, Value) space as given by Equation (1). It does not require color normalization and has strong robustness to illumination. Only when the following conditions are met can the skin be segmented:
{ ( 0 o H 25 o ) ( 335 o H 360 o ) ( 0.2 S 0.6 ) ( 0.4 V )
The heart rate measurement effect of skin segmentation and non-skin segmentation was compared, and the calculation average error is shown in Figure 2. The results in Figure 2 illustrate that the heart rate measured by skin segmentation was closer to the ground truth, that is, the measurement accuracy can be improved by skin pixel extraction.

3.1.4. Raw Traces Extraction and Signal Processing

The raw red, green, blue (RGB) signals are obtained by calculating the average pixel value of the skin pixels within the ROI region over time. Then, the whole video sequence is transformed into three one-dimensional signals as raw signals, as shown in Figure 3 step a. After extracting the raw signals, signal-processing techniques such as detrending, normalization, smoothing, and filtering are employed to refine the signal. Firstly, we used detrending to remove linear trends from the raw signal, then the raw signal was normalized dividing the raw signal by its absolute value. Secondly, we employed a five-point sliding average filter to remove random noise.

3.1.5. Independent Component Analysis

The RGB signals contain information about HR, but it is always mixed with noise. FastICA, one of the ICA methods, is an effective technique that can be utilized to eliminate noise artifacts. We used FastICA to extract the raw source signals from signals with noise. After FastICA, we obtained three unsorted independent source components from the extracted signals, as shown in Figure 3 step b. Verkruysse’s study showed that the green channel signal contains the strongest plethysmographic signal among all three channels [7]. Therefore, Pearson correlation analysis was performed between the three independent source signals and green channel signal. The Pearson correlation coefficients of the three components were 0.89, 0.46, and 0.04, respectively. This demonstrates a strong correlation between the upper example in Figure 3 step b and the green channel. Thus, the upper example in Figure 3 step b was selected as the BVP signal, as shown in Figure 3 step c. The other two components were discarded.

3.1.6. Heart Rate Calculation

In this step, we applied a Hamming window-based bandpass filter with cutoff frequencies of 0.75 and 4 Hz to refine the BVP signal. This bandpass is common in previous research. The filtered signal is shown in Figure 3 step d. Then, we employed Fourier transform to obtain the frequency spectrum of the refined BVP signal, as shown in Figure 3 step e. The measured heart rate value can be calculated by finding the frequency corresponding to the highest peak value of the spectrum.

3.2. Symmetry Substitution Method

The subjects lose the left cheek or right cheek information in the region of interest when they face the camera sideways. On the basis of the statistical experiment (see Table 4 include some data), there was no significant difference in heart rate measurement between the left and right cheek. In this paper, a symmetrical substitution method is proposed, which replaces the undetected ROI of the left and right cheek with the one that can be detected. As a result of the limitations of the face detection algorithm, the face cannot be detected if the deflection angle is too large. Through experimentation, we found that the effective angle range of the symmetry substitution method proposed in this paper is 30 degrees to 45 degrees.
The goal of the symmetry substitution method is to calculate the Sleft and Sright of the ROIleft and ROIright (here, this refers to the left and right cheek in the video, contrary to the actual situation), and then judge whether there is a measurement missing through the area ratio. If it is missing, the symmetrical substitution method is used; if not, the original region of interest is used. Assuming that the area ratio of left cheek to right cheek is a, when a is greater than 1.5, the right cheek disappears; when a is less than 0.66, the left cheek disappears; when a is between 0.66 and 1.5, it indicates that no region of interest disappears as given by Equation (2).
a = S l e f t S r i g h t ; R O I = { R O I n o s e + 2 R O I l e f t , a > 1.5 R O I n o s e + 2 R O I r i g h t , a < 0.66 R O I n o s e + R O I l e f t + R O I r i g h t , 0.66 a 1.5

3.3. Heart Rate Estimation by Face–Eye Location

In the case of no face detection, the coordinates of human eyes are used in this paper to determine the region of interest in the face. The coordinate positions of the two eyes on the face are determined and marked with a rectangle, as shown in Figure 4b and Figure 5b. Suppose that the coordinates of point A of the left eye are (x1, y1), the coordinates of point B are (x2, y2), the coordinates of point C of the right eye are (x3, y3), and the coordinates of point D are (x4, y4). The position of facial ROI is determined by the coordinates of two eyes. When the subject’s face is too close to the camera, the coordinates of two points of E (x5, y5) and F(x6, y6) are those shown in Formula (3). When the subject is wearing a mask, the coordinates of two points of M (x7, y7) and N(x8, y8) are those shown in Formula (4).
{ x 5 = x 1 y 5 = y 2 x 6 = x 4 y 6 = y 2 + ( y 2 y 1 ) = 2 y 2 y 1
According to the coordinates of the rectangle, the ROI region can be determined, as shown in Figure 4c and Figure 5c. The heart rate is further calculated according to the region. Because the height of the eye area is basically fixed each time, we found that the coordinate of y5 is equal to y2, the coordinates of y6 are equal to y2 plus the rectangle box height, y7 is obtained by subtracting three times the rectangle box height from the coordinates of y2, and y8 is obtained by subtracting one times the rectangle box height from the coordinates of y3. The size of the area enclosed by point M and N is suitable for the subsequent calculation.
{ x 7 = x 2 y 7 = y 2 3 ( y 2 y 1 ) = 3 y 1 2 y 2 x 8 = x 3 y 8 = y 3 ( y 4 y 3 ) = 2 y 3 y 4

3.4. Participants

Subjects included three males and three females in our dataset. All subjects wore glasses or sunglasses. One of the female subjects covered her forehead and only showed her eyes and below. All subjects were asked to record videos without makeup.

3.5. Experimental Environment and Benchmark Dataset

Several groups of videos were collected in two static states: head to camera and head side to camera. Videos were recorded in the natural environment at home. During the recording process, the subjects were required to hold the finger clip pulse oximeter (model DB12) for the real-time measurement. Various parameters are shown in Table 5. The measurement results can be displayed in the video, so that the recorded video contained not only facial information, but also the results of the detection using professional equipment (the real value), which was convenient for comparison with the calculated heart rate. Our dataset environment is shown in Figure 6a. The status of students in online class is shown in Figure 6b.
We created our own dataset containing 70 video sequences (each video lasting 60 s) using the Chicony USB 2.0 camera webcam. The light source was a mixture of fluorescent lamps and natural light. All videos were recorded in a 24-bit RGB color format with a resolution of 640 × 480, at 29.97 frames/s (NTSC video standard), and were stored as uncompressed data. When recording the video, the subjects were required to sit away from the camera (35~60 cm), keeping their head from shaking as much as possible. Their sight line was consistent with the camera level, and natural blinking was allowed. Our dataset description is shown in Table 6.

3.6. Emotion Experiments Design

Emotion can affect heart rate when the subject is stimulated [5]. We conducted two experiments using our method to prove that it can provide effective data support for emotion recognition. Ten students were subjected to emotional stimuli and asked to imitate emotions in the experiments.
The purpose of Experiment 1 was to verify that heart rate changes significantly when the emotion changes. We designed two scenarios in which the subjects could be angry while looking at a slide. When the subjects were watching the slide and their emotions changed, significant heart rate changes were detected by our method. Because the subjects watched at different speeds, the time it took to begin viewing certain scenes varied.
The purpose of Experiment 2 was to discover the heart rate changes when the subjects imitated the expressions, such as happiness, anger, surprise, etc., according to the cue.

4. Results

4.1. Results of the Symmetrical Substitution Method

This paper investigates the effects of two different ROI symmetric substitution methods, namely, the band region and the three rectangular regions.
For the band region, the comparison of the RMSE results between the symmetric substitution method and non-symmetric substitution method are shown in Figure 7.
Furthermore, we compared the effects of using the symmetric substitution method and not using the symmetric substitution method in the case of the frontal face. The symmetrical substitution method included two cases: replacing the right face with the left face and replacing the left face with the right face. The RMSE results are shown in Table 7. The optimal situation is expressed in bold.
Because the three rectangular regions cannot estimate heart rate without a region of interest, only the symmetric substitution method can be considered. The RMSE is shown in Figure 8.

4.2. Results of Heart Rate Estimated by Face–Eye Location

When the whole face is not detected, the coordinates of the eyes are selected to determine the region of interest in the face. In the case of a frontal face, the method can detect human eyes whether the eyes are open or closed, and whether the eyes are looking out of the computer screen. In the case of a profile, if only one eye can be detected, the corresponding skin pixel under the eye is taken. The results of heart rate measurement with the face–eye location method are shown in Table 8. Although this allows for heart rate detection without facial detection, the detection time is increased.

4.3. Comparison with Other Methods

For special cases, we used the following methods on our database to compare with our method, and the results are shown in Table 9.

4.4. Results of Emotion Experiments

Firstly, the heart rate time series acquired using our method are illustrated in Figure 9. From 20 s to 160 s, there are two kinds of visual stimulation causing emotional changes in the subjects. According to the time point where the subjects switch slides, we marked two dotted lines in each graph. The section before the first dashed line represents the subjects before they were stimulated, and the section between the first and second dashed line represents the subjects seeing the scene that made them angry. The heart rates of subjects 1, 4, and 7 changed obviously. The heart rates of subjects 2, 9, and 10 changed slightly. The heart rates of subjects 3, 5, and 6 remained basically unchanged. The heart rate of subject 8 was undetected.
Secondly, we observed the changes in the heart rate of the subjects when they imitated expressions. The results are shown in Figure 10. The section of the subjects imitating the expressions was between the first and second dashed lines. The heart rates of subjects 3, 5, 6, and 10 remained basically unchanged. The heart rates of subjects 1, 7, and 9 were relatively stable in the second half of the imitation. The heart rates of subjects 2 and 4 changed obviously. The heart rate of subject 8 was undetected.

5. Discussion

On the basis of our study results, we illustrated the effectiveness of using the symmetrical substitution and face-eye location methods to estimate heart rate.
According to the definition of ROI in Table 4, the minimum average error of heart rate measured in each line appeared most frequently in the region of interest composed of the left and right cheek and nose. This is because the whole face area contains a mouth and eyes, and the signal-to-noise ratio is reduced due to unavoidable actions such as breathing and blinking. There are more capillaries in the cheek area, so the rPPG signal quality in this band region is the best. Compared with the average error measured by the left cheek or right cheek, it can be seen that there was no significant difference between them, which also provides a theoretical basis for symmetrical substitution.
When one cheek disappeared, the physiological signal was measured using the symmetrical substitution method. For the banded region of interest, we compared the ROI of skin pixels composed of the left and right cheek and nose with those obtained without the symmetrical substitution method. The RMSE, as shown in Figure 7, reveals that the same video measured using the symmetrical substitution method can be up to 2 bpm when the ROI disappears and the ROI is complete. We found that the RMSE of the 10 videos were distributed in the range of 1–8 bpm. We can conclude that the symmetrical substitution method does not have a significant influence on the measurement accuracy in the case of a profile. As shown in Table 7, we found that whether the left cheek is replaced by the right cheek or the right cheek is replaced by the left cheek, the measured heart rate results were similar to those without symmetrical replacement. For the three rectangular regions, when a region of interest disappears, the heart rate measurement cannot be carried out. In this case, the symmetrical substitution method is discussed. We found that the RMSE of the measurement was generally between 1–7 bpm as shown in Figure 8. Another result worth highlighting is that the darker the video, the higher the RMSE value.
Face–eye location can solve the problem of the face–camera distance being too small to detect the face. The ROI on the face under the eyes can be found through the coordinates, and then the heart rate can be estimated. Similarly, in the case of wearing a mask, if it is not possible to detect the face, our method locates the two eyes, then finds the ROI on the forehead, before extracting the skin pixels and detecting the heart rate. When the profile is positioned towards the camera, that is, the head is turned, our method can also locate a person’s eye, detect the skin area under the eye, and then detect the heart rate. Therefore, as long as the eyes are located, i.e., at least part of the forehead, left cheek, and right cheek are visible, the heart rate can be detected. It can be seen from Table 8 that the MAE and RMSE of the heart rate measurement value was less than 6.15 bpm, which indicates that the method of locating the ROI with human eyes is relatively accurate. When both the face and eyes could not be detected, we could detect the heart rate by detecting the skin pixels. However, cases in which face information could not be detected completely was not within the scope of the study. In our dataset, the videos were recorded with complete face information and partial face information.
From Figure 9 and Figure 10, we found that when subjects received stimulation, their heart rates increased, while simple imitation did not have a great impact on their heart rates. In online teaching, it is possible to analyze students’ real emotions through the change of students’ heart rates. Our method confirms that emotional changes can cause heart rate changes. However, there was no significant change in heart rate when the subjects imitated the expression according to the cue. The results show that our method can provide effective data support for emotion recognition.
Furthermore, in the absence of information, our RMSE was in a range of 4.79 to 6.15 bpm. Our accuracy was compared with other methods in special situations, as shown in the Table 9. Our method can estimate heart rate in the absence of information (while a mask is being worn, the face–camera distance is too small, or the profile is being presented), while traditional methods cannot estimate heart rate in these cases.
We use the value of non-contact heart rate detection to provide data support for changes in students’ emotion in online teaching scenarios. There is a lot of facial information missing in e-learning, which makes it impossible to measure heart rate with traditional methods based on rPPG. In the proposed method, we use the symmetrical substitution method to make up for the lack of ROI when the profile of the face is angled towards the camera, and use face–eye location instead of face location to solve the problem of masks and the face being too close to the camera, which are seldom mentioned by other researchers. The traditional methods cannot complete the heart rate detection in these situations.
From the analyses, we think that heart rate can be used as one of the reference indicators to detect real emotions, just like expressions. Therefore, our method can provide data support for learning state detection in smart education.

6. Conclusions

Using physiological signals measured using the rPPG method to evaluate students’ emotional problems in education and teaching has received a lot of attention in recent years. However, it is susceptible to problems when the head is rotated and the distance between the head and camera changes. Firstly, we presented a symmetrical substitution method. When the subjects’ head is rotated 30 degrees to 45 degrees, the detected ROI data are copied to the undetected areas. Our results indicate that the heart rate measurement accuracy does not decrease compared with that obtained in the full face condition. Secondly, this paper presents a method of heart rate detection based on face–eye location instead of face detection as in traditional methods. It can also detect heart rate under special circumstances, with a RMSE under 7.64 bpm. Thirdly, we found that the heart rate rises when the subjects receive stimulation, and the heart rate remains basically unchanged when they imitate expressions. In addition, the profile dataset contains videos recorded under special circumstances, providing data support for the research of physiological signal detection based on rPPG in real situations. Future research will be devoted to improving the real-time stability and accuracy of the method.

Author Contributions

Conceptualization, K.Z. and J.Z.; methodology, K.Z.; software, K.C.; validation, K.C., J.K. and J.C.; formal analysis, J.Z. and K.C.; investigation, K.C.; resources, K.C. and J.K.; data curation, K.C., J.C. and J.K.; writing—original draft preparation, K.C.; writing—review and editing, K.Z. and J.C.; visualization, K.C.; supervision, K.Z.; project administration, J.Z.; funding acquisition, K.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Beijing Education Science Planning Project, grant number CADA18069.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pekrun, R.; Goetz, T.; Titz, W.; Perry, R.P. Academic Emotions in Students’ Self-Regulated Learning and Achievement: A Program of Qualitative and Quantitative Research. Educ. Psychol. 2002, 37, 91–105. [Google Scholar] [CrossRef]
  2. Chettupuzhakkaran, P.; Sindhu, N. Emotion Recognition from Physiological Signals Using Time-Frequency Analysis Methods. In Proceedings of the 2018 International Conference on Emerging Trends and Innovations in Engineering and Technological Research (ICETIETR), Ernakulam, India, 11–13 July 2018; pp. 1–5. [Google Scholar]
  3. Joesph, C.; Rajeswari, A.; Premalatha, B.; Balapriya, C. Implementation of physiological signal based emotion recognition algorithm. In Proceedings of the 2020 IEEE 36th International Conference on Data Engineering (ICDE), Dallas, TX, USA, 20–24 April 2020; pp. 2075–2079. [Google Scholar]
  4. Bland, M.W.; Morrison, E. The Experimental Detection of an Emotional Response to the Idea of Evolution. Am. Biol. Teach. 2015, 77, 413–420. [Google Scholar] [CrossRef]
  5. Stavroulia, K.E.; Christofi, M.; Baka, E.; Michael-Grigoriou, D.; Magnenat-Thalmann, N.; Lanitis, A. Assessing the emotional impact of virtual reality-based teacher training. Int. J. Inf. Learn. Technol. 2019, 36, 192–217. [Google Scholar] [CrossRef] [Green Version]
  6. Thompson, N.; McGill, T.J. Genetics with Jean: The design, development and evaluation of an affective tutoring system. Educ. Technol. Res. Dev. 2016, 65, 279–299. [Google Scholar] [CrossRef]
  7. Verkruysse, W.; Svaasand, L.O.; Nelson, J.S. Remote plethysmographic imaging using ambient light. Opt. Express 2008, 16, 21434–21445. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Chen, D.-Y.; Wang, J.-J.; Lin, K.-Y.; Chang, H.-H.; Wu, H.-K.; Chen, Y.-S.; Lee, S.-Y. Image Sensor-Based Heart Rate Evaluation from Face Reflectance Using Hilbert–Huang Transform. IEEE Sens. J. 2015, 15, 618–627. [Google Scholar] [CrossRef]
  9. Lin, K.-Y.; Chen, D.-Y.; Tsai, W.-J. Face-Based Heart Rate Signal Decomposition and Evaluation Using Multiple Linear Regression. IEEE Sens. J. 2016, 16, 1351–1360. [Google Scholar] [CrossRef]
  10. Chen, X.; Chen, Q.; Zhang, Y.; Wang, Z.J. A Novel EEMD-CCA Approach to Removing Muscle Artifacts for Pervasive EEG. IEEE Sens. J. 2018, 19, 8420–8431. [Google Scholar] [CrossRef]
  11. Cheng, J.; Chen, X.; Xu, L.; Wang, J. Illumination Variation-Resistant Video-Based Heart Rate Measurement Using Joint Blind Source Separation and Ensemble Empirical Mode Decomposition. IEEE J. Biomed. Health Inform. 2017, 21, 1422–1433. [Google Scholar] [CrossRef]
  12. Xu, L.; Cheng, J.; Chen, X. Illumination variation interference suppression in remote PPG using PLS and MEMD. Electron. Lett. 2017, 53, 216–218. [Google Scholar] [CrossRef]
  13. Poh, M.-Z.; McDuff, D.; Picard, R.W. Non-contact, automated cardiac pulse measurements using video imaging and blind source separation. Opt. Express 2010, 18, 10762–10774. [Google Scholar] [CrossRef] [PubMed]
  14. De Haan, G.G.; Jeanne, V. Robust Pulse Rate from Chrominance-Based rPPG. IEEE Trans. Biomed. Eng. 2013, 60, 2878–2886. [Google Scholar] [CrossRef]
  15. Bousefsaf, F.; Maaoui, C.; Pruski, A. Continuous wavelet filtering on webcam photoplethysmographic signals to remotely assess the instantaneous heart rate. Biomed. Signal Process. Control. 2013, 8, 568–574. [Google Scholar] [CrossRef]
  16. Prakash, S.K.A.; Tucker, C.S. Bounded Kalman filter method for motion-robust, non-contact heart rate estimation. Biomed. Opt. Express 2018, 9, 873–897. [Google Scholar] [CrossRef] [PubMed]
  17. McDuff, D.J.; Blackford, E.B.; Estepp, J.R. The Impact of Video Compression on Remote Cardiac Pulse Measurement Using Imaging Photoplethysmography. In Proceedings of the 2017 12th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2017), Washington, DC, USA, 30 May–3 June 2017; pp. 63–70. [Google Scholar]
  18. Zhao, C.; Lin, C.-L.; Chen, W.; Li, Z. A Novel Framework for Remote Photoplethysmography Pulse Extraction on Compressed Videos. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 1380–1389. [Google Scholar]
  19. Zhao, C.; Chen, W.; Lin, C.-L.; Wu, X. Physiological Signal Preserving Video Compression for Remote Photoplethysmography. IEEE Sens. J. 2019, 19, 4537–4548. [Google Scholar] [CrossRef]
  20. Nooralishahi, P.; Loo, C.K.; Shiung, L.W. Robust remote heart rate estimation from multiple asynchronous noisy channels using autoregressive model with Kalman filter. Biomed. Signal Process. Control. 2019, 47, 366–379. [Google Scholar] [CrossRef]
  21. Yang, Z.; Yang, X.; Jin, J.; Wu, X. Motion-resistant heart rate measurement from face videos using patch-based fusion. Signal Image Video Process. 2019, 13, 423–430. [Google Scholar] [CrossRef]
  22. Yang, Z.; Yang, X.; Wu, X. Motion-tolerant heart rate estimation from face videos using derivative filter. Multimed. Tools Appl. 2019, 78, 26747–26757. [Google Scholar] [CrossRef]
  23. Bobbia, S.; Luguern, D.; Benezeth, Y.; Nakamura, K.; Gomez, R.; Dubois, J. Real-Time Temporal Superpixels for Unsupervised Remote Photoplethysmography. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 1422–14227. [Google Scholar]
  24. Van Der Kooij, K.M.; Naber, M. An open-source remote heart rate imaging method with practical apparatus and algorithms. Behav. Res. Methods 2019, 51, 2106–2119. [Google Scholar] [CrossRef]
  25. Qi, L.; Yu, H.; Xu, L.; Mpanda, R.S.; Greenwald, S.E. Robust heart-rate estimation from facial videos using Project_ICA. Physiol. Meas. 2019, 40, 085007. [Google Scholar] [CrossRef]
  26. Fouad, R.M.; Omer, O.A.; Aly, M.H. Optimizing Remote Photoplethysmography Using Adaptive Skin Segmentation for Real-Time Heart Rate Monitoring. IEEE Access 2019, 7, 76513–76528. [Google Scholar] [CrossRef]
  27. Qiu, Y.; Liu, Y.; Arteaga-Falconi, J.S.; Dong, H.; El Saddik, A. EVM-CNN: Real-Time Contactless Heart Rate Estimation From Facial Video. IEEE Trans. Multimed. 2018, 21, 1778–1787. [Google Scholar] [CrossRef]
  28. Rong, M.; Fan, Q.; Li, K. Research on non-contact physiological parameter measurement algorithm based on IPPG. Biomed. Eng. Res. 2018, 37, 27–31. [Google Scholar]
  29. Dalal, N.; Triggs, B. Histograms of Oriented Gradients for Human Detection. In Proceedings of the 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’05), San Diego, CA, USA, 20–25 June 2005; pp. 886–893. [Google Scholar]
  30. Viola, P.; Jones, M.J. Robust Real-Time Face Detection. Int. J. Comput. Vis. 2004, 57, 137–154. [Google Scholar] [CrossRef]
  31. Zhang, K.; Zhang, Z.; Li, Z.; Qiao, Y. Joint Face Detection and Alignment Using Multitask Cascaded Convolutional Networks. IEEE Signal Process. Lett. 2016, 23, 1499–1503. [Google Scholar] [CrossRef] [Green Version]
  32. Kowalski, M.; Naruniec, J.; Trzcinski, T. Deep Alignment Network: A Convolutional Neural Network for Robust Face Alignment. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2017; pp. 2034–2043. [Google Scholar]
  33. Soleymani, M.; Lichtenauer, J.; Pun, T.; Pantic, M. A Multimodal Database for Affect Recognition and Implicit Tagging. IEEE Trans. Affect. Comput. 2012, 3, 42–55. [Google Scholar] [CrossRef] [Green Version]
  34. Koelstra, S.; Muhl, C.; Soleymani, M.; Lee, J.-S.; Yazdani, A.; Ebrahimi, T.; Pun, T.; Nijholt, A.; Patras, I. DEAP: A Database for Emotion Analysis Using Physiological Signals. IEEE Trans. Affect. Comput. 2011, 3, 18–31. [Google Scholar] [CrossRef] [Green Version]
  35. Tulyakov, S.; Alameda-Pineda, X.; Ricci, E.; Yin, L.; Cohn, J.F.; Sebe, N. Self-Adaptive Matrix Completion for Heart Rate Estimation from Face Videos under Realistic Conditions. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2396–2404. [Google Scholar]
  36. Stricker, R.; Muller, S.; Gross, H.-M. Non-contact video-based pulse rate measurement on a mobile service robot. In Proceedings of the the 23rd IEEE International Symposium on Robot and Human Interactive Communication, Edinburgh, UK, 25–29 August 2014; pp. 1056–1062. [Google Scholar]
  37. Li, X.; Alikhani, I.; Shi, J.; Seppanen, T.; Junttila, J.; Majamaa-Voltti, K.; Tulppo, M.; Zhao, G. The OBF Database: A Large Face Video Database for Remote Physiological Signal Measurement and Atrial Fibrillation Detection. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), Xi’an, China, 15–19 May 2018; pp. 242–249. [Google Scholar]
  38. Niu, X.; Shan, S.; Han, H.; Chen, X. RhythmNet: End-to-End Heart Rate Estimation from Face via Spatial-Temporal Representation. IEEE Trans. Image Process. 2020, 29, 2409–2423. [Google Scholar] [CrossRef] [Green Version]
  39. Poh, M.-Z.; McDuff, D.; Picard, R.W. Advancements in Noncontact, Multiparameter Physiological Measurements Using a Webcam. IEEE Trans. Biomed. Eng. 2010, 58, 7–11. [Google Scholar] [CrossRef] [Green Version]
  40. Lam, A.; Kuno, Y. Robust Heart Rate Measurement from Video Using Select Random Patches. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 3640–3648. [Google Scholar]
  41. Wang, W.; Brinker, A.C.D.; Stuijk, S.S.; De Haan, G.G. Algorithmic Principles of Remote PPG. IEEE Trans. Biomed. Eng. 2017, 64, 1479–1491. [Google Scholar] [CrossRef] [Green Version]
  42. Henriques, J.F.; Caseiro, R.; Martins, P.; Batista, J. Exploiting the Circulant Structure of Tracking-by-Detection with Kernels. Lect. Notes Comput. Sci. 2012, 7575, 702–715. [Google Scholar]
Figure 1. Outline of the proposed algorithm.
Figure 1. Outline of the proposed algorithm.
Sensors 20 07021 g001
Figure 2. Comparison of the measurement effect between skin segmentation and non-skin segmentation.
Figure 2. Comparison of the measurement effect between skin segmentation and non-skin segmentation.
Sensors 20 07021 g002
Figure 3. The detailed flowchart of our proposed method. (step a) Three one-dimensional signals. (step b) Three unsorted independent source components. (step c) The BVP signal. (step d) The filtered signal. (step e) The frequency spectrum of the refined BVP signal.
Figure 3. The detailed flowchart of our proposed method. (step a) Three one-dimensional signals. (step b) Three unsorted independent source components. (step c) The BVP signal. (step d) The filtered signal. (step e) The frequency spectrum of the refined BVP signal.
Sensors 20 07021 g003
Figure 4. ROI location by eye while closing to camera.
Figure 4. ROI location by eye while closing to camera.
Sensors 20 07021 g004
Figure 5. ROI location by eye while wearing a mask.
Figure 5. ROI location by eye while wearing a mask.
Sensors 20 07021 g005
Figure 6. Dataset and experiment environment.
Figure 6. Dataset and experiment environment.
Sensors 20 07021 g006
Figure 7. Comparison of the effect of the symmetrical substitution method in the band region.
Figure 7. Comparison of the effect of the symmetrical substitution method in the band region.
Sensors 20 07021 g007
Figure 8. Comparison of the effect of the symmetrical substitution method in the three rectangular regions.
Figure 8. Comparison of the effect of the symmetrical substitution method in the three rectangular regions.
Sensors 20 07021 g008
Figure 9. Heart rate changes when subjects were stimulated.
Figure 9. Heart rate changes when subjects were stimulated.
Sensors 20 07021 g009
Figure 10. Heart rate changes when subjects imitated the expressions.
Figure 10. Heart rate changes when subjects imitated the expressions.
Sensors 20 07021 g010
Table 1. Comparison of different region of interest (ROI) definitions and measurement results.
Table 1. Comparison of different region of interest (ROI) definitions and measurement results.
ArticleROI DefinitionMethodResult
[13]full faceICAThe root mean square error of the static dataset was 2.29 bpm, and that of moving dataset was 4.63 bpm.
[25]full faceProject_ICAIn static, human–computer interaction and exercise recovery scenarios, the mean absolute deviation were 3.30, 3.93, and 9.80 bpm, respectively.
[16]three rectangular regionsbounded Kalman filterThe average measurement error was 3 bpm when the subjects walked to the camera from 4 feet away.
[20]three rectangular regionsRADICALThe average error was 1.42 bpm in a well-controlled dataset.
[26]three rectangular regionsPCAThe accuracy rate of heart rate measurement of five subjects was above 98%.
[27]band regionEVM + CNN74.13% of the test data were well estimated.
[28]band regionICAThe root mean square error was 2.258 bpm under static condition.
Table 2. Advantages and disadvantages of the four face detection methods.
Table 2. Advantages and disadvantages of the four face detection methods.
Face Detection EffectPerformance DescriptionAdvantageDisadvantage
HOG [29] Sensors 20 07021 i001greatly outperforms the wavelet, PCA-SIFT, and Shape Context methodsfast running speed; 68 facial feature pointsgreatly influenced by light intensity and direction; inaccurate location of feature points on profile
Viola–Jones [30] Sensors 20 07021 i002extremely rapid image processing, while achieving high detection ratesfast running speed no facial feature point; greatly influenced by light intensity and direction
MTCNN [31] Sensors 20 07021 i003can achieve very fast speed in joint face detection and alignmentaccurate face detection; less affected by light intensity and directioncomplicated models; complex calculation;
slow running speed; only five feature points can be marked
DAN [32] Sensors 20 07021 i004reduces the state-of-the-art failure rate by up to 70%accurate location of feature points on profile; less affected by light intensity and direction; can mark 68 facial feature pointscomplicated models; complex calculation;
slow running speed
Table 3. Available video datasets for physiological parameter detection.
Table 3. Available video datasets for physiological parameter detection.
DatasetsNumber of SubjectsNumber of VideosCamera ParametersVideo Parameters
MAHNOB-HCI [33]27527Allied Vision Stingray F-046C; F-046BRGB videos
780 × 580 @ 60 fps
DEAP [34]32120Sony DCR-HC27ERGB videos
800 × 600 @ 50 fps
MMSE-HR [35]40102RGB 2D color cameraRGB image sequences
1040 × 1392 @ 25 fps
PURE [36]1060eco274CVGERGB videos
640 × 480 @ 30 fps
OBF [37]1062120Blackmagic URFA miniRGB:1920 × 1080 @ 60 fps
Camera boxNIR:640 × 480 @ 30 fps
VIPL-HR [38]1073130Logitech C310960 × 720 @ 25 fps
Realsense F200NIR:640 × 480 @ 30 fps
Color:1920 × 1080 @ 30 fps
Huawei P9 smartphoneColor:1920 × 1080 @ 30 fps
Table 4. Comparison of different ROI definitions and measurement MAE (L = left cheek, R = right cheek, F = forehead, L-F = left cheek and forehead, R-F = right cheek and forehead, L-R = left cheek and right cheek, L-R-F = left cheek and right cheek and forehead, L-R-N = left cheek and right cheek and nose).
Table 4. Comparison of different ROI definitions and measurement MAE (L = left cheek, R = right cheek, F = forehead, L-F = left cheek and forehead, R-F = right cheek and forehead, L-R = left cheek and right cheek, L-R-F = left cheek and right cheek and forehead, L-R-N = left cheek and right cheek and nose).
No.LRFL-FR-FL-RL-R-FL-R-N
112.7312.3810.2816.3514.1312.0516.688.15
28.756.158.507.857.0810.756.385.50
38.139.957.957.258.338.856.185.68
48.466.108.059.417.2111.418.366.97
52.931.333.693.363.552.361.291.48
61.571.951.503.192.881.902.573.12
74.360.694.383.812.361.863.740.38
85.931.3418.1012.4915.711.2715.541.39
923.4011.1018.5221.3616.174.5716.332.86
109.814.797.953.673.143.293.884.81
Table 5. DB12 parameters.
Table 5. DB12 parameters.
ParameterValue
Pulse Rate Display25–250 bpm
Resolution1 bpm
Measurement accuracy2 bpm
Table 6. Dataset description.
Table 6. Dataset description.
ClassificationDistance (cm)Numbers
Frontal facewith sunglasses35–6010
Wearing a mask35–6010
too small a face–camera distance<2010
no occlusion of the face35–6030
Profileno occlusion of the face35–6010
Table 7. Comparison of the effect of the symmetrical substitution method in the frontal face (L-R = replace right cheek with left cheek; R-L = replace left cheek with right cheek; N-U-S = No use of symmetrical substitution).
Table 7. Comparison of the effect of the symmetrical substitution method in the frontal face (L-R = replace right cheek with left cheek; R-L = replace left cheek with right cheek; N-U-S = No use of symmetrical substitution).
Methods12345678910
L-R5.535.422.572.153.982.445.435.993.690.99
R-L1.325.434.691.373.623.781.762.361.251.47
N-U-S2.256.863.771.415.201.753.812.772.361.88
Table 8. Heart rate measurement results from the face–eye location method.
Table 8. Heart rate measurement results from the face–eye location method.
VideoMAE (bpm)RMSE (bpm)
Too small a face-camera distance4.005.87
Wearing a mask4.196.15
Table 9. Comparison of heart rate measured by different methods.
Table 9. Comparison of heart rate measured by different methods.
MethodsRMSE in Different Conditions (bpm)
Frontal FaceProfileToo Small a Face-Camera DistanceWearing a Mask
CHROM_De Haan [14]Viola–Jones [30]1.1UndetectedUndetectedUndetected
Skin segmentation1.4410.519.4521.34
ICA_Poh [39]Viola–Jones [30]1.24UndetectedUndetectedUndetected
Skin segmentation6.5311.1830.2526.20
POS_Wang [41]OC-SVM [42]8.04UndetectedUndetectedUndetected
Skin segmentation8.3720.7413.6124.69
Bounded Kalman filter [16]Viola–Jones [30]5.56UndetectedUndetectedUndetected
Back Projection6.62UndetectedUndetectedUndetected
Proposed algorithm3.494.795.876.15
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zheng, K.; Ci, K.; Cui, J.; Kong, J.; Zhou, J. Non-Contact Heart Rate Detection When Face Information Is Missing during Online Learning. Sensors 2020, 20, 7021. https://doi.org/10.3390/s20247021

AMA Style

Zheng K, Ci K, Cui J, Kong J, Zhou J. Non-Contact Heart Rate Detection When Face Information Is Missing during Online Learning. Sensors. 2020; 20(24):7021. https://doi.org/10.3390/s20247021

Chicago/Turabian Style

Zheng, Kun, Kangyi Ci, Jinling Cui, Jiangping Kong, and Jing Zhou. 2020. "Non-Contact Heart Rate Detection When Face Information Is Missing during Online Learning" Sensors 20, no. 24: 7021. https://doi.org/10.3390/s20247021

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop