Next Article in Journal
A Systematic Review on Machine Learning and Deep Learning Models for Electronic Information Security in Mobile Networks
Previous Article in Journal
Integrating GEDI and Landsat: Spaceborne Lidar and Four Decades of Optical Imagery for the Analysis of Forest Disturbances and Biomass Changes in Italy
Previous Article in Special Issue
Contribution of Different Subbands of ECG in Sleep Apnea Detection Evaluated Using Filter Bank Decomposition and a Convolutional Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Vision-Based System for In-Sleep Upper-Body and Head Pose Classification

1
Department of Computer Science and Information Engineering, National Taiwan University, Taipei 10167, Taiwan
2
Tainan National University of the Arts, Tainan 72045, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(5), 2014; https://doi.org/10.3390/s22052014
Submission received: 31 December 2021 / Revised: 18 February 2022 / Accepted: 1 March 2022 / Published: 4 March 2022

Abstract

:
Sleep quality is known to have a considerable impact on human health. Recent research shows that head and body pose play a vital role in affecting sleep quality. This paper presents a deep multi-task learning network to perform head and upper-body detection and pose classification during sleep. The proposed system has two major advantages: first, it detects and predicts upper-body pose and head pose simultaneously during sleep, and second, it is a contact-free home security camera-based monitoring system that can work on remote subjects, as it uses images captured by a home security camera. In addition, a synopsis of sleep postures is provided for analysis and diagnosis of sleep patterns. Experimental results show that our multi-task model achieves an average of 92.5% accuracy on challenging datasets, yields the best performance compared to the other methods, and obtains 91.7% accuracy on the real-life overnight sleep data. The proposed system can be applied reliably to extensive public sleep data with various covering conditions and is robust to real-life overnight sleep data.

1. Introduction

Good sleep quality helps the mind and body remain healthy. Irregular sleeping patterns can increase the risk of diabetes, obesity, and cardiovascular disease [1,2]. Numerous studies have shown that sleep postures can serve as an indicator of sleep health. Monitoring in-bed postures provides valuable information regarding the intensity of dreams [3], risk of pressure ulcers [4], patients’ mobility [5], obstructive sleep apnea syndrome [6], risk of spinal symptoms [7], and quality of sleep [8]. Therefore, sleep behavior monitoring is a critical aspect of healthcare management. Moreover, home sleep testing is becoming critical at present owing to the overwhelmed healthcare system resulting from the COVID-19 pandemic [9]. A contact-free sleep monitoring system is necessary for healthcare in the non-contact era.
Traditionally, trunk posture has been used to study the impact on sleep, but a growing number of studies have indicated that head posture has a crucial impact on sleep as well. For example, head posture has a noticeable impact on obstructive sleep apnea (OSA) independent of trunk posture [10,11], and rotation of the head from the supine to lateral posture decreased the OSA severity significantly while the trunk was in the supine posture [10]. A recent study [12] found that head posture during sleep affects the clearance of neurotoxic proteins from the brain. In addition, head posture is related to sleep quality. Sleepers who spend more time on their backs with their heads straight have a high probability of experiencing poor sleep quality [13]. As whole-body shapes under blankets cannot be seen using popular home security cameras, only head and upper-body appearance are considered in this study. Monitoring in-bed head and body posture simultaneously can help us address these issues and better understand sleep behaviors.
Non-invasive technology, such as a computer vision-based system, is better suited to monitoring sleep posture with no disturbance and low cost. However, region detection is usually performed separately from pose classification, and the detecting region is traditionally set as the bed region instead of the human body. To make sleep monitoring practically useful, it needs efficient and robust head and body detection modules to localize heads and bodies with large variations in illumination and coverings for the pose classification tasks. Therefore, this paper presents a sleep monitoring system combining efficient body localization and pose classification using a home security camera. Our system is applicable to both RGB and infrared (IR) modes.
The number of sleep posture datasets with annotations is limited due to privacy and cost concerns. Moreover, the sleep streams are of very long duration, and it needs an efficient way to process the enormous amount of data. The multi-task learning method is applied to overcome the challenge of the sleep dataset. A unified framework is proposed to simultaneously detect and classify head and upper-body posture. All detection and classification tasks are trained and tested simultaneously in an all-in-one convolutional neural network (CNN). This decreases training times, reduces latency during inference, reduces storage space, and is easier to train, deploy, and maintain. The proposed method can save time and the amount paid for processing sleep videos.
To our knowledge, this is the first attempt to merge in-sleep head and upper-body detection as well as head and upper-body pose classification together in one unified network. In addition to recognizing sleep posture, synopses of sleep postures, sleep duration, and related sleep quality indicators are provided. This information can help medical professionals make suitable health recommendations. It can also help users better understand their sleep habits and find ways to improve and maintain sleep health.
The contributions of this work are as follows: (1) this is the first sleep posture study using a single infrared camera and a unified framework for simultaneously detecting and classifying upper-body and head poses during sleep, (2) an all-in-one CNN model with high accuracy is proposed to detect and classify sleep posture, which is robust to realistic sleep conditions, various covering conditions, and variable illumination, (3) a synopsis of sleep postures is provided for analysis and diagnosis of sleep behavior.
The remainder of this paper is organized as follows. Section 2 reviews related work of sleep pose classification and multi-task learning. Section 3 gives an overview of the proposed framework and explains the SleePose-FRCNN-Net architecture and SleePose-FRCNN-Net training strategy in detail. Section 4 describes the datasets used for training and testing and provides experimental results and comparisons with other methods. The discussion of this study is presented in Section 5. Finally, our work is concluded in Section 6.

2. Related Works

This work is related to sleep posture classification and multi-task learning. In this section, the literature is briefly reviewed regarding these two topics.

2.1. Sleep Posture Classification

Sleep posture classification is challenging owing to various reasons, such as variation in the viewing conditions while monitoring human sleep and lack of publicly available sleep datasets because of privacy and cost concerns. In addition, occlusion of the body under a blanket increases the difficulty of detecting and tracking human poses.
Several studies estimate the posture of bodies sleeping in beds using a high-cost pressure-sensing mat [14] or high-resolution thermal camera [15]. However, these devices are too expensive for home use.
Some studies use depth cameras [16,17,18,19]. Grimm et al. [16] use depth maps to detect sleep posture. However, their method does not distinguish between supine and prone postures. Klishkovskaia et al. [19] use Kinect v2 skeleton for the automatic classification of human posture detection. However, their method requires the subject to avoid any covering. In addition, when using depth cameras, the depth measurements may suffer from various noise factors such as ambient light, scene geometry, and glossy reflection [20,21]. The ambient light has an influence on the measurement and correlation of disparities. The scene geometry includes the distance to the object that impacts the error of depth measurement and the occluded or shadowed scenes that can lead to outliers in 3D reconstruction. The reflective surfaces can lead to noise in the depth measurement [21]. Thus, using depth cameras does not always guarantee high accuracy. In recent years, IR cameras have been commonly used for home surveillance. Taking into consideration deployment and prevalence factors, a sleep monitoring system using a single IR camera was chosen to design.
Few studies have focused on head and upper-body pose classification during sleep in comparison with full-body pose classification. Choe et al. [22] employ motion analysis to determine sleep and wake states using a general-purpose head detector. An accuracy of 50% is achieved overnight. A CNN model for tracking upper body joints in the clinical environments from RGB-video is proposed in [23]. The system focuses on monitoring the patient’s pose for clinical studies. Liu and Ostadabbas [24] propose a pre-trained CNN called convolutional pose machine for in-bed pose estimation using a near-infrared modality. Their data is obtained from mannequins in a simulated hospital room. In another approach, a CNN algorithm is developed to detect upper-body and head posture with blankets during sleep [25]. It focuses on the classification of posture without detecting the upper-body and head region.
Recently, benefited from computer vision technologies, an increased number of research efforts have advanced camera-based sleep posture monitoring. A recent deep learning method distinguishes obstructive sleep apnea and central sleep apnea by tracking body movements using an IR camera [26]. Lyu et al. [27] use an object detection algorithm and human pose estimation algorithm to classify sleep posture without covering. A domain adaption-based training strategy [28] is proposed to estimate in-bed human poses using RGB and thermal images. Another study [29] presents a non-contact sleep monitoring system through a transfer learning strategy. Their study mainly focuses on trunk posture instead of head posture.
Table 1 compares the proposed method and previous methods for sleep posture classification. To the best of our knowledge, no previous work has integrated head detection, upper-body detection, head pose classification, and upper-body pose classification into one sleep monitoring system.

2.2. Multi-Task Learning (MTL)

MTL is an excellent solution to share common knowledge among multiple related tasks. Learning correlated tasks jointly can improve performance and offers good generalization ability compared with single-task learning [30].
In a previous study [31], a heterogeneous multi-task model is trained for human pose estimation. They show that the regression network benefits considerably from the various body-part detection tasks. In another study [32], an accurate and cost-efficient MTL framework is employed for simultaneous face detection and 3D head pose estimation.
Research on applying MTL to sleep posture classification is limited. Piriyajitakonkij et al. [33] propose a multi-task learning network to classify sleep postural transition and sleep turning transition. Their application is based on the Ultra-Wideband radar system.
To our knowledge, this is the first attempt to merge in-sleep head and upper-body detection as well as head and upper-body pose classification in a single network.

3. Materials and Methods

This study designs and implements a non-contact sleep monitoring system to perform head and upper-body detection and pose classification. Figure 1 shows an overview of our framework, which consists of two modules:
3.1. Motion detection: A motion detection algorithm is applied to trigger sleep video processing.
3.2. SleePose-FRCNN-Net: Head and upper-body detection and pose classification: a deep multi-task learning (DMTL) network for head and upper-body detection and pose classification (supine, prone, left, or right).

3.1. Motion Detection

A motion detection algorithm called Visual Background Extractor (ViBe) [34] is applied to reduce video processing. The ViBe is a rapid background modeling technique for video sequences. The method is robust and efficient for natural background scenes. The basic idea of the algorithm is to collect background samples for each pixel point to build background models. The elements in the background samples are selected randomly from the pixels around the pixel and used to update the background model. ViBe algorithm mainly consists of three aspects: the classification process of pixels, the initialization of the background mode, and the update strategy of the background model [34].

3.1.1. The Classification Process of Pixels

A background model is built of N background samples for each pixel of the video frame. Equation (1) represents the background model:
M ( x ) = { v 1 ,   v 2 ,     , v N }   ,
where N is the number of background sample sets and v ( x )   is the value of pixel x in the image. The Euclidean distance between v ( x ) and the pixel of model samples M ( x ) is used to determine whether a pixel is similar to the background model samples. If the distance is less than threshold R, the count of the set intersection of the pixel and model samples increases. According to Equation (2), if the count of set intersection is larger than or equal to a given threshold # m i n , the pixel is classified as background; otherwise, it is classified as foreground.
      # { S R ( v ( x ) ) { v 1 ,   v 2 ,     , v N } } # m i n   x   background # { S R ( v ( x ) ) { v 1 ,   v 2 ,     , v N } } < # m i n   x   foreground
where S R ( v ( x ) ) is a circle centered on v(x) and has radius R.

3.1.2. The Initialization of the Background Mode

The first video frame is used to initialize the background model. Certain pixels are randomly selected from their neighborhood to fill the background sample set for each pixel.

3.1.3. The Update Strategy of the Background Mode

The update strategy of the ViBe algorithm is the random replacement. A pixel classified as the background has 1 chance in 16 of being selected to update its background sample set with its current pixel value. This method covers a large time window and better detects the slow-moving targets.

3.1.4. Pre- and Post-Processing

Preprocessing includes histogram equalization for adjusting image intensities across varying bedroom lighting. After applying the ViBe algorithm, the segmentation mask that includes moving object pixels in the frame is obtained. The erosion and dilation processes are used to process the segmentation mask to improve accuracy. The operation is applied to remove foreground blobs whose area is smaller and fill holes to keep the completion of objects.

3.2. SleePose-FRCNN-NET—Head and Upper-Body Detection and Pose Classification

The network aims to recognize the head and upper-body poses during sleep from an image. To this end, the SleePose-FRCNN-Net, a DMTL network, is introduced to jointly detect and classify head pose and upper-body pose from the input image. The Faster R-CNN [35] is adopted as a basic detection framework in our work. The main reason for choosing the Faster R-CNN framework to detect sleep postures is that it has higher accuracy than other deep learning-based detectors [36]. The head detection, upper-body detection, head pose classification, and upper-body pose classification are combined in a unified convolutional neural network.

3.2.1. SleePose-FRCNN-NET Architecture

As shown in Figure 2, the framework consists of four major modules: feature extraction module, region proposal network (RPN) module, head and upper-body detection module, and head and upper-body pose classification module.
  • Feature Extraction Module
From a given image, convolutional features are extracted first. The ResNet-50 [37] network is employed as the backbone architecture in the feature extraction module and initializes the network using the weights pre-trained on the ImageNet dataset. The ResNet-50 model consists of five stages, each with a residual block. Each residual block consists of a set of repeated layers. The Faster R-CNN model allows different input image sizes in the RGB channels. The shortest side can be any number larger than 600 pixels, and the longest side smaller than 1000 pixels, as suggested by the Faster R-CNN paper [35].
2.
RPN Module
The RPN in Faster R-CNN [35] is developed for proposing candidate regions. The RPN predicts a set of candidate object proposals and corresponding objectness scores from feature maps. It is built on top of the res4f layer of the ResNet-50 network, followed by an intermediate 3 × 3 convolutional layer with 512 channels and two sibling 1 × 1 convolutional layers for classification and bounding box regression. The anchor-based method [38] is proposed to detect objects with multiple scales and aspect ratios. This method generates nine bounding boxes, which includes three scales ( 128 2 , 256 2 , and 512 2 pixels) and three aspect ratios (1:1, 1:2, and 2:1). Finally, the RPN classifies each bounding box’s category (an object or not) and regresses four coordinates. The non-maximum suppression (NMS) algorithm prunes redundant, overlapping bounding boxes at the finetuning step.
3.
Head and Upper-Body Detection Module
With the predicted bounding boxes generated by RPN, the RoI pooling layer [38] is adopted to extract feature maps from regions. The feature map passes through a residual block and an average pooling layer. Finally, a fully connected layer and softmax are added to classify the results into three classes (head, upper-body, and background) and output bounding box regression.
4.
Head and Upper-Body Pose Classification Module
The head and upper-body pose class can be considered as a subcategory in head and upper-body detection. Two fully connected layers are appended with four outputs (supine, prone, left side, right side). Each of the fully connected layers makes predictions for individual tasks.

3.2.2. SleePose-FRCNN-NET Training

The training DMTL network contains the ResNet50 backbone network and sub-networks. The multi-task loss function is used to train head and upper-body detection, head and upper-body pose classification, and bounding box regression. Here, the loss functions for each task are discussed.
1.
Head and Upper-Body Detection
This detection performs head and upper-body classification and uses categorical cross-entropy loss. The loss function is defined as the following Equation (3).
L c l a s s = i = 1 k y i log ( y ^ i ) ,
where y i is the ground truth and   y ^ is the predicted score for each class i .
2.
Head Pose Classification
Predicting the head pose is a multiple-class problem. The categorical cross-entropy is used as a loss function. The loss is considered only if the box is detected as a head. The loss function is defined as the following Equation (4).
L h e a d = i = 1 k y i log ( y ^ i ) ,
where y i is the ground truth pose class and   y ^ is the predicted score for the pose class.
3.
Upper-Body Pose Classification
Predicting the upper-body pose is a multiple-class problem. The categorical cross-entropy is used as a loss function. The loss is considered only if the box is detected as the upper body. The loss function is defined as the following Equation (5).
L b o d y = i = 1 k y i log ( y ^ i ) ,
where y i is the ground truth and   y ^ is the predicted score for each pose class i .
4.
Bounding Box Regression
The bounding box regression is used to tighten the bounding boxes for each identified region. For bounding box regression, the smooth L1 loss is used. The loss function is defined as the following Equation (6).
L b o x = i ϵ { x , y , w , h } k L 1 s m o o t h ( t i k t ^ i ) ,
where t k = ( t x k , t y k , t w k , t h k ) denotes four predicted parameterized coordinates for class k and t ^ i indicates ground truth coordinates.
5.
Overall Multi-Task Loss
The overall loss function is the weighted average of all the losses defined above. The overall loss function is defined as the following Equation (7).
L = λ c l a s s L c l a s s + λ h e a d L h e a d + λ b o d y L b o d y + λ b o x L b o x ,
where λ i , i { c l a s s ,   h e a d ,   b o d y ,   b o x } are loss weights to balance their contributions to the overall loss.
6.
Parameter Setting
The model was trained on an Nvidia GeForce RTX 2070 GPU using the Adam optimizer with a learning rate of 0.00001. All the tasks were trained end-to-end, with a validation-based early stopping, in order to avoid overfitting. The development environment is built on the Keras framework.

3.3. Data Augmentation

Data augmentation is used to expand the training dataset to reduce overfitting and improves generalization. For this study, data augmentation was performed when training the live streaming dataset, in the form of applying randomly rotating input images between 0 and 15 degrees, and randomly adjusting contrast and brightness of images.

3.4. Sleep Analysis—Posture Focused

Our system provides a pictorial representation of sleep postures and posture-related indicators that have been proven to be highly associated with sleep quality [39]. The indicators include (a) shifts in sleep posture, (b) the number of postures that last longer than 15 min, (c) average duration in postures, and (d) sleep efficiency (based on turning) [39]. Table 2 gives detailed descriptions of indicators of sleep quality.
Sleep postures and movements are associated with sleep quality. Our system classifies head and upper-body sleep postures into four clinical standard categories: supine, left, right, and prone.
Several studies reveal that the number of shifts in sleep posture, postures that last longer than 15 min, and nocturnal movements are related to lifestyle and insomnia symptoms. Generally, healthy people shift their posture between 10 and 30 times per night [40]. However, too much tossing and turning during the night indicates poor sleep quality. The duration in a particular posture has been associated with various medical conditions. Patients remaining in the same posture for long periods have an increased risk of pressure injury. Several clinical guidelines recommend that patients change posture at least every two hours [41].
Here, the synopsis of sleep posture and movement that can be assessed for further sleep measurement and study is provided.

4. Experimental Results and Analysis

4.1. Datasets

Several datasets were used to train and test our CNN model. The Simultaneously-Collected Multimodal Lying Pose (SLP) dataset [42] with annotated head and upper-body position was used for training and testing. The imLab@NTU Sleep Posture Dataset (iSP) dataset [43] was used, and a YouTube dataset was created for training and testing.

4.1.1. Simultaneously-Collected Multimodal Lying Pose (SLP)

Simultaneously-collected multimodal Lying Pose (SLP) is a large-scale in-bed pose dataset. The SLP dataset contains images from RGB and thermal cameras. RGB images were used for our study.
Images from 109 subjects were collected under three main categories: supine, left side, and right side. Example images of the SLP dataset are shown in Figure 3. The data from 102 subjects was collected in the living room and seven subjects in the hospital room. The samples were collected under various cover conditions: no covering, a thin sheet, and a thick blanket. Table 3 describes the details of the SLP dataset.

4.1.2. imLab@NTU Sleep Posture Dataset (iSP)

1.
Pilot Experiment
An experimental environment was set up for recording sleep postures. Microsoft Kinect was mounted at a horizontal distance of 50 cm and vertical distance of 55 cm above the bed, so that its field of view can cover the bed and entire body of a subject (Figure 4).
RGB and depth images were collected simultaneously. Only RGB images were used in this work, as shown in Figure 5. Thirty-six subjects participated in the experiment, of which 29 subjects were male and seven were female. Based on Sleep Assessment and Advisory Service (SAAS)’s research, 10 types of sleep postures were chosen, including the left side in fetal posture, left side in yearner posture, left side in log posture (LEFT set), right side in fetal posture, right side in yearner posture, right side in log posture (RIGHT set), supine in hand-up posture, supine in hand-down posture (SUPINE set), prone in hand-up posture, and prone in hand-down posture (PRONE set) as shown in Figure 6. The total time of the experiment of each subject was 10 min, and each subject mimicked each type of sleep posture lasting 1 min via manual guidelines. Sleep postures were collected under three cover conditions: no covering, a thin sheet, and a thick blanket. Table 4 describes the details of the iSP dataset.
2.
Real-Life Sleep Experiment
Four healthy adult subjects participated in the real-life study. The video was recorded at 20 frames per second from a home security camera during each subject’s sleep. The duration of each subject’s record and the number of frames used for evaluation are shown in Table 5. This dataset includes RGB images in day mode and IR images in night mode, as Figure 7 shows. A home security day-and-night camera was fit onto a custom-built mount that stood about 200 cm tall to the front of the bed. To increase posture variance and diversity, image frames were manually labeled and selected from minor movement and non-movement periods.

4.1.3. YouTube Dataset

With the growing popularity of video-sharing sites such as YouTube, many people continuously broadcast daily events in their life, including while sleeping. Four sleep streams were collected from YouTube, whose duration varied from 5 to 8 h. The dataset included RGB and IR images. It contained sleep data with various poses, coverings, and illumination. A sampling rate of one frame per second was used for our experiments and the redundant frames were discarded. The time of each sleep stream and the number of frames are presented in Table 6.

4.2. Evaluation on SLP Dataset

The model was trained on 4050 samples of the first 90 subjects recorded in the room and validated on 540 samples of the remaining 12 subjects. The model was tested on 315 samples recorded in the hospital. As the SLP dataset is not annotated for head and upper-body detection, ground truth labels for head and upper-body regions were annotated manually.
The mean average precision (mAP) result of head and upper-body detection is presented in Table 7. As can be seen from the table, our method is more robust to environmental changes than the YOLOv3 [44] and YOLOv4 [45] methods. Our approach applies whether the image is captured in the room or the hospital.
For head and upper-body pose classification, our method was compared with Akbarian et al.’s work [25], Mohammadi et al.’s work [46], and the Inception network [47], which was used to extract deep features in the study of Torres et al. [48] for the classification of sleep poses. Given that the above methods cannot detect heads and upper bodies, detections from the SleePose-FRCNN-Net are provided. Table 8 and Table 9 present the results of our approach and the other methods on the SLP dataset. As can be seen, our method achieves the highest accuracy. Confusion matrices of our models are shown in Figure 8 and Figure 9 for head and upper-body poses, respectively.

4.3. Evaluation on the iSP Dataset

4.3.1. iSP Pilot Dataset

Next, the model was trained and tested on the iSP pilot dataset. In this experiment, 15,000 samples from 25 subjects were used to train the model, and the remaining 6600 samples from 11 subjects were used to test the model. The model was trained and tested with three covering conditions. The detection results in terms of mAP are presented in Table 10. The average processing time of SleePose-FRCNN-Net is about 0.7 s. Although the processing time by our method is longer than YOLOv3 (0.43 s), our accuracy is higher than YOLOv3′s accuracy, as shown in Table 7 and Table 10.
  • Comparison with Posture Classification
Head and upper-body classifications were evaluated on the iSP pilot dataset. Our model was compared to the current state-of-the-art sleep posture classification model. The results can be seen in Table 11 and Table 12. The results show that our model achieves higher accuracy in three cover conditions. As illustrated in the confusion matrix in Figure 10 and Figure 11, incorrect predictions occur rarely.
2.
Comparison with General Human Pose Estimation
Pose estimation results were compared with SimpleBaseline [49] and OpenPose [50], which is the most classical pose estimation model and is used to recognize the types of sleep postures in the study of Lyu et al. [24]. Although these models have satisfactory performance in the general case, they do not perform well in the iSP dataset. This is likely because some body parts of an individual are occluded by a blanket. For example, Figure 12 illustrates that SimpleBaseline and OpenPose do not detect the individual’s body parts under the covering.

4.3.2. Real-Life Sleep Experiment

Head and upper-body classifications were evaluated on the iSP real-life dataset. Training and testing sets were chosen from different fragments of sleep. The data ratio is 80% and 20% for training and testing sets, respectively. After training, the model obtains an overall classification accuracy of 91.67% for detecting the upper-body sleep postures and 94.37% for detecting the head sleep postures.
Experimental results demonstrate that the proposed model is applicable to the real environment. Our model can handle sleep posture detection and classification with RGB and IR images, as shown in Figure 13.

4.4. Evaluation on YouTube Dataset

The individual model was trained using the data from each subject. Training and testing sets were chosen from different fragments of sleep. The data ratio is 80% and 20% for training and testing sets. The detection results in terms of mAP are presented in Table 13. The averaged accuracy of each subject is presented in Table 14 and Table 15. Across four subjects, our method achieves an overall classification accuracy of 96.92% for detecting the four main sleep postures: supine, prone, right, and left.
The ability of our system to detect both sleep posture and movement is exhibited with a pictorial representation, as shown in Figure 14. The circular diagram shows the state of 12 h. It is divided into segments to represent the duration of different sleep poses. A grey area indicates a waking state. The diagram shows that subject S2 had few body movements, few shifts in sleep posture, high sleep efficiency, and few awakenings.
Our system provides posture-related indicators of sleep quality, as presented in Table 16. According to research, a higher percentage of time in bed with fewer shifts indicates more efficient sleep.

4.5. Computational Performance

The average processing time per image by our method is about 0.7 s. The experiment was performed on the desktop computer explained in Section 3.2.2. In addition, our algorithm was implemented in python using the Keras mechanism. Python is a slow programming language because of its dynamically typed feature and limited to using a single processor during its application. Therefore, if our algorithm is implemented in the C language in the future, the processing time will be significantly reduced with the same accuracy.

5. Discussion

The paper presents a new approach to sleep posture monitoring based on a deep multi-task learning network. The innovation of this work is integrating head detection, upper-body detection, head pose classification, and upper-body pose classification into one sleep monitoring system. The upper-body pose and head pose information during sleep are obtained simultaneously. The proposed method was evaluated using laboratory-based simulation datasets such as the SLP, iSP, and real-life datasets such as YouTube, and both achieved impressive results. The work demonstrates practical value because the system uses a single 2D IR video camera and applies to various covering conditions and variable illumination.
Most of the existing techniques for sleep posture monitoring only focus on posture classification methods; however, the proposed method combines head detection, upper-body detection, head pose classification, and upper-body pose classification into one united framework. The proposed method is more robust and accurate than the methods used in other papers [25,46,47], which use a deep learning classification model for sleep monitoring. The proposed system takes a video feed from an IR camera and analyzes the video stream. The ability of the proposed approach to accurately detect both upper-body and head postures provides valuable information for sleep studies.
Although our method was evaluated on adults, the proposed method can be applied to babies and children using transfer learning based on a pre-trained network. The model trained on the adult data is initialed, then finetunes that model on the infant and children dataset.
To analyze the generalization ability of the proposed model, the model was trained on the SLP training dataset and tested on the iSP pilot testing dataset. The model obtains an average classification accuracy of 89.17% for classifying the upper-body sleep postures and 91.25% for classifying the head sleep postures. Table 17 shows that our model generalizes well on previously unseen, new data. In addition, the camera placement and monitoring zone have a significant impact on the overall performance of video analysis. For posture detection and classification, to increase confidence in the system, it is recommended that a camera is installed at the 2–2.5 m height. The whole body can be covered in the view. It is necessary to avoid the extreme angle view (from the side).
This research, however, is subject to several limitations. Due to lack of manpower and labeled sleep posture data, a small amount of data in the real-life dataset was used in this study. A large amount of labeled data is needed to adapt to challenging environments, such as variations in human appearance and arbitrary camera viewpoints. This limitation could be mitigated with a semi-automatic annotation tool in the future.

6. Conclusions

This paper presented a non-contact video-based framework for simultaneous head and upper-body detection and pose classification during sleep. All detections and classifications were trained and tested simultaneously in a single multi-task network. Experimental investigations on three available datasets show that the proposed system can be applied reliably to extensive public sleep data with various covering conditions and is robust to real-life overnight sleep data. The real-life application achieves a high accuracy of 91.7% in upper-body pose classification. Furthermore, a sleep posture and movement synopsis is provided to assess sleep quality and irregular sleeping habits.

Author Contributions

Conceptualization, Y.-Y.L., S.-J.W. and Y.-P.H.; methodology, Y.-Y.L., S.-J.W. and Y.-P.H.; software, Y.-Y.L.; writing—original draft preparation, Y.-Y.L.; writing—review and editing, S.-J.W. and Y.-P.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Huang, T.; Redline, S. Cross-sectional and Prospective Associations of Actigraphy-Assessed Sleep Regularity with Metabolic Abnormalities: The Multi-Ethnic Study of Atherosclerosis. Diabetes Care 2019, 42, 1422–1429. [Google Scholar] [CrossRef]
  2. Huang, T.; Mariani, S.; Redline, S. Sleep Irregularity and Risk of Cardiovascular Events: The Multi-Ethnic Study of Atherosclerosis. J. Am. Coll. Cardiol. 2020, 75, 991–999. [Google Scholar] [CrossRef] [PubMed]
  3. Yu, C.K.C. The effect of sleep position on dream experiences. Dreaming 2012, 22, 212–221. [Google Scholar] [CrossRef]
  4. Neilson, J.; Avital, L.; Willock, J.; Broad, N. Using a national guideline to prevent and manage pressure ulcers. Nurs. Manag. 2014, 21, 18–21. [Google Scholar] [CrossRef] [Green Version]
  5. Hoyer, E.H.; Friedman, M.; Lavezza, A.; Wagner-Kosmakos, K.; Lewis-Cherry, R.; Skolnik, J.L.; Byers, S.P.; Atanelov, L.; Colantuoni, E.; Brotman, D.J.; et al. Promoting mobility and reducing length of stay in hospitalized general medicine patients: A quality-improvement project. J. Hosp. Med. 2016, 11, 341–347. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. van Maanen, J.P.; Richard, W.; van Kesteren, E.R.; Ravesloot, M.J.; Laman, D.M.; Hilgevoord, A.A.; de Vries, N. Evaluation of a new simple treatment for positional sleep apnoea patients. J. Sleep Res. 2012, 21, 322–329. [Google Scholar] [CrossRef]
  7. Cary, D.; Briffa, K.; McKenna, L. Identifying relationships between sleep posture and non-specific spinal symptoms in adults: A scoping review. BMJ Open 2019, 9, e027633. [Google Scholar] [CrossRef]
  8. Kwasnicki, R.M.; Cross, G.W.V.; Geoghegan, L.; Zhang, Z.; Reilly, P.; Darzi, A.; Yang, G.Z.; Emery, R. A lightweight sensing platform for monitoring sleep quality and posture: A simulated validation study. Eur. J. Med. Res. 2018, 23, 28. [Google Scholar] [CrossRef]
  9. Yoon, D.W.; Shin, H.W. Sleep Tests in the Non-Contact Era of the COVID-19 Pandemic: Home Sleep Tests Versus In-Laboratory Polysomnography. Clin. Exp. Otorhinolaryngol. 2020, 13, 318–319. [Google Scholar] [CrossRef]
  10. van Kesteren, E.R.; van Maanen, J.P.; Hilgevoord, A.A.; Laman, D.M.; de Vries, N. Quantitative effects of trunk and head position on the apnea hypopnea index in obstructive sleep apnea. Sleep 2011, 34, 1075–1081. [Google Scholar] [CrossRef]
  11. Zhu, K.; Bradley, T.D.; Patel, M.; Alshaer, H. Influence of head position on obstructive sleep apnea severity. Sleep Breath. 2017, 21, 821–828. [Google Scholar] [CrossRef] [PubMed]
  12. Levendowski, D.J.; Gamaldo, C.; St Louis, E.K.; Ferini-Strambi, L.; Hamilton, J.M.; Salat, D.; Westbrook, P.R.; Berka, C. Head Position During Sleep: Potential Implications for Patients with Neurodegenerative Disease. J. Alzheimer’s Dis. 2019, 67, 631–638. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. De Koninck, J.; Gagnon, P.; Lallier, S. Sleep positions in the young adult and their relationship with the subjective quality of sleep. Sleep 1983, 6, 52–59. [Google Scholar] [CrossRef]
  14. Clever, H.M.; Kapusta, A.; Park, D.; Erickson, Z.; Chitalia, Y.; Kemp, C. 3d human pose estimation on a configurable bed from a pressure image. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Madrid, Spain, 1–5 October 2018; pp. 54–61. [Google Scholar]
  15. Liu, S.; Ostadabbas, S. Seeing Under the Cover: A Physics Guided Learning Approach for In-Bed Pose Estimation. In Proceedings of the International Conference on Medical Image Computing and Computer Assisted Intervention, Shenzhen, China, 13–17 October 2019; pp. 236–245. [Google Scholar]
  16. Grimm, T.; Martinez, M.; Benz, A.; Stiefelhagen, R. Sleep position classification from a depth camera using bed aligned maps. In Proceedings of the International Conference on Pattern Recognition, Cancun, Mexico, 4–8 December 2016. [Google Scholar]
  17. Chang, M.-C.; Yu, T.; Duan, K.; Luo, J.; Tu, P.; Priebe, M.; Wood, E.; Stachura, M. In-bed patient motion and pose analysis using depth videos for pressure ulcer prevention. In Proceedings of the IEEE International Conference on Image Processing, Beijing, China, 17–20 September 2017; pp. 4118–4122. [Google Scholar]
  18. Li, Y.Y.; Lei, Y.J.; Chen, L.C.L.; Hung, Y.P. Sleep Posture Classification with Multi-Stream CNN using Vertical Distance Map. In Proceedings of the International Workshop on Advanced Image Technology, Chiang Mai, Thailand, 7–9 January 2018. [Google Scholar]
  19. Klishkovskaia, T.; Aksenov, A.; Sinitca, A.; Zamansky, A.; Markelov, O.A.; Kaplun, D. Development of Classification Algorithms for the Detection of Postures Using Non-Marker-Based Motion Capture Systems. Appl. Sci. 2020, 10, 4028. [Google Scholar] [CrossRef]
  20. Sarbolandi, H.; Lefloch, D.; Kolb, A. Kinect range sensing: Structured-light versus Time-of-Flight Kinect. Comput. Vis. Image Underst. 2015, 139, 1–20. [Google Scholar] [CrossRef] [Green Version]
  21. Khoshelham, K.; Elberink, S.O. Accuracy and Resolution of Kinect Depth Data for Indoor Mapping Applications. Sensors 2012, 12, 1437–1454. [Google Scholar] [CrossRef] [Green Version]
  22. Choe, J.; Montserrat, D.M.; Schwichtenberg, A.J.; Delp, E.J. Sleep Analysis Using Motion and Head Detection. In Proceedings of the IEEE Southwest Symposium on Image Analysis and Interpretation, Las Vegas, NV, USA, 8–10 April 2018; pp. 29–32. [Google Scholar]
  23. Chen, K.; Gabriel, P.; Alasfour, A.; Gong, C.; Doyle, W.K.; Devinsky, O.; Friedman, D.; Dugan, P.; Melloni, L.; Thesen, T.; et al. Patient-specific pose estimation in clinical environments. IEEE J. Transl. Eng. Health Med. 2018, 6, 2101111. [Google Scholar] [CrossRef]
  24. Liu, S.; Yin, Y.; Ostadabbas, S. In-bed pose estimation: Deep learning with shallow dataset. IEEE J. Transl. Eng. Health Med. 2019, 7, 4900112. [Google Scholar] [CrossRef]
  25. Akbarian, S.; Delfi, G.; Zhu, K.; Yadollahi, A.; Taati, B. Automated Noncontact Detection of Head and Body Positions During Sleep. IEEE Access 2019, 7, 72826–72834. [Google Scholar] [CrossRef]
  26. Akbarian, S.; Montazeri Ghahjaverestan, N.; Yadollahi, A.; Taati, B. Distinguishing Obstructive Versus Central Apneas in Infrared Video of Sleep Using Deep Learning: Validation Study. J. Med. Internet Res. 2020, 22, e17252. [Google Scholar] [CrossRef]
  27. Lyu, H.; Tian, J. Skeleton-Based Sleep Posture Recognition with BP Neural Network. In Proceedings of the IEEE 3rd International Conference on Computer and Communication Engineering Technology, Beijing, China, 14–16 August 2020; pp. 99–104. [Google Scholar]
  28. Afham, M.; Haputhanthri, U.; Pradeepkumar, J.; Anandakumar, M.; De Silva, A. Towards Accurate Cross-Domain In-Bed Human Pose Estimation. arXiv 2021, arXiv:2110.03578. [Google Scholar]
  29. Mohammadi, S.M.; Enshaeifar, S.; Hilton, A.; Dijk, D.-J.; Wells, K. Transfer Learning for Clinical Sleep Pose Detection Using a Single 2D IR Camera. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 29, 290–299. [Google Scholar] [CrossRef] [PubMed]
  30. Zhang, Y.; Yang, Q. A survey on multi-task learning. arXiv 2018, arXiv:1707.08114. [Google Scholar] [CrossRef]
  31. Kokkinos, I. UberNet: Training a universal convolutional neural network for low-, mid-, and high-level vision using diverse datasets and limited memory. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  32. Li, T.; Xu, Z. Simultaneous Face Detection and Head Pose Estimation: A Fast and Unified Framework. In Proceedings of the Asian Conference on Computer Vision, Perth, Australia, 2–6 December 2018. [Google Scholar]
  33. Piriyajitakonkij, M.; Warin, P.; Lakhan, P.; Leelaarporn, P.; Pianpanit, T.; Niparnan, N.; Mukhopadhyay, S.C.; Wilaiprasitporn, T. SleepPoseNet: Multi-View Multi-Task Learning for Sleep Postural Transition Recognition Using UWB. arXiv 2020, arXiv:2005.02176. [Google Scholar] [CrossRef]
  34. Barnich, O.; Van Droogenbroeck, M. ViBe: A universal background subtraction algorithm for video sequences. IEEE Trans. Image Process. 2011, 20, 1709–1724. [Google Scholar] [CrossRef] [Green Version]
  35. Ren, S.; He, K.; Ross, G.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. In Advances in Neural Information Processing Systems, Proceedings of the 28th Conference on Neural Information Processing Systems, Montreal, Canada, 7–12 December 2015; MIT Press: Cambridge, MA, USA, 2015; pp. 91–99. [Google Scholar]
  36. Huang, J.; Rathod, V.; Sun, C.; Zhu, M.; Korattikara, A.; Fathi, A.; Fischer, I.; Wojna, Z.; Song, Y.; Guadarrama, S.; et al. Speed/accuracy trade-offs for modern convolutional object detectors. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3296–3305. [Google Scholar]
  37. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  38. Girshick, R. Fast R-CNN. In Proceedings of the International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015. [Google Scholar]
  39. Wrzus, C.; Brandmaier, A.M.; von Oertzen, T.; Müller, V.; Wagner, G.G.; Riediger, M. A new approach for assessing sleep duration and postures from ambulatory accelerometry. PLoS ONE 2012, 7, e48089. [Google Scholar] [CrossRef] [Green Version]
  40. De Koninck, J.; Lorrain, D.; Gagnon, P. Sleep positions and position shifts in five age groups: An ontogenetic picture. Sleep 1992, 15, 143–149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  41. Lyder, C.H. Pressure ulcer prevention and management. JAMA 2003, 289, 223–226. [Google Scholar] [CrossRef]
  42. Liu, S.; Huang, X.; Fu, N.; Li, C.; Su, Z.; Ostadabbas, S. Simultaneously-collected multimodal lying pose dataset: Towards in-bed human pose monitoring under adverse vision conditions. arXiv 2020, arXiv:2008.08735. [Google Scholar]
  43. Chen, L.C.L. On Detection and Browsing of Sleep Events. Ph.D. Thesis, National Taiwan University, Taipei City, Taiwan, 2016. [Google Scholar]
  44. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  45. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  46. Mohammadi, S.M.; Alnowami, M.; Khan, S.; Dijk, D.J.; Hilton, A.; Wells, K. Sleep Posture Classification using a Convolutional Neural Network. In Proceedings of the Conference of the IEEE Engineering in Medicine and Biology Society, Honolulu, HI, USA, 17–21 July 2018; pp. 1–4. [Google Scholar]
  47. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  48. Torres, C.; Fried, J.; Rose, K.; Manjunath, B. Deep EYE-CU (decu): Summarization of patient motion in the ICU. In Proceedings of the European Conference on Computer Vision Workshops, Amsterdam, The Netherlands, 8–16 October 2016; pp. 178–194. [Google Scholar]
  49. Xiao, B.; Wu, H.; Wei, Y. Simple baselines for human pose estimation and tracking. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 472–487. [Google Scholar]
  50. Cao, Z.; Hidalgo, G.; Simon, T.; Wei, S.-E.; Sheikh, Y. Openpose: Realtime multi-person 2d pose estimation using part affinity fields. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 172–186. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Architecture of the proposed framework.
Figure 1. Architecture of the proposed framework.
Sensors 22 02014 g001
Figure 2. SleePose-FRCNN-Net architecture.
Figure 2. SleePose-FRCNN-Net architecture.
Sensors 22 02014 g002
Figure 3. Sample images for each pose class from the SLP database [42].
Figure 3. Sample images for each pose class from the SLP database [42].
Sensors 22 02014 g003
Figure 4. Illustration of the system in the experiment [43].
Figure 4. Illustration of the system in the experiment [43].
Sensors 22 02014 g004
Figure 5. Examples of images from the iSP pilot dataset [43].
Figure 5. Examples of images from the iSP pilot dataset [43].
Sensors 22 02014 g005
Figure 6. The 10 types of sleep postures [43].
Figure 6. The 10 types of sleep postures [43].
Sensors 22 02014 g006
Figure 7. Examples of RGB/IR images from the iSP real-life dataset.
Figure 7. Examples of RGB/IR images from the iSP real-life dataset.
Sensors 22 02014 g007
Figure 8. Confusion matrix for head pose classification on the SLP dataset.
Figure 8. Confusion matrix for head pose classification on the SLP dataset.
Sensors 22 02014 g008
Figure 9. Confusion matrix for upper-body pose classification on the SLP dataset.
Figure 9. Confusion matrix for upper-body pose classification on the SLP dataset.
Sensors 22 02014 g009
Figure 10. Confusion matrix for head pose classification on the iSP pilot dataset.
Figure 10. Confusion matrix for head pose classification on the iSP pilot dataset.
Sensors 22 02014 g010
Figure 11. Confusion matrix for upper-body pose classification on the iSP pilot dataset.
Figure 11. Confusion matrix for upper-body pose classification on the iSP pilot dataset.
Sensors 22 02014 g011
Figure 12. Repurposed SimpleBaseline [49] (first row) and OpenPose [50] (second row) performance on the iSP dataset.
Figure 12. Repurposed SimpleBaseline [49] (first row) and OpenPose [50] (second row) performance on the iSP dataset.
Sensors 22 02014 g012
Figure 13. Sleep posture detection on real-life data. The left column shows images in IR mode and the right column shows in RGB mode.
Figure 13. Sleep posture detection on real-life data. The left column shows images in IR mode and the right column shows in RGB mode.
Sensors 22 02014 g013
Figure 14. Pictorial representation of sleep poses and motion over time of subjects.
Figure 14. Pictorial representation of sleep poses and motion over time of subjects.
Sensors 22 02014 g014
Table 1. Comparison of previous studies and the proposed method on sleep posture classification.
Table 1. Comparison of previous studies and the proposed method on sleep posture classification.
SensorMethodDataset UsedAdvantagesLimitations
Pressure-sensing mat3D human pose estimation based on deep learning method [14]Simulation datasetA pressure-sensing mat is robust to covering.A pressure-sensing mat has high cost
and complex maintenance for home use.
Thermal cameraHuman pose estimation based on deep learning method [15]Simulation datasetA thermal camera is robust to illuminationchanges and covering.A thermal camera has high cost for home use.
Depth cameraSleep posture classification
based on deep learning method [16,18]
Simulation datasetA depth camera is robust to low light intensity.
-
Depth measurements of depth cameras suffer from various noise factors.
-
The depth camera is not prevalent at home surveillance.
Infrared cameraSleep vs. wake states detection in young children based on motion analysis [22]Real sleep data
-
A single infrared camera is convenient and low-cost.
-
A general-purpose head detector is used.
This method only succeeds for 50% of nights.
Human pose estimation based on deep learning method (OpenPose) [24,27]Simulation datasetThe method can extract features of the skeleton effectively.Their data [24] is obtained from mannequins in a simulated hospital room, and this method cannot perform well on real data [25].
Sleep posture classificationbased on deep learning method [25,29]Simulation datasetThe deep learning method can achieve good accuracy.It focuses on classifying posture without detecting the upper-body and head region.
Sleep posture detection and classification based on deep learning method (proposed method)Simulation and real sleep datasetA unified framework for simultaneously detecting and classifying upper-body pose and head pose is proposed.Training personal data to learn CNN is required.
Table 2. Indicators of sleep quality.
Table 2. Indicators of sleep quality.
IndicatorDescriptionUnit
Shifts in sleep postureCount of posture changesCount
Number of postures
that last longer than 15 min
Count of postures with duration longer than 15 minCount
Average duration in a postureMean of the posture durationMinutes
Sleep efficiencyPercentage of time without turningPercentage
Table 3. The details of the SLP dataset.
Table 3. The details of the SLP dataset.
DatasetSleep Postures CategoriesCovering ConditionSetRecorded EnvironmentNumber of SubjectsNumber of Images per Covering Condition
SLP [42]Supine, left side, and right sideNo covering, thin sheet, and thick blanketTrainLiving room904050
ValidationLiving room12540
TestHospital room7315
Table 4. The details of the iSP pilot dataset.
Table 4. The details of the iSP pilot dataset.
DatasetSleep Postures CategoriesCovering ConditionSetRecorded EnvironmentNumber of SubjectsNumber of Images per Covering Condition
iSP [43]Supine, left side, right side, and proneNo covering,
thin sheet,
and thick blanket
TrainLab2515,000
TestLab116600
Table 5. The iSP real-life dataset summary.
Table 5. The iSP real-life dataset summary.
SubjectHoursNumber of Frames
S18576,000
S26432,000
S37504,000
S48576,000
Table 6. YouTube dataset summary.
Table 6. YouTube dataset summary.
SubjectHoursNumber of Frames
S1828,800
S2518,800
S3—Day 1828,800
S3—Day 27.527,000
Table 7. Results of detection on the SLP dataset.
Table 7. Results of detection on the SLP dataset.
ModelmAP
(No Covering)
mAP
(Thin Sheet)
mAP
(Thick Blanket)
ValTestValTestValTest
YOLOv3 [44]99.2696.8499.5869.3099.7064.34
YOLOv4 [45]98.6795.5999.5491.2799.2983.84
SleePose-FRCNN-Net99.8299.8699.9799.9199.6492.31
Table 8. Results of head pose classification on the SLP dataset.
Table 8. Results of head pose classification on the SLP dataset.
ModelAccuracy
(No Covering)
Accuracy
(Thin Sheet)
Accuracy
(Thick Blanket)
ValTestValTestValTest
Akbarian [25]95.6187.5096.7083.6697.0085.52
Mohammadi [46]85.5678.4779.4271.9085.1971.03
Inception [47]85.7778.4779.6172.2288.2070.69
SleePose-FRCNN-Net99.2596.1599.0790.4898.7092.70
Table 9. Results of upper-body pose classification on the SLP dataset.
Table 9. Results of upper-body pose classification on the SLP dataset.
ModelAccuracy
(No Covering)
Accuracy
(Thin Sheet)
Accuracy
(Thick Blanket)
ValTestValTestValTest
Akbarian [25]98.2785.0696.2988.5696.1988.64
Mohammadi [46]85.3677.6091.8473.2088.5768.18
Inception [47]86.5166.1289.9882.2088.5764.29
SleePose-FRCNN-Net99.4495.2499.0791.1198.7091.11
Table 10. Results of detection on the iSP pilot dataset.
Table 10. Results of detection on the iSP pilot dataset.
ModelmAP
(No Covering)
mAP
(Thin Sheet)
mAP
(Thick Blanket)
YOLOv3 [44]97.1193.6395.38
YOLOv4 [45]99.6194.1198.59
SleePose-FRCNN-Net99.8294.6799.64
Table 11. Results of head pose classification on the iSP pilot dataset.
Table 11. Results of head pose classification on the iSP pilot dataset.
ModelAccuracy
(No Covering)
Accuracy
(Thin Sheet)
Accuracy
(Thick Blanket)
Akbarian [25]87.6292.1091.80
Mohammadi [46]77.9074.8271.05
Inception [47]83.9584.1678.61
SleePose-FRCNN-Net96.6597.3394.58
Table 12. Results of upper-body pose classification on the iSP pilot dataset.
Table 12. Results of upper-body pose classification on the iSP pilot dataset.
ModelAccuracy
(No Covering)
Accuracy
(Thin Sheet)
Accuracy
(Thick Blanket)
Akbarian [25]87.6292.1091.80
Mohammadi [46]77.9074.8271.05
Inception [47]83.9584.1678.61
SleePose-FRCNN-Net97.7998.3394.72
Table 13. Results of detection on the YouTube dataset.
Table 13. Results of detection on the YouTube dataset.
MethodmAP
S1S2S3—Day 1S3—Day 2
SleePose-FRCNN-Net96.3399.5799.6696.30
Table 14. Results of head pose classification on the YouTube dataset.
Table 14. Results of head pose classification on the YouTube dataset.
MethodAccuracy
S1S2S3—Day 1S3—Day 2
SleePose-FRCNN-Net89.8790.6795.4492.38
Table 15. Results of upper-body pose classification on the YouTube dataset.
Table 15. Results of upper-body pose classification on the YouTube dataset.
MethodAccuracy
S1S2S3—Day 1S3—Day 2
SleePose-FRCNN-Net98.2196.1097.6298.25
Table 16. Sleep indicators of subjects.
Table 16. Sleep indicators of subjects.
Subject
S1S2S3—Day1S3—Day2
Shifts in sleep posture (n/hour)1.631.002.382.53
Number of postures that last longer than 15 min (n/hour)1.381.201.501.75
Average duration in a posture (min)19.2060.0025.2623.68
Sleep efficiency (percentage)89.3896.0091.6783.56
Table 17. Results of the proposed model that was trained on the SLP dataset and tested on the iSP pilot dataset.
Table 17. Results of the proposed model that was trained on the SLP dataset and tested on the iSP pilot dataset.
PoseAccuracy
(No Covering)
Accuracy
(Thin Sheet)
Accuracy
(Thick Blanket)
Head92.5794.7486.44
Upper-Body92.9291.1083.51
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, Y.-Y.; Wang, S.-J.; Hung, Y.-P. A Vision-Based System for In-Sleep Upper-Body and Head Pose Classification. Sensors 2022, 22, 2014. https://doi.org/10.3390/s22052014

AMA Style

Li Y-Y, Wang S-J, Hung Y-P. A Vision-Based System for In-Sleep Upper-Body and Head Pose Classification. Sensors. 2022; 22(5):2014. https://doi.org/10.3390/s22052014

Chicago/Turabian Style

Li, Yan-Ying, Shoue-Jen Wang, and Yi-Ping Hung. 2022. "A Vision-Based System for In-Sleep Upper-Body and Head Pose Classification" Sensors 22, no. 5: 2014. https://doi.org/10.3390/s22052014

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop