Next Article in Journal
Sensitive pH Monitoring Using a Polyaniline-Functionalized Fiber Optic—Surface Plasmon Resonance Detector
Previous Article in Journal
Dynamic Spectrum Sharing for Future LTE-NR Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Depth-Camera Based Energy Expenditure Estimation System for Physical Activity Using Posture Classification Algorithm

1
Department of Computer Science and Information Engineering, National Taipei University, New Taipei City 237303, Taiwan
2
College of Electrical Engineering and Computer Science, National Taipei University, New Taipei City 237303, Taiwan
3
Department of Computer Science and Information Engineering, National Taiwan University of Science and Technology, Taipei 106335, Taiwan
4
Department of Computer Science and Information Engineering, Tamkang University, New Taipei City 251301, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(12), 4216; https://doi.org/10.3390/s21124216
Submission received: 25 May 2021 / Revised: 14 June 2021 / Accepted: 18 June 2021 / Published: 19 June 2021
(This article belongs to the Section Wearables)

Abstract

:
Insufficient physical activity is common in modern society. By estimating the energy expenditure (EE) of different physical activities, people can develop suitable exercise plans to improve their lifestyle quality. However, several limitations still exist in the related works. Therefore, the aim of this study is to propose an accurate EE estimation model based on depth camera data with physical activity classification to solve the limitations in the previous research. To decide the best location and amount of cameras of the EE estimation, three depth cameras were set at three locations, namely the side, rear side, and rear views, to obtain the kinematic data and EE estimation. Support vector machine was used for physical activity classification. Three EE estimation models, namely linear regression, multilayer perceptron (MLP), and convolutional neural network (CNN) models, were compared and determined the model with optimal performance in different experimental settings. The results have shown that if only one depth camera is available, optimal EE estimation can be obtained using the side view and MLP model. The mean absolute error (MAE), mean square error (MSE), and root MSE (RMSE) of the classification results under the aforementioned settings were 0.55, 0.66, and 0.81, respectively. If higher accuracy is required, two depth cameras can be set at the side and rear views, the CNN model can be used for light-to-moderate activities, and the MLP model can be used for vigorous activities. The RMSEs for estimating the EEs of standing, walking, and running were 0.19, 0.57, and 0.96, respectively. By applying the different models on different amounts of cameras, the optimal performance can be obtained, and this is also the first study to discuss the issue.

1. Introduction

Sedentary lifestyles are common in modern society. Insufficient physical activity increases the risk of noncommunicable diseases (NCDs) [1], such as cardiovascular diseases, respiratory diseases, cancers, stroke, and diabetes. Studies have indicated that NCDs account for more than 70% of global deaths [2]. To address the problem of insufficient physical activity, physical activities must be quantified for the design of appropriate exercise plans. However, some elderly are unable to go outside to meet the basic daily requirements of physical activities because of certain healthy issues. For those elderly who can only stay at home, it is necessary to estimate their EE of indoor activities and ensure they meet the basic requirement of physical activity. Energy expenditure (EE) is one factor used for quantifying physical activities. Conventionally, people use the portable metabolic analyzer to measure EE during exercise. The common commercial models of portable metabolic analyzers include COSMED K 4 b 2 [3], PNOĒ, and METAMAX® 3B, and the COSMED K 4 b 2 system is the most common system for measuring EE. Although the portable metabolic analyzers provide the most direct and accurate method for measuring calorie consumption, users of the system find it inconvenient when performing physical activities because they must carry an instrument and wear an oxygen mask during exercise.
Several studies have proposed systems using various sensors to map the EE measured by metabolic analyzers. Sensor systems can be divided into two types: contact-based [4,5,6,7] and noncontact [8,9,10,11,12]. In a contact-based sensor system, users must wear devices on body segments to measure EE. Most related research on this sensor system has adopted inertial measurement units (IMUs) for data collection. Altini et al. [4] adopted a customized platform that combined electrocardiogram sensors and accelerometers to obtain data and measure EE when users performed different physical activities in daily living. Three EE estimation methods were compared in the aforementioned study. The sensor number and the positioning of each sensor were also evaluated. Cvetković et al. [5] proposed an approach for estimating the EE of various activities in different scenarios. In addition to evaluating the EE, they analyzed the number and contributions of various sensors. Park et al. [6] evaluated the EE of six activities in daily living by combining various sensors, including an IMU and electrocardiogram sensor. Three feature selection methods were combined with the linear support vector machine (SVM) to find the best input feature set. The best feature set was tested in various models. The activity-specific approach was also adopted to improve the accuracy of EE estimation. However, the activities were only classified into two types, including static and dynamic activities. It is also not convenient to wear an IMU and electrocardiogram sensor for EE estimation. Hedegaard et al. [7] proposed an EE estimation system for seven activities during which 17 IMU sensors must be worn on body segments. Several parameters, such as gender, heart rate, acceleration, center of body weight, were input into multiple linear regression to estimate EE. Although this system can accurately estimate the EE of seven activities, wearing the clothing with 17 sensors is inconvenient. The main advantage of contact-based systems is their suitability for outdoor activities. However, wearing many sensors when performing physical activities is impractical. Sensor positioning greatly affects the accuracy of EE estimation. Moreover, the battery life of wearable devices is a critical challenge. Noncontact approaches have been proposed to solve the problems of using wearable devices to estimate EE. Kim et al. [8] developed a system based on Doppler radar to estimate the EE of upper-limb motions with a regression model. However, the average error of the aforementioned system for EE estimation was 12.75%, which is high. The other limitation of the aforementioned system is its inability to capture vertical motions, which is essential for estimating the EE of running activity. Moreover, setting a Doppler radar sensor is difficult in home settings.
Another noncontact approach for estimating EE involves a depth camera or traditional camera as a data collection tool. Yang et al. [9] proposed an EE estimation system based on smartphone cameras for workouts. The system can accurately count the repetitions and estimate the EE of four types of workouts: sit-ups, push-ups, jumping jacks, and squats. However, the system provides quantification for only the four aforementioned workouts, which are uncommon in daily living. Koporec et al. [10] developed a noncontact method for EE estimation by using a camera and classical image processing approaches. However, the error of EE estimation in the aforementioned study was high. With the development of artificial intelligence, deep learning approaches have also been adopted for EE estimation. Na et al. [11] proposed a deep-learning-based method for estimating the EE of different levels of physical activities. A convolutional neural network (CNN)-based method was used in their study for EE prediction [13]. The aforementioned research is representative of the applications of deep learning algorithms to EE estimation; however, the EEs of only light physical activities were estimated in the aforementioned study.
Lin et al. [12] proposed an EE estimation system by using Kinect depth cameras. The cameras were installed at three locations, namely the rear view, rear side view, and side view. Moreover, five regression-based models were compared to determine the model with the highest accuracy. Lin et al. estimated the EEs of only moderate-to-vigorous activities. Light activities, such as standing, were not considered; however, standing is an everyday physical activity. Another limitation of the aforementioned study was the high computational time required because of the complicated feature extraction process.
The proposed study can be regarded as an extension of the study of Lin et al. The motivation of the proposed study is to propose a Kinect-based EE estimation system to solve the problems and limitations associated with the previous research of Lin et al. This study investigated if applying the physical activity classification algorithm to the EE estimation system can improve the accuracy of EE estimation, which has not been confirmed in previous studies. Moreover, the proposed study investigated the best model for each type of physical activity for different camera settings. The study adopted a physical activity recognition algorithm to classify the postures captured by Kinect depth cameras into three activities, namely standing, walking, and running. Only 18 velocity-based features were used for EE estimation in this study to lower the computational complexity for feature extraction and selection. Three types of models were adopted to develop an independent model for each activity, and the model with the best performance for each activity was selected. The proposed system hypothesized that using the posture classification algorithm can improve the accuracy of EE estimation and is expected to obtain the optimal model for each physical activity for different camera settings.

2. Methodology

Figure 1 depicts the system architecture used in this study. The system was used to perform five tasks: data acquisition, data preprocessing, feature extraction, physical activity classification, and EE estimation. The performance of three models, namely linear regression (LR), multilayer perceptron (MLP), and convolutional neural network (CNN) models, for estimating the EE of each activity were compared. After physical activity classification, the independent models for activity prediction were used for training and EE estimation.

2.1. Data Acquisition

The dataset used in this study was obtained from a study conducted in 2019 [12]. Three XBOX 360 Kinect cameras (Microsoft Corp., Redmond, WA, USA) were placed along three directions, namely the side, rear side, and rear view directions. The positions of the Kinect cameras are illustrated in Figure 2. Each Kinect camera was installed and fixed to a tripod 0.9 m high. The distance between each camera and the participant was set to approximately 2 m, which ensured that the Kinect cameras could capture the participant’s entire body. The sampling frequency of Kinect is 30 Hz.
During experiments, participants were asked to wear a portable metabolic analyzer ( K 4 b 2 , COSMED, Rome, Italy) to record the pulmonary gas exchange by breath with an accuracy of deviation of ±0.02% O2 and ± 0.01% CO2.
A total of 21 subjects (10 men and 11 women) were recruited for the experiment. Their gender, ages, weights, and body fat ratios are listed in Table 1. The mean and standard deviation (SD) of the age of the participants is 21.90 ± 1.55 years old. The mean and SD of the weight is 60.20 ± 7.60 kg. The mean and SD of the body fat rate is 20.69 ± 7.37%.
People with infectious or chronic diseases were excluded from this study. All participants provided informed consent before participating in the experiments. The experimental procedures used in this study were approved by the Institutional Review Board (IRB) of Cathay General Hospital, Taipei, Taiwan (IRB code: CGH-NTPU105001).
Experimental details for each activity are listed in Table 2. Various physical activities, including standing, walking, and running at various speeds, were included in the experiments. Standing, walking, and running are regarded as the light, moderate, and vigorous activities in this study. The walking speeds considered were 4.8, 5.6, and 6.4 km/h, and the running speeds considered were 8.0 and 8.3 km/h and were set on the treadmill. Participants performed each activity for 5 min on a treadmill. In the pilot study, a treadmill was used instead of free-living conditions because a stable movement speed can be maintained with the treadmill. The resting times between each light-to-moderate activities and moderate-to-vigorous activities were 5 and 10 min, respectively.
The skeleton sequential data were collected with Kinect cameras while the participants were performing the physical activity on the treadmill. The first 3 min of each activity represented a non-steady state, and data from this period were discarded in this study [14]. Two minutes of steady-state data were used for physical activity classification and EE prediction.
To develop an EE estimation model, the expenditure data for each physical activity should be collected as the ground truth. In this study, the metabolic equivalent of task (MET) was regarded as the measurement unit. The MET is calculated as the ratio of the rate of energy expended during physical activity to the mass of a person. This ratio is used by many aerobic training organizations to estimate exercise intensity and EE.

2.2. Data Preprocessing

Data preprocessing can be divided into two parts: coordinate system transformation and noise reduction. Coordinate system transformation aims to transform the reference coordinate of each skeletal value into the body center. The initial reference coordinate of skeletal values was the center of a Kinect camera, but the varying object distances from a Kinect camera would cause problems of action recognition. For example, when the participants were walking or running on the treadmill, their bodies moved back and forth, which caused unstable variations in the skeletal data. According to the previous study, the referenced coordinate should be translated from the Kinect camera to the origin of body center, which is the shoulder center [15]. In this study, the reference coordinate was set to the center of the shoulder as the previous study had suggested. The reference coordinate of other joints was transformed by using the relative distance between the joint and the shoulder center. The formula for coordinate system transformation can be expressed as follows:
( x i ,   y i ,   z i ) = ( x i x r ,   y i y r ,   z i z r ) ,
where the coordinate of the shoulder center is ( x r , y r , z r ) , i = 1 ,   ,   j is the index of the joint, j is the total selected number of joints, which is six and will be explained in the following section, and ( x i , y i , z i ) denotes the skeleton coordinates after transformation.
Kinect cameras are sensitive to backgrounds, light, and surrounding objects. Noise occurs during skeletal tracking [16]. Therefore, a moving average filter, which averages the coordinates of two previous and two subsequent frames [17], can be used to remove the noise. The formula for applying the moving average filter is:
( x k * ,   y k * ,   z k * ) = 1 5 ( n = k 2 k + 2 x n ,   n = k 2 k + 2 y n ,   n = k 2 k + 2 z n )
where k   represents the current frame, k 2 to k + 2 represent two previous frames and two subsequent frames, and ( x k * ,   y k * ,   z k * ) is the new skeleton coordinate after processing with the moving average filter.

2.3. Feature Extraction

Because the speed of each activity is varying, the mean velocity between every two consecutive frames might be an important factor to discriminate the activities. The formula is presented in (3). The velocity of a joint is calculated by obtaining the difference of velocity between two consecutive frames [18]. v x k ,   v y k , and v z k means the mean velocity in x, y, and z directions and will be used to train the models. The accuracy of the positions of joints from Kinect was also discussed and validated in the previous study to prove that the data used for this study are reliable [19].
( v x k ,   v y k , v z k ) = ( x k * ,   y k * ,   z k * ) ( x k 1 * ,   y k 1 * ,   z k 1 * ) t k t k 1
According to the related research [20], treadmill workouts, such as standing, walking, and running, involve periodic and symmetric movements. Thus, to reduce the computational complexity, only half of the joints in the body were used to analyze physical activities. The six joints selected in this study were the shoulder, elbow, wrist, hip, knee, and ankle on the left side (red circles in Figure 3). The six selected joints were imaged in three directions simultaneously by using the three cameras. Three-dimensional data were obtained for each joint. Therefore, 18 velocities were obtained as features for further analysis.
Principal component (PC) analysis (PCA) is used to reduce feature dimensionality. In this study, PCs that cumulatively accounted for more than 90% of the overall variances were retained [21]. The numbers of feature subsets from the side, rear side, and rear views were 10, 9, and 7, respectively. The cumulative explained variance of the retained PCs is plotted in Figure 4.

2.4. Physical Activity Classification

To improve the accuracy of EE estimation, physical activities were classified as standing, walking, or running. An SVM algorithm was adopted to classify physical activities [22]. SVM is a widespread method to solve the classification problems. It can convert the raw data from lower to higher dimension using the kernel functions. After converting to higher dimension, the optimal hyper plane can be found to separate the data from different classes with the maximized margin. The advantage of SVM is that it can solve linear inseparable problems after projecting the raw data to a higher dimension. In this study, three SVM classifiers were built to classify the physical activities into three activities with one-against-all approach, and radial basis function kernel was used as the kernel function.

2.5. EE Prediction

To obtain the optimal model for estimating the EE during physical activities, the estimation results obtained by using three models, namely LR, MLP, and CNN, were compared. The reason of testing the three models is that these are the most common models in statistical, machine learning, and deep learning fields. The LR model can be used to determine whether the ground truth and estimated output are highly correlated. The MLP model is a type of feedforward artificial neural network, which comprises an input layer, an output layer, and several hidden layers [23]. The MLP model used in this study contained five layers: one input layer, three hidden layers, and one output layer, as presented in Figure 5. Different numbers of nodes n were tested to obtain the optimal number of neurons in the hidden layers to achieve optimal regression results. The logistic function was used as the activation function to estimate EE for different numbers of neurons, and mean square error (MSE) was used to evaluate the performance of different numbers of nodes. The number of nodes varied from 10 to 90 in increments of 10 neurons.
CNN is widely used to recognize objects and patterns in images [24]. In this study, CNN was used to automatically extract the features of kinematic data. The structure of the constructed CNN model is depicted in Figure 6. The CNN model used in this study comprised one input layer, two convolutional layers, two max-pooling layers, and one dense layer. The kernel size for each convolutional layer was 3 × 3. The rectified linear unit activation function was used in each convolutional layer [25]. Eighteen velocity data were input as the 18 × 1 input layer. The two convolutional layers were used to detect the features of the input data and create the feature maps automatically. The fully connected layers comprised two hidden layers with 16 hidden units per layer. In addition, the final output layer will output the result of the MET.
To evaluate the performance of EE estimation and compare the performance with the related research, three indicators used in the related research [5,6,7,10,12], mean absolute error (MAE), MSE, and root MSE (RMSE), were adopted to evaluate the performance. The formulas for calculating MAE, MSE, and RMSE are as follows:
MAE = 1 N k = 1 N | Y ^ k Y k |
MSE = 1 N k = 1 N ( Y ^ k Y k ) 2
RMSE = k = 1 N ( Y ^ k Y k ) 2 N
where Y ^ k denotes the estimated EE from each model, and Y k denotes the actual EE from K 4 b 2 portable calorimetry system. k = 1, …, N is the index of the data point. N is the amount of the data points. The unit for the MAE and RMSE was the MET. The ranges for MAE, MSE, and RMSE are [0, ∞].
Ten-fold cross-validation was used to evaluate the accuracy of different EE predictive models [26]. The final accuracy of each model was the average accuracy obtained through 10-fold cross-validation.

3. Results

For the MLP model, different nodes were tested to obtain the optimal model for the further comparisons. Figure 7 indicated that the lowest MSE was achieved when 70 neurons per hidden layer were used. The MLP with three hidden layers which contains 70 neurons per layer was adopted for the further comparisons.
The performance evaluation of the models was divided into three parts. The EE estimation performance of (i) a general model for all physical activities, (ii) the physical activity classification method, and (iii) independent models for different physical activities were evaluated.

3.1. EE Estimation Performance of the General Model

A general model was built for estimating the energy expenditure of all physical activities for two purposes. The first is for comparison with independent models for physical activity classification. The second is that if only one Kinect camera is available, the optimal position of the general model can be used.
Data for different physical activities were used to evaluate the EE estimation performance of the LR, MLP, and CNN models. The performance of the models with and without PCA was also examined to evaluate whether PCA improves the performance of the general model. The performance of the models with and without PCA is presented in Table 3, Table 4 and Table 5. According to Table 3, Table 4 and Table 5, when data from all physical activities were used, the models without PCA outperformed those with PCA. The MLP and CNN models exhibited smaller errors than did the LR model. Moreover, the smallest estimation error was obtained when a Kinect camera was set along the side view and the MLP model without PCA was adopted. The MAE, MSE, and RMSE under the aforementioned settings were 0.55, 0.66, and 0.81, respectively.

3.2. Accuracy of Physical Activity Classification

Because the EE estimation method proposed in this study is mainly based on physical activity classification, the accuracy of physical activity classification must be examined. The performance of physical activity classification with and without PCA is presented in Table 6.
The performance differences in physical activity classification with and without PCA were nonsignificant. Moreover, the accuracy of physical activity classification without PCA was higher than that with PCA. Accuracy without PCA could reach 99.55%.
Therefore, PCA was not adopted for further comparisons of the models and experimental settings for different activities.

3.3. EE Estimation Performance with Physical Activity Classification

Table 7, Table 8 and Table 9 indicate the EE estimation performance with physical activity classification. The smallest error in the EE estimation for standing was achieved when a Kinect camera was installed along the rear view and the CNN model was used. The MAE, MSE, and RMSE of the EE estimation for standing were 0.15, 0.04, and 0.19, respectively. The smallest error in the EE estimation for walking was achieved when a Kinect camera was set along the side view and the CNN model was used. The MAE, MSE, and RMSE of the EE estimation for walking were 0.45, 0.33, and 0.57, respectively. The smallest error in the EE estimation for running was obtained when a Kinect camera was set along the rear view and the MLP model was used. The MAE, MSE, and RMSE of the EE estimation for running were 0.66, 0.94, and 0.96, respectively.

4. Discussion

In the proposed method, physical activity classification is performed before the EE estimation model is developed. To compare with the performance in different models and experimental settings, the performance evaluation was divided into three parts. First, the performance of a general model for all physical activities was investigated. According to the results in Table 3, Table 4 and Table 5, the smallest error in EE was obtained when the Kinect camera was set along the side view and the MLP model was used. The MAE, MSE, and RMSE in the EE under the aforementioned settings were 0.55, 0.66, and 0.81, respectively. Thus, when only one Kinect camera is available, optimal EE estimation performance is obtained when the camera is set along the side view and the MLP model is used. For an individual who weighs 60 kg and exercises for 1 h, the error between the EE estimated with the general model and the actual EE is 33 kcal.
Second, the accuracy of physical activity classification was examined with and without PCA. The classification accuracies with and without PCA were examined because the primary aim of this study was to determine whether physical activity classification improves EE estimation accuracy. Therefore, evaluating the accuracy of physical activity classification was crucial. As presented in Table 6, the accuracy of EE estimation along the side view was 99.55% without PCA and 98.49% with PCA. Therefore, superior EE estimation performance was obtained without PCA. A possible reason for this finding is the loss of velocity characteristics after using PCA because PCA results in the projection of data from high to low dimensions. Although the accuracy of using PCA is lower in this study, it was still helpful while collecting a larger dataset in future works because larger dataset will increase the computational cost of the model, and PCA can reduce the complexity of classification problems and make the predictive model more stable.
Third, the performance of EE estimation with physical activity classification was investigated. Table 7, Table 8 and Table 9 present the EE estimation performance of different models under different experimental settings for three physical activities. The results indicated that the CNN model provided the highest EE estimation accuracy for light and moderate-intensity activities, such as standing and walking. The MLP model exhibited the highest EE estimation accuracy for vigorous activities such as running. The optimal setting for achieving optimal EE estimation performance with physical activity classification is illustrated in Figure 8. The optimal setting involves two Kinect cameras along the side and rear views as well as a combination of MLP and CNN models. After physical activity classification, if the physical activity is classified as standing, walking, or running, then the EE is estimated using the CNN model for standing, CNN model for walking, or MLP model for running, respectively. The combination of various models reduced computational complexity because fewer data were required for independent model training than for general model training.
The data from the rear side view did not provide optimal results for any experimental setting possibly because the data provided by the camera on the rear side view have already been projected on the other two planes. Therefore, the skeletal data from the Kinect cameras along only the side and rear views should be used in future studies. Moreover, the results show that after applying physical activity classification for EE estimation, the performance of estimating EE for light to moderate activity, such as standing and walking, will be improved; however, the performance for vigorous activity will not be improved by applying physical activity classification according to the results.
Table 10 presents a comprehensive comparison of the proposed method with methods presented in related studies. Several criteria, such as sensors, accuracy, indoor/outdoor usage, limitation of battery time, and requirement of wearable devices, are discussed among all the listed studies in Table 10. The proposed method can estimate the EE of light-to-vigorous activities. Only Kinect cameras were used in this study. Of all the compared methods, only the proposed method does not require personal data, such as weight, heart rate, and gender. The proposed method does not have the problem of battery life but can be used for indoor activities only. None of the compared camera-based systems require wearable devices. The accuracy of the proposed general model is the same as that of the system proposed by Cvetković et al. [5] but lower than that of the system proposed by Lin et al. [12]. However, in the EE estimation for standing and walking, the proposed system outperforms that of Lin et al. This result is acceptable because standing and walking account for most of the exercising time. The proposed method only uses velocity-based features (only 18 features) for EE estimation. Thus, the computational complexity of the proposed method is lower than that of other methods.
This study has certain limitations. First, not all models considered in this study were fine tuned to their optimal parameters. In this study, only the neurons of the hidden layers in the MLP model were fine tuned. Second, two Kinect cameras are still required to obtain the best EE prediction performance. However, a practical solution requiring the use of only one Kinect camera would be more appropriate.
To solve the aforementioned limitations, in the future studies, the optimal parameters for all models will be investigated. Deep learning approaches for sequential datasets, such as the algorithms based on recurrent neural networks [27,28] and long short-term memory networks [29], will also be considered for EE estimation. Moreover, it is necessary to build a robust model with only one Kinect camera for more practical usage. The approaches for converting the coordinates of a Kinect camera from one direction to other direction will be considered for improving the performance and efficiency of the EE estimation model on any direction.

5. Conclusions

This paper proposed an EE estimation system based on physical activity classification. Three depth cameras were set at three locations, namely the side, rear side, and rear views, to obtain the kinematic data and test the performance of EE estimation. Three EE estimation models, including LR, MLP, and CNN models, were compared and determined the model with optimal performance in varying experimental settings.
The experimental results indicated that when only one camera is available, the optimal solution involves setting the camera along the side view and estimating EE with the MLP model. The MAE, MSE, and RMSE of the EE under the aforementioned settings were 0.55, 0.66, and 0.81, respectively. In physical activity classification, the highest EE estimation accuracy for light-to-moderate activities, such as standing and walking, was obtained when Kinect cameras were set along the side and rear views and the CNN model was used. The highest EE estimation accuracy for vigorous activities, such as running, was obtained when Kinect cameras were set along the side and rear views and the MLP model was used. The RMSEs for estimating the EEs of standing, walking, and running were 0.19, 0.57, and 0.96, respectively. The camera on the rear side view did not yield optimal performance in all the experimental settings in this study. The results indicated that after applying physical activity classification, the performance of estimating EE of standing and walking was improved compared with that of the general model; however, the performance of running was not improved after applying physical activity classification. By using the proposed system, the EE of physical activities can be accurately estimated.
The main innovation of this research is that this is the first study to discuss the performance of EE estimation in different experimental settings, including one and two cameras, and the speed of different types of activities. With the proposed system, the subject can choose the best setting depending on their exercising environments to estimate EE.

Author Contributions

Conceptualization, B.-S.L. and C.-S.F.; methodology, M.-L.W.; software, Y.-F.L.; validation, Y.-F.L., W.-J.C. and M.-L.W.; formal analysis, Y.-F.L.; investigation, W.-J.C.; resources, B.-S.L. and C.-S.F.; data curation, I.-J.L.; writing—original draft preparation, Y.-F.L., M.-L.W. and C.-S.F.; writing—review and editing, W.-J.C., I.-J.L. and B.-S.L.; visualization, B.-S.L. and C.-S.F.; supervision, B.-S.L. and C.-S.F.; project administration, I.-J.L. and M.-L.W.; funding acquisition, B.-S.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partly supported by Ministry of Science and Technology in Taiwan, under grants MOST 107-2221-E-011-113-MY2, MOST 107-2221-E-305-014, MOST 108-2314-B-305-001, MOST 109-2314-B-305-001, and MOST 109-2221-E-305-001-MY2. This research was also partly supported by University System of Taipei Joint Research Program, under grants USTP-NTPU-TMU-109-03, Faculty Group Research Funding Sponsorship by National Taipei University, under grants 2020-NTPU-ORDA-02, and “Academic Top-Notch and Features Field Project” Funding Sponsorship of National Taipei University, Taiwan (R.O.C), under 109-NTPU_ORDA-F-005.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board of Cathay General Hospital, Taipei, Taiwan (protocol code: CGH-NTPU105001).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. World Health Organization: Physical Activity. Available online: https://www.who.int/news-room/fact-sheets/detail/physical-activity (accessed on 24 May 2021).
  2. Causes of Death—Our World in Data. Available online: https://ourworldindata.org/causes-of-death (accessed on 24 May 2021).
  3. McLaughlin, J.E.; King, G.A.; Howley, E.T.; Bassett, D.R.; Ainsworth, B.E. Validation of the COSMED K4b2 portable metabolic system. Int. J. Sports Med. 2001, 22, 280–284. [Google Scholar] [CrossRef] [PubMed]
  4. Altini, M.; Penders, J.; Vullers, R.; Amft, O. Estimating energy expenditure using body-worn accelerometers: A comparison of methods, sensors number and positioning. IEEE J. Biomed. Health Inform. 2015, 19, 219–226. [Google Scholar] [CrossRef]
  5. Cvetković, B.; Milić, R.; Luštrek, M. Estimating Energy Expenditure with Multiple Models Using Different Wearable Sensors. IEEE J. Biomed. Health Inform. 2016, 20, 1081–1087. [Google Scholar] [CrossRef]
  6. Park, H.; Dong, S.Y.; Lee, M.; Youn, I. The Role of Heart-Rate Variability Parameters in Activity Recognition and Energy-Expenditure Estimation Using Wearable Sensors. Sensors 2017, 17, 1698. [Google Scholar] [CrossRef]
  7. Hedegaard, M.; Anvari-Moghaddam, A.; Jensen, B.K.; Jensen, C.B.; Pedersen, M.K.; Samani, A. Prediction of energy expenditure during activities of daily living by a wearable set of inertial sensors. Med. Eng. Phys. 2020, 75, 13–22. [Google Scholar] [CrossRef] [PubMed]
  8. Kim, Y.; Choudhury, S.; Kong, H.J. Application of Micro-Doppler Signatures for Estimation of Total Energy Expenditure in Humans for Walking/Running Activities. IEEE Access 2016, 4, 1560–1569. [Google Scholar] [CrossRef] [Green Version]
  9. Yang, Y.; Liu, C.; Tsow, F.; Shao, D.; Yu, H.; Xia, S.; Tao, N. Remote Quantification of Workout Energy Expenditure with a Cell Phone Camera. IEEE Sens. J. 2016, 16, 8263–8270. [Google Scholar] [CrossRef]
  10. Koporec, G.; Vučković, G.; Milić, R.; Perš, J. Quantitative contact-less estimation of energy expenditure from video and 3D imagery. Sensors 2018, 18, 2435. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Na, H.S.; Choi, J.H.; Kim, H.S.; Kim, T. Development of a human metabolic rate prediction model based on the use of Kinect-camera generated visual data-driven approaches. Build. Environ. 2019, 160, 106216. [Google Scholar] [CrossRef]
  12. Lin, B.-S.; Wang, L.-Y.; Hwang, Y.-T.; Chiang, P.-Y.; Chou, W.-J. Depth-Camera-Based System for Estimating Energy Expenditure of Physical Activities in Gyms. IEEE J. Biomed. Health Inform. 2019, 23, 1086–1095. [Google Scholar] [CrossRef]
  13. Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional neural networks: An overview and application in radiology. Insights Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef] [Green Version]
  14. Gordon, K.E.; Ferris, D.P.; Kuo, A.D. Metabolic and mechanical energy costs of reducing vertical center of mass movement during gait. Arch. Phys. Med. Rehabil. 2009, 90, 136–144. [Google Scholar] [CrossRef] [PubMed]
  15. Taha, A.; Zayed, H.H.; Khalifa, M.E.; El-Horbaty, E.S. Human Action Recognition based on MSVM and Depth Images. Int. J. Comput. Sci. 2014, 11, 42–51. [Google Scholar]
  16. Livingston, M.A.; Sebastian, J.; Ai, Z.; Decker, J.W. Performance measurements for the Microsoft Kinect skeleton. In Proceedings of the 2012 IEEE Virtual Reality Workshops (VRW), Costa Mesa, CA, USA, 4–8 March 2012; pp. 119–120. [Google Scholar]
  17. Chou, Y.L. Statistical Analysis: With Business and Economic Applications; Holt McDougal: New York, NY, USA, 1975; ISBN 978-003-073-095-5. [Google Scholar]
  18. Ren, P.; Bosch Bayard, J.F.; Dong, L.; Chen, J.; Mao, L.; Ma, D.; Sanchez, M.A.; Morejon, D.M.; Bringas, M.L.; Yao, D.; et al. Multivariate Analysis of Joint Motion Data by Kinect: Application to Parkinson’s Disease. IEEE Trans. Neural Syst. Rehabil. Eng. 2020, 28, 181–190. [Google Scholar] [CrossRef]
  19. Khoshelham, K.; Elberink, S.O. Accuracy and Resolution of Kinect Depth Data for Indoor Mapping Applications. Sensors 2012, 12, 1437–1454. [Google Scholar] [CrossRef] [Green Version]
  20. Winter, D.A. Biomechanics and Motor Control of Human Movement, 4th ed.; John Wiley & Sons, Inc.: Hoboken, NJ, USA, 2009; ISBN 978-047-054-914-8. [Google Scholar]
  21. Shlens, J.A. A Tutorial on Principal Component Analysis. arXiv 2014, arXiv:1404.1100. [Google Scholar]
  22. Chang, C.C.; Lin, C.J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. 2011, 2, 1–27. [Google Scholar] [CrossRef]
  23. Hagan, M.T.; Demuth, H.B.; Beale, M.H. Neural Network Design; Martin Hagan: Stillwater, OK, USA, 2014; ISBN 978-097-173-211-7. [Google Scholar]
  24. Namatēvs, I. Deep Convolutional Neural Networks: Structure, Feature Extraction and Training. Inf. Technol. Manag. Sci. 2018, 20, 40–47. [Google Scholar] [CrossRef] [Green Version]
  25. Arora, R.; Basu, A.; Mianjy, P.; Mukherjee, A. Understanding deep neural networks with rectified linear units. In Proceedings of the 6th International Conference on Learning Representations 2018 (ICLR 2018), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  26. Fushiki, T. Estimation of prediction error by using K-fold cross-validation. Stat. Comput. 2011, 21, 137–146. [Google Scholar] [CrossRef]
  27. Jain, L.C.; Medsker, L.R. Recurrent Neural Networks: Design and Applications, 1st ed.; CRC Press, Inc.: Boca Raton, FL, USA, 1999; ISBN 978-0-8493-7181-3. [Google Scholar]
  28. Schuster, M.; Paliwal, K.K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 1997, 45, 2673–2681. [Google Scholar] [CrossRef] [Green Version]
  29. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
Figure 1. System architecture used in the proposed method.
Figure 1. System architecture used in the proposed method.
Sensors 21 04216 g001
Figure 2. Locations of the Kinect cameras.
Figure 2. Locations of the Kinect cameras.
Sensors 21 04216 g002
Figure 3. Map of the Kinect-tracked skeleton joints. The red-marked joints represent the selected joints.
Figure 3. Map of the Kinect-tracked skeleton joints. The red-marked joints represent the selected joints.
Sensors 21 04216 g003
Figure 4. Cumulative and individual explained variances of the retained PCs for each Kinect placement: (a) side view, (b) rear side view, and (c) rear view.
Figure 4. Cumulative and individual explained variances of the retained PCs for each Kinect placement: (a) side view, (b) rear side view, and (c) rear view.
Sensors 21 04216 g004
Figure 5. Structure of the constructed MLP model.
Figure 5. Structure of the constructed MLP model.
Sensors 21 04216 g005
Figure 6. Structure of the constructed CNN model.
Figure 6. Structure of the constructed CNN model.
Sensors 21 04216 g006
Figure 7. MSEs for different numbers of nodes (from 10 to 90) in the MLP model.
Figure 7. MSEs for different numbers of nodes (from 10 to 90) in the MLP model.
Sensors 21 04216 g007
Figure 8. Optimal setting of the proposed system.
Figure 8. Optimal setting of the proposed system.
Sensors 21 04216 g008
Table 1. Participants’ data.
Table 1. Participants’ data.
Number of Subjects21
Sex10 males, 11 females
Age (years old)21.90 ± 1.55
Weight (kg)60.20 ± 7.60
Body Fat Rate (%)20.69 ± 7.37
Table 2. Experimental details of each activity.
Table 2. Experimental details of each activity.
TypeActivitySpeed (km/h)Time (min)Rest Time (min)
LightStanding050
ModerateWalking4.855
Walking5.655
Walking6.455
VigorousRunning8.0510
Running8.3510
Table 3. LR model results obtained with and without PCA.
Table 3. LR model results obtained with and without PCA.
SideRear SideRear
Without PCAWith PCAWithout PCAWith PCAWithout PCAWith PCA
MAE (MET)1.291.381.511.541.251.50
MSE (MET2)2.943.463.854.042.623.69
RMSE (MET)1.711.861.962.011.621.92
Table 4. MLP model results obtained with and without PCA.
Table 4. MLP model results obtained with and without PCA.
SideRear SideRear
Without PCAWith PCAWithout PCAWith PCAWithout PCAWith PCA
MAE (MET)0.550.670.851.060.610.82
MSE (MET2)0.661.121.652.520.811.52
RMSE (MET)0.811.061.281.580.901.23
Table 5. CNN model results obtained with and without PCA.
Table 5. CNN model results obtained with and without PCA.
SideRear SideRear
Without PCAWith PCAWithout PCAWith PCAWithout PCAWith PCA
MAE (MET)0.630.780.921.080.700.89
MSE (MET2)0.811.301.662.360.951.60
RMSE (MET)0.901.141.291.530.981.26
Table 6. Accuracy of the SVM algorithm with and without PCA for classifying different activities at three Kinect camera locations.
Table 6. Accuracy of the SVM algorithm with and without PCA for classifying different activities at three Kinect camera locations.
SideRear SideRear
Without PCA99.55%96.33%98.70%
With PCA98.49%92.16%95.72%
Table 7. EE estimation performance of independent LR models for different activities.
Table 7. EE estimation performance of independent LR models for different activities.
SideRear SideRear
StandingWalkingRunningStandingWalkingRunningStandingWalkingRunning
MAE (MET)0.240.701.130.220.751.230.200.621.12
MSE (MET2)0.090.822.260.080.902.550.060.622.11
RMSE (MET)0.300.901.500.280.951.600.250.781.45
Table 8. EE estimation performance of independent MLP models for different activities.
Table 8. EE estimation performance of independent MLP models for different activities.
SideRear SideRear
StandingWalkingRunningStandingWalkingRunningStandingWalkingRunning
MAE (MET)0.180.440.690.170.590.880.160.470.66
MSE (MET2)0.060.340.960.050.591.640.040.410.94
RMSE (MET)0.240.580.980.230.771.280.200.640.96
Table 9. EE estimation performance of independent CNN models for different activities.
Table 9. EE estimation performance of independent CNN models for different activities.
SideRear SideRear
StandingWalkingRunningStandingWalkingRunningStandingWalkingRunning
MAE (MET)0.160.450.760.170.560.870.150.470.73
MSE (MET2)0.040.331.180.050.541.430.040.370.98
RMSE (MET)0.210.571.080.220.741.190.190.600.99
Table 10. Comparison of the proposed method with methods proposed in related studies.
Table 10. Comparison of the proposed method with methods proposed in related studies.
Cvetković et al. [5]Park et al. [6]Koporec et al. [10]Lin et al. [12]Hedegaard et al. [7]The Proposed Method
Year201620172018201920202020
ActivityLight to vigorousLight to vigorousLight to vigorousModerate to vigorousLight to vigorousLight to vigorous
SensorsMulti-sensorsMulti-sensorsCameraKinect cameraIMU sensorsKinect camera
Personal DataYesYesNoYesYesNo
Indoor/OutdoorBothBothIndoorIndoorBothIndoor
Limitation of battery timeYesYesNoNoYesNo
Wearable devices are neededYesYesNoNoYesNo
RMSE (MET)0.810.892.220.76Sedentary—0.29General model—0.81
Standing—0.19
Dynamic—1.14Walking—0.57
Running—0.96
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lin, B.-S.; Lee, I.-J.; Fahn, C.-S.; Lee, Y.-F.; Chou, W.-J.; Wu, M.-L. Depth-Camera Based Energy Expenditure Estimation System for Physical Activity Using Posture Classification Algorithm. Sensors 2021, 21, 4216. https://doi.org/10.3390/s21124216

AMA Style

Lin B-S, Lee I-J, Fahn C-S, Lee Y-F, Chou W-J, Wu M-L. Depth-Camera Based Energy Expenditure Estimation System for Physical Activity Using Posture Classification Algorithm. Sensors. 2021; 21(12):4216. https://doi.org/10.3390/s21124216

Chicago/Turabian Style

Lin, Bor-Shing, I-Jung Lee, Chin-Shyurng Fahn, Yi-Fang Lee, Wei-Jen Chou, and Meng-Luen Wu. 2021. "Depth-Camera Based Energy Expenditure Estimation System for Physical Activity Using Posture Classification Algorithm" Sensors 21, no. 12: 4216. https://doi.org/10.3390/s21124216

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop