Next Article in Journal
Parkinson’s Kinetigraph for Wearable Sensor Detection of Clinically Unrecognized Early-Morning Akinesia in Parkinson’s Disease: A Case Report-Based Observation
Next Article in Special Issue
An Optical Sensor for Measuring Displacement between Parallel Surfaces
Previous Article in Journal
3D VSP Imaging Using DAS Recording of P- and S-Waves in Vertical and Lateral Well Sections in West Texas
Previous Article in Special Issue
Towards a Remote Patient Monitoring Platform for Comprehensive Risk Evaluations for People with Diabetic Foot Ulcers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A CNN Model for Physical Activity Recognition and Energy Expenditure Estimation from an Eyeglass-Mounted Wearable Sensor

by
Md Billal Hossain
1,
Samuel R. LaMunion
2,
Scott E. Crouter
2,
Edward L. Melanson
3 and
Edward Sazonov
1,*
1
Department of Electrical and Computer Engineering, The University of Alabama, Tuscaloosa, AL 35487, USA
2
Department of Kinesiology, Recreation and Sport Studies, The University of Tennessee, Knoxville, TN 37996, USA
3
USA Division of Endocrinology, Metabolism, and Diabetes, University of Colorado Anschutz Medical Campus, Aurora, CO 80045, USA
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(10), 3046; https://doi.org/10.3390/s24103046
Submission received: 4 April 2024 / Revised: 4 May 2024 / Accepted: 10 May 2024 / Published: 11 May 2024
(This article belongs to the Special Issue Applications of Body Worn Sensors and Wearables)

Abstract

:
Metabolic syndrome poses a significant health challenge worldwide, prompting the need for comprehensive strategies integrating physical activity monitoring and energy expenditure. Wearable sensor devices have been used both for energy intake and energy expenditure (EE) estimation. Traditionally, sensors are attached to the hip or wrist. The primary aim of this research is to investigate the use of an eyeglass-mounted wearable energy intake sensor (Automatic Ingestion Monitor v2, AIM-2) for simultaneous recognition of physical activity (PAR) and estimation of steady-state EE as compared to a traditional hip-worn device. Study data were collected from six participants performing six structured activities, with the reference EE measured using indirect calorimetry (COSMED K5) and reported as metabolic equivalents of tasks (METs). Next, a novel deep convolutional neural network-based multitasking model (Multitasking-CNN) was developed for PAR and EE estimation. The Multitasking-CNN was trained with a two-step progressive training approach for higher accuracy, where in the first step the model for PAR was trained, and in the second step the model was fine-tuned for EE estimation. Finally, the performance of Multitasking-CNN on AIM-2 attached to eyeglasses was compared to the ActiGraph GT9X (AG) attached to the right hip. On the AIM-2 data, Multitasking-CNN achieved a maximum of 95% testing accuracy of PAR, a minimum of 0.59 METs mean square error (MSE), and 11% mean absolute percentage error (MAPE) in EE estimation. Conversely, on AG data, the Multitasking-CNN model achieved a maximum of 82% testing accuracy in PAR, a minimum of 0.73 METs MSE, and 13% MAPE in EE estimation. These results suggest the feasibility of using an eyeglass-mounted sensor for both PAR and EE estimation.

Graphical Abstract

1. Introduction

Energy intake in humans comes from consuming food and beverages, while energy expenditure (EE) results from various physiological processes and physical activities. Specifically, EE comprises the resting metabolic rate, thermic effect of feeding, physical activity-induced energy, and non-exercise activity thermogenesis. The thermic effect of feeding accounts for a small fraction of EE, typically 8–15% [1]. The most variable forms of EE are physical activity-induced energy and non-exercise activity thermogenesis, which can be controlled through physical activity and exercise such as walking or running [2]. EE is approximately constant in a steady state of a specific physical activity (e.g., walking at a constant pace). However, EE is not constant during physical activity transitions (e.g., start running after walking). The intensity of EE is expressed in kilojoules (kJ) or kilocalories (kcal) per unit of time and body weight, or as metabolic equivalents (METs). EE in humans is typically determined using indirect calorimetry (e.g., metabolic carts, portable metabolic systems, or the doubly labeled water method). Although those methods provide higher accuracy, they are not practical for application in the general population. Therefore, researchers are exploring different technological approaches to estimate EE in daily life.
Energy balance refers to the state where the energy intake of an individual is equivalent to their EE [3]. Imbalances in energy intake and EE lead to alterations in body mass [4]. Sustained positive energy balance (i.e., energy intake > energy expenditure) results in weight gain, with approximately 60–80% of the gained weight attributed to body fat [5,6]. Conversely, sustained negative energy balance (energy intake < energy expenditure) results in a loss in body mass. Positive energy balance, often associated with overeating and a sedentary lifestyle, has become a significant concern in modern society due to the potential risk of metabolic syndromes such as type-2 diabetes, obesity, hypertension, dyslipidemia, stroke, and cardiovascular disease. The possible solution to minimize the positive imbalance is to reduce overeating and increase physical activity. Increasing physical activity increases EE, which minimizes the effect of positive imbalance and helps control weight gain [7,8,9]. The prevalence of metabolic syndromes has significantly increased, highlighting the need for a comprehensive energy intake, EE, and physical activity monitoring system.
Physical activity refers to the movement of skeletal muscles that results in EE [10]. Physical Activity Recognition (PAR) aims to identify the type of physical activity, ranging from simple to complex activities. Once identified, the EE of a given activity can be estimated using standard values or individually calibrated values [11]. PAR is an actively researched topic due to its various real-world applications in areas such as human-computer interaction [12,13], security and surveillance systems [14,15], and healthcare systems [16,17,18,19]. In some studies [20,21], researchers have integrated PAR and EE estimation through wearable technology.
Wearable sensor devices are becoming increasingly popular due to their portability, light weight, and application diversity. Researchers are experimenting with different sensors, such as accelerometers, gyroscopes, heart rate monitors, and temperature sensors, in different body locations, such as the waist, chest, arm, leg, hip, and head. Among sensors, the accelerometer-gyroscope combination (Inertial Measurement Unit (IMU)) is frequently used due to its ability to measure multiple aspects of human motion [22,23]. The hip and wrist are the most popular body positions for the IMU-based PAR and EE estimations. Research has shown that data collected from wrist-worn devices (for example, the ActiGraph GT3X+ [24]) can overestimate EE compared to hip-worn devices [25]. A large-scale experiment [26] tested PAR and EE estimation performance using six different wear positions (wrist, hip, ankle, upper arm, and thigh). It concluded that the hip position was the best for both PAR and EE estimation. Other experiments have introduced wearables (eSense) and head-worn sensors placed in headbands for PAR [27,28]. To the best of our knowledge, no research has been conducted on simultaneously estimating PAR and EE using eyeglass-mounted head-worn wearable sensor devices. Eyeglass-mounted, head-worn wearable sensors may offer new possibilities. Location on the frame of eyeglasses may allow to integrate PAR and EE monitoring with energy intake monitoring through a variety of eating detection methods [29,30]. Traditional wearable devices for PAR and EE estimation are incapable of tracking energy intake, which presents a significant advantage for eyeglass-mounted wearable sensor devices that naturally follow the eye gaze during eating and could be used to capture images of food. Eyeglasses are a prevalent wearable, with over 75% of American [31] and 64% of global [32] adults wearing glasses for vision correction. The use of PAR sensors by regular eyeglass users may potentially reduce the user burden and improve compliance with wear. Since the movement patterns of the head are different from those at the hip or wrist, this paper attempts to evaluate the accuracy of PAR and EE estimations with respect to established methodology. Additionally, it also features hands-free operation and passive monitoring, making it the ideal choice for our research experiment. While there are minor considerations, such as the added weight, these benefits make them a promising option.

2. Previous Studies and Our Contribution

Traditionally, PAR and EE estimation models were developed based on classical signal processing techniques [21,23,26], statistical and classical machine learning [20]. Nevertheless, the superior performance of deep learning, such as convolutional neural networks (CNNs), long short-term memory (LSTM), and attention mechanisms, makes them attractive for PAR and EE estimation applications [33,34]. In [35], CNN and LSTM with self-attention mechanisms improved PAR performance, and a similar experiment in [36] enhanced arm exercise activity recognition. CNN-LSTM models showed improved performance and were used by other recent researchers [37,38]. The real-time feasibility of the IMU sensor and CNN-based PAR was introduced in [39]. A soft-voting and self-learning-based PAR method was proposed in [40], where the researcher enhanced accuracy by employing multiple machine learning models. The researcher also utilized a self-training mechanism to generate training data and retrain the model iteratively. Furthermore, a study in [41] proposed CNN-based EE estimation for ambulatory activities, reporting a lower error with 30 participants and five activities. In [42], multilayer perceptron (MLP) was applied to EE estimation data from 102 healthy adults engaged in two structured activities. A graph-based PAR method is proposed in [43], incorporating a combination of the deterministic selection algorithm, the lion optimization algorithm, and the association pattern algorithm (referred to as gSpan).
Notably, two significant research experiments involve multitasking models for PAR and EE estimation. Multitasking is superior to the single-task model when applied to a collection of related tasks using shared representations to learn [44]. These shared representations increase data efficiency and potentially yield faster learning and inference speeds. Additionally, simultaneously monitoring physical activity and estimating EE offers several advantages, including a more comprehensive understanding of an individual’s energy balance and activity levels, allowing for real-time tracking of EE during different activities, aiding in personalized activity recommendations, and enhancing the accuracy of energy balance assessments. Moreover, it can provide valuable insights into the relationship between activity patterns and energy intake, facilitating more effective strategies for weight management and overall health improvement [45]. In [46], a multitasking LSTM network classified activity types and estimated activity intensity using raw sensor data and MET values. Another multitasking model for PAR and EE estimation was proposed in [47], based on computer vision and heart rate sensors. Although substantial progress has been achieved, a significant research scope remains unexplored in multitasking model development for PAR and EE estimation.
The primary objective of this pilot study is to investigate the viability of using the head-mounted wearable device known as the ‘Automatic Ingestion Monitor version 2 (AIM-2) [48]’ for comprehensive estimation of PAR and steady-state EE during six different physical activities, utilizing a multitasking model. Our contributions are as follows:
  • For the first time, we demonstrated the feasibility of using an eyeglass-based wearable sensor (AIM-2) for simultaneous PAR and EE estimation.
  • We proposed a multitasking CNN (Multitasking-CNN) model for comprehensive estimation of PAR and EE. We employ a two-step progressive training approach to enhance the model’s performance.
  • We compared the performance of Multitasking-CNN in PAR and EE estimation using data from both the hip-worn ActiGraph GT9X device and the head-mounted AIM-2 device.

3. Materials and Methods

3.1. Wearable Sensor Devices

Two wearable sensor devices collected the experimental data: AIM-2 and ActiGraph GT9X (AG) [49]. AIM-2 is an eyeglass-mounted passive wearable device primarily designed for dietary assessment (Figure 1). AIM-2 consists of a miniature 5 Megapixel camera with a 170-degree wide-angle gaze-aligned lens, a low-power 3D accelerometer (ADXL362 from Analog Devices, Norwood, MA, USA), and a gyroscope (LSM6DS3TR-C from STMicroelectronics N.V.). The IMU from AIM-2 has a full-scale range of ±8 g and ±2000 degrees per second, respectively. Data from the AIM-2 IMU were sampled at 128 Hz and stored on an SD card. The AIM-2 can be attached to any eyeglass using simple, double-sided adhesive, eliminating the need for a specialized eyeglass frame.
AG is a lightweight (14 g) and non-invasive commercially available wearable sensor device that is widely used in the PAR research community [33]. The AG device is equipped with a triaxial primary accelerometer (full scale range ± 8 g), a triaxial gyroscope (full scale range ± 2000 degrees per second), a triaxial magnetometer (full scale range ± 4800 micro-Tesla), and an internal sensor thermometer to measure the temperature inside the sensor housing. The primary accelerometer in the AG device has a selectable sampling rate of 30–100 Hz (90 Hz selected), and the gyroscope has a fixed sampling rate of 100 Hz.
A portable indirect calorimeter (COSMED K5) [50] was used for collecting the ground truth EE. The COSMED K5 measures oxygen consumption and carbon dioxide production using a breath-by-breath or mixing chamber mode. The COSMED K5 unit was calibrated prior to each use using a five-step manufacturer specified procedure, including room air calibration, carbon dioxide scrubber calibration, reference gas calibration using a mixture of 16.0% O2, 5.0% CO2, and balanced nitrogen, volume calibration using a 3-liter Hans Rudolf syringe, and delay calibration to check the timing of air flow from the face mask to the analyzers. The accuracy of the device is reported in [51].

3.2. Study Design

A study was conducted at The University of Tennessee, Knoxville, with six participants (mean ± SD; age 23.8 ± 3.9 y, BMI 24.25 ± 5.77 kg/m2, body weight 77.1 ± 20.2 kg; 5 males, 1 female). The University of Tennessee Knoxville Institutional Review Board approved the study protocol (IRB#: UTK IRB-14-01988-XP). Participants’ height and weight were measured at the beginning of the study visit. All measurements were taken without shoes and in lightweight clothing. Height was measured using a wall-mounted stadiometer, and weight was taken using a calibrated physician’s scale. Participants completed the following six structured activities for six minutes per activity during a single laboratory visit: seated computer work, sweeping, stationary cycling (50 W), variable treadmill walking (3 mph, 0% grade; 4 mph, 0% grade; 3 mph, 5% grade), treadmill running (6 mph, 0% grade), and one-vs.-one basketball. The AG was worn on the right hip, an AIM-2 was attached to the right arm of a pair of non-prescription eyeglasses, and a COSMED K5 was worn to measure gas exchange throughout the activity protocol (Figure 1). A human annotator annotated each activity’s start and stop times to obtain the periods of interest. Table 1 shows the summary of the duration of each activity (six participants identified by participant ID 101–106). The computer activity was discarded for participant 105 because the AG device started recording data after starting the computer activity data collection. Figure 2 displays the raw sensor signal samples collected by the AIM-2 device from participant 101 across various activities.
After each study session, breath-by-breath (averaged into a 30-s window) EE data were extracted from the COSMED OMNIA software. The relative VO2 (mLkg−1min−1) was obtained by dividing the mean VO2 (mLmin−1) from each 30-second epoch by the participant’s body mass in kg. For all weight-bearing activities, an additional 2 kg of weight was added to the body weight (the weight of the K5 unit and harness system). The EE in METs was then obtained by dividing the relative VO2 (mLkg−1min−1) by 3.5 mLkg−1min−1. The last 30-s window was discarded to obtain steady-state EE, and the average of the four preceding 30 s windows of data was used as the steady-state EE for each activity. The steady-state EE value for each activity was then matched by timestamp and criterion activity label to the corresponding 10 s activity periods for each participant. The transition between the activities was not considered for analysis. All analyses in this paper refer to steady-state EE estimation.

3.3. Study Data Processing

3.3.1. IMU Signal Analysis

Due to historical reasons, the AG device sampled acceleration at 90 Hz and the gyroscope at 100 Hz [52]. The accelerometer signal was up-sampled to 100 Hz using the MATLAB resample function to ensure the uniformity of the signals. The gyroscope signals from the AIM-2 and AG devices were clipped to ±400 degrees per second to remove extreme outliers. We did not apply any filters to remove the noise from the IMU signal. The accelerometer and clipped gyroscope sensor signals were used in the next step for sensor signal-to-image conversion.

3.3.2. Sensor Signal to Digital Image Conversion

In this step, sensor signals were represented as grayscale images to facilitate the application of convolutional neural networks that compute features used in PAR and EE estimation. First, each channel (X, Y, and Z) of the accelerometer and gyroscope sensor signals was rescaled individually to [0,255] by Equation (1).
X s c a l e d = N M + X X m i n X m a x X m i n × N M a
where NM = 0 is the new minima; NMa = 255 is the new maxima; Xscaled = scaled sensor signal; X = sensor signal; Xmin = minimum value of sensor signal; and Xmax = maximum value of sensor signal. The values of X m i n and X m a x were obtained with the bounds function in MATLAB.
Then, the fractional part of the converted value was removed with floor operation (Figure 3d) in MATLAB. Next, the rescaled sensor signals were grouped together into the following three combinations: accelerometer and gyroscope (Acc+Gyr); accelerometer only (Acc); and gyroscope only (Gyr), as shown in Figure 3.
A sliding window of 10 s was applied to the grouped sensor data, which was then saved as a grayscale image (Figure 4). The choice of window size significantly affects EE and PAR performance. A shorter window aids in fast-tracking and transitions but may result in false positives. Previous studies [53] suggest a 10 s window with a 55 Hz sampling rate is optimal for activity recognition. Further research [54] examined various sensors and data from 20 participants in 52 activities, indicating that longer windows generally improve PAR, except for short activities. A parallel trend emerged in EE-focused research articles. For instance, in [42], researchers utilized a 10 s window for EE estimation. Subsequently, a more exhaustive study detailed in [55], encompassing 100 young participants, underscored that a 10-s window size yields a reduced estimation error for EE. Hence, we chose a 10 s window for a balanced PAR and EE estimation performance. The sliding window operation was performed for each activity and participant’s data, with no overlap between windows from different activities or participants.
The training and validation datasets were produced using sequential and random windows (Figure 3b,c). The sequential window was precisely aligned with the activity boundaries, and the ground truth was generated based on the corresponding activity label and EE (Figure 3b). In contrast, the starting point of random windows (Figure 3c) was selected randomly, which can be considered a form of data augmentation. The testing dataset was produced using only sequential windows. Grayscale images from the sensor combination Acc+Gyr, Acc, and Gyr were saved for both AIM-2 and AG devices. Image tensor sizes (heigh×twidth×chanel) were 1280 × 6 × 1 for Acc+Gyr, 1280 × 3 × 1 for Acc, 1280 × 3 × 1 for Gyr in AIM-2; and 1000 × 6 × 1 for Acc+Gyr, 1000 × 3 × 1 for Acc, 1000 × 3 × 1 for Gyr in AG.

3.4. Proposed Model for PAR and EE Estimation

Our proposed model (Figure 5) integrates the computation of PAR and EE estimations into a single operation. The IMU sensor data are converted into grayscale images and supplied as input to the Multitasking-CNN model. Given the pivotal role of body weight in determining EE, this value was incorporated as an additional external input to the Multitasking-CNN. Multitasking-CNN comprises a CNN model with a fully connected dense layer. Despite the computational complexity of CNN-based deep learning models, they eliminate the need for handcrafted feature selection, a step in traditional signal processing methods. The latest microcontrollers include low-power hardware for acceleration of inference in deep neural networks, thus enabling the use of deep learning in wearables. Therefore, we opted for a deep learning-based approach. The subsequent section provides a more detailed description of Multitasking-CNN.

3.5. Multitasking-CNN Architecture

Multitasking demonstrates superiority over the single-task model when applied to a set of interrelated tasks. Therefore, we chose to adopt a multitasking architecture for our experiment. A typical multitasking model architecture is designed to handle multiple related tasks simultaneously. Generally, a multitasking model consists of an input layer, a shared layer, a task-specific layer, and a head. In Multitasking-CNN (Figure 6), the shared layer consists of the CNN blocks that apply filters to the input data by performing convolution operations. Each CNN feature extraction block consists of a series of layers, including convolution, average pooling, batch normalization, and dropout. A head is the final part of a multitasking model that produces the desired outputs or predictions. The proposed Multitasking-CNN model has a separate head for PAR and EE estimation (Head-1 and Head-2 in Figure 6).
To determine the optimal configuration for the CNN blocks (number of kernels, filter size, and the number of neurons in dense layers), a grid search technique was employed by monitoring the model’s PAR performance. Twelve combinations (as presented in Table 2 were evaluated and compared to determine the optimal CNN feature block combination for EE estimation. As an example, in Table 2, the entry ‘block 1 + 2 + 3’ signifies that Head-2 was trained using the features extracted from the output of CNN blocks 1, 2, and 3. The Multitasking-CNN model achieved the best EE estimation when Head-2 took features from CNN blocks 1 and 2. Thus, this configuration was selected as our proposed model and used for model performance analysis. The model predicts the EE over a 10-s interval, and the final steady-state EE was calculated as the average EE across the entire activity.

3.6. Multitasking-CNN Training

The commonly used multitasking training approaches are hard parameter sharing, soft parameter sharing, task-specific layers, progressive training, and reinforcement learning with auxiliary tasks [56]. In this study, we evaluated two-step progressive training and one-step hard parameter sharing approaches.
In the two-step progressive training, we trained the model for the task of PAR (Head-1). The model weights were then held constant, and Head-2 was trained for EE estimation.
PAR is a classification problem, whereas EE estimation is a regression problem. Thus, during training, the categorical cross-entropy loss function (Equation (2)) was used for PAR and the mean squared error loss function (Equation (3)) for EE estimation.
In the one-step hard parameter sharing approach, the model was trained for both PAR and EE estimation simultaneously, utilizing a combined loss function. The combined loss was calculated based on a weighted combination of PAR and EE estimation losses (Equation (4)). The optimal values of w 1 and w 2 were found by a grid search.
L o s s P A R = c = 1 M y i , c log ( P i , c )
where M = n u m b e r   o f   c l a s s e s and i = o b s e r v a t i o n   n u m b e r .
L o s s E E = 1 n i = 1 n ( y i y i ^ ) 2
where y i = a c t u a l   v a l u e   a n d   y i ^ = p r e d i c t e d   v a l u e .
L o s s c o m b i n e d = w 1 × L o s s P A R + w 2 × L o s s E E   and   w 1 + w 2 = 1
The model’s performance was evaluated by a 6-fold leave-one-out procedure, where the entire data from one (different for each fold) participant was used as the test set. The data from the remaining 5 participants were split by a 3:1 ratio into the training and validation sets, respectively. The training/validation/testing procedures were repeated 6 times, so that on each of the 6 folds, the training and validation data were completely independent from the testing data. The average performance across 6 folds was measured in terms of accuracy, precision, recall, and F1 score (equations are given in [32]) for the PAR classification problem and mean squared error (MSE) (Equation (5)) and mean absolute percentage error (MAPE) (Equation (6)) for the EE estimation regression problem. The following hyperparameters were used during training: a batch size of 100, an initial learning rate of 0.00002 with a 50% decay after 10 epochs, and a maximum of 200 epochs with an early stopping criterion at a patience value of 25.
M S E y , y ^ = 1 N   i = 0 N 1 ( y i y i ^ ) 2  
M A P E = 1 N i = 0 P 1 y i y i ^ y i  
where N = S a m p l e   N u m b e r ,   y = a c t u a l ,   a n d   y ^ = p r e d i c t e d   v a l u e .

4. Results and Discussion

4.1. Two-Step Progressive Training

The mean accuracy, precision, recall, and F1 score for both the AIM-2 and AG devices are reported in Table 3. The highest (highlighted in bold) average PAR testing accuracy of 95% was achieved by a model utilizing gyroscope data from the AIM-2 device. In comparison, a testing accuracy of 82% was attained using accelerometer and gyroscope data from the AG device. The table provides evidence that the model achieves higher PAR performance with AIM-2 collected data compared to AG collected data, regardless of sensor combination.
Table 2 displays the MSE and MAPE results for twelve different CNN block combinations using accelerometer data. The MSE and MAPE values for Multitasking-CNN with various sensor combinations are tabulated in Table 4. The findings from both Table 2 and Table 4 show the features from CNN blocks 1 and 2 offer good performance (highlighted in bold) for EE estimation. Table 4 lists the testing MSE and MAPE for EE estimation for the two-step progressive training approach. The Multitasking-CNN model achieved the lowest (highlighted with bold) MSE and MAPE values for the AIM-2 device-collected data. The MSE value was 0.59 METs, and the MAPE value was 11%. In comparison, the AG-collected data had higher values, with an MSE of 0.73 METs and a MAPE of 12%.
The model achieves the lowest MSE and MAPE for accelerometer and gyroscope sensors in combination with the AIM-2 device-collected data. For the accelerometer sensor combination, the AIM-2 data result in an MSE of 0.75 METs and a MAPE of 14%, which is highly competitive compared to the AG device data with an MSE of 0.73 METs and a MAPE of 12%. The AG device data outperform the EE estimation with the gyroscope sensor (compared to AIM-2 gyroscope data). However, when considering overall performance, the AIM-2 device achieves the lowest or, at the very least, comparable results when compared to the AG device data.
Figure 7 presents the predicted vs. actual EE estimation for the best-performing models with data from the AIM-2 and AG devices. The EE estimation is depicted for six activities, including “seated computer work”, “sweeping”, “stationary cycling”, “treadmill walking with different paces”, “treadmill running”, and “one-vs.-one basketball” in that order. Note that participant 105 did not perform “seated computer work”, so “sweeping” was their first activity.
Overall, the results from two-step progressive training indicate that the AIM-2 device data yield higher PAR accuracy with testing accuracy values of 91%, 88.0%, and 95%, compared to the AG device data, which had testing accuracy values of 82%, 78.0%, and 80%. For the AIM-2 device, the combination of accelerometer and gyroscope sensors achieved the lowest MSE of 0.59 METs and MAPE of 11%, while for the AG device, the accelerometer sensor yielded the best EE estimation performance (MSE of 0.73 METs and MAPE of 12%). The Multitasking-CNN demonstrated a good prediction for “sweeping”, “stationary cycling”, “treadmill running”, and “one-vs.-one basketball” for both the AIM-2 and AG device data. Our proposed Multitasking-CNN model, trained in two steps, demonstrates superior PAR and EE predictions for both AIM-2 and AG device collected data.

4.2. One-Step Hard Parameter Sharing Training

Table 5 shows the results of the grid search for parameters w 1 and w 2 . The results in the table show that the higher the value of w 1 , the better the PAR performance is (accuracy, precision, recall, and F1 score). This is an expected behavior, as the loss function for PAR is prioritized over the EE loss function. However, the reverse effect was not observed if the value of w 2 was higher ( w 1 is lower in Table 5). The MSE and MAPE were higher when w 2 was set at 0.9 (i.e., w 1 = 0.1). The results from Table 5 suggest that a combination of w 1 = 0.3 and w 2 = 0.7 provides balanced performance in both PAR and EE estimation for the accelerometer sensor. Similar results were obtained for ‘Acc+Gyr’ and ‘Gyr’ sensor combinations.
Table 6 shows the PAR performance of one-step training for different sensor combinations. The model achieved the best PAR accuracy for gyroscope data from AIM-2, having an accuracy of 88%, 91% precision, 88% recall, and an 89% F1 score. AG device data achieve a maximum of 80%, 78%, 80%, and 79% accuracy, precision, recall, and F1 score with accelerometer sensor data. The table also shows that the model achieves higher PAR performance with AIM-2 collected data compared to AG collected data, irrespective of the sensor combination.
Table 7 shows the model EE performance for AIM-2 and AG collected data by using one-step hard parameter sharing training. The model achieved the best MSE and MAPE with accelerometer sensor data for both AIM-2 and AG devices. The best-performing EE estimation model achieved minimum MSE and MAPE of 2.69 METs and 28% for AIM-2 and 2.79 METs and 30% for AG, respectively. The model achieved lower MSE and MAPE for ‘Acc’ and ‘Gyr’ sensor combination with AIM-2 collected data (with compared to AG collected data). However, when considering the ‘Acc+Gyr’ sensor combination from the AG device, the model achieved lower MSE and MAPE compared to the AIM-2 device. In summary, the AIM-2 device consistently achieved comparable results when compared to the AG device data.
The result from one-step hard parameter sharing and two-step progressive training concludes that the two-step training method achieves better performance for both PAR (maximum accuracy, precision, recall, and F1 score of 95%, 96%, 95%, and 95%, respectively) and EE estimation (lowest MSE and MAPE of 0.59 METs and 11%). The findings indicate that when utilizing eyeglass-mounted AIM-2 collected data, it consistently outperforms the AG data in terms of PAR, irrespective of the sensor combination and training strategy employed. Regarding EE estimation, the model achieved the lowest MSE and MAPE when using AIM-2 data, regardless of whether the one-step or two-step training method was employed. This research outcome suggests that eyeglass-mounted wearable sensor devices could be a potential platform for comprehensive monitoring of physical activity and energy expenditure.
Although the models achieve the highest PAR and EE estimation using different sensor combinations, there are notable variations. For instance, in the case of AIM-2 collected data, the gyroscope signal demonstrates the highest PAR accuracy (Table 3), while the accelerometer and gyroscope signals achieve low MSE and MAPE (Table 4). However, a well-balanced performance in PAR and EE estimation is consistently observed when using accelerometer and gyroscope signals, regardless of the device and training strategy employed.
Table 8 presents a comparative view of PAR and EE estimation using wearable sensor devices in various body locations. Our focus is on simultaneous PAR and EE estimation, aligning with our experiment. A direct comparison of model performance is challenging due to differences in data collection procedures, performance metrics, and physical activities. We provide a comparative overview of our experiment’s pros and cons (compared to our experiment). Our primary contribution is experimenting with simultaneous PAR and EE monitoring using the eyeglass-mounted AIM-2 device, offering new possibilities. The AIM-2 device, which can be easily affixed to any eyeglass, eliminates the need for additional gadgets, a significant advantage over other methods (Table 8). Our AIM-2-based experiments also offer advantages such as collecting ground truth EE from COSMED-K5 (compared with [45]), convenience (compared with all methods in Table 8), and no handcrafted feature selection (compared with [26]). Our experiment’s limitations are discussed in Section 5. Numerically, using the AIM-2 device, we achieved a maximum testing accuracy of 95% for PAR, a minimum MSE of 0.59 METs, and an 11% MAPE in EE estimation compared to methods listed in Table 8.

5. Conclusions

We have developed a multitasking model called Multitasking-CNN for estimating PAR and steady-state EE using data collected from the AIM-2 and AG devices. The model was trained using both a two-step progressive training approach and a one-step hard parameter sharing approach. Our results indicate that the two-step progressive training approach outperforms the one-step hard parameter-sharing approach. With the AIM-2 device data and two-step progressive training, the model achieves a maximum testing accuracy of 95% for PAR estimation, a minimum MSE of 0.59 METs for EE estimation, and an MAPE of 11%. Conversely, for the AG device data, the model achieves a maximum testing accuracy of 82% for PAR estimation, a minimum MSE of 0.73 METs for EE estimation, and an MAPE of 13%.
Regardless of the training strategy and sensor combination, the models consistently demonstrate higher PAR performance when using the AIM-2 data. For EE estimation, the model achieves the lowest MSE and MAPE for AIM-2 collected data, irrespective of training strategy. These findings suggest that eyeglass-mounted wearable devices hold promising potential for further research in PAR and EE estimation.
This experiment has some limitations. The human study was cut short in 2020 due to the COVID-19 pandemic, and for logistical reasons, the study could not be restarted in 2022, when human metabolic research became possible again. The small number of study participants is a major limitation that impacts the ability to draw broad conclusions. The findings may not accurately reflect the responses of broader demographic groups, such as different age ranges, ethnicities, or socioeconomic backgrounds. Moreover, biological and physiological differences between males and females could influence the outcomes of the study. With only one female participant, the result might skew toward the male gender. Finally, the narrow range of ages (from approximately 20 to 30 years) and BMI values may restrict the generalizability of the findings to other age groups or individuals with different BMI ranges. However, we used rigorous (e.g., leave-one-out) training and validation methods to derive generalizable results (based on available data) that justify further work with head-worn sensors. Future research with larger and more diverse samples is necessary to confirm and extend these results to broader populations.
Another limitation is that the experiment was carried out in a controlled environment with a small number of activities. Moreover, the experiment only compares the eyeglass-mounted AIM-2 with a hip-worn wearable sensor device. Nonetheless, our goal is to demonstrate the feasibility of using eyeglass-mounted wearable sensor devices for PAR and EE estimation. A more rigorous study with a larger population, a protocol inclusive of more activities of daily living, and more sophisticated learning models (such as LSTM or CNN-LSTM) needs to be conducted to generalize the performance. The limitations of this work will be addressed in our future work.
Future research could delve into AIM-2’s potential for comprehensive monitoring of both energy intake and EE, which would aid in achieving energy balance and potentially mitigating the risk of chronic metabolic syndromes.

Author Contributions

Conceptualization, investigation, methodology, validation, data curation, and writing—original draft preparation: M.B.H.; writing—review and editing, and study data collection: S.R.L. and S.E.C.; writing—review and editing: E.L.M.; supervision, project administration, and writing—review and editing: E.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board (or Ethics Committee) of The University of Tennessee Knoxville (IRB#: UTK IRB-14-01988-XP).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data are available upon reasonable and responsible request for non-commercial research purposes.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. von Loeffelholz, C.; Birkenfeld, A. The Role of Non-Exercise Activity Thermogenesis in Human Obesity. In Endotext [Internet]; MDText.com: South Dartmouth, MA, USA, 2018. [Google Scholar]
  2. Poehlman, E.T. A Review: Exercise and Its Influence on Resting Energy Metabolism in Man. Med. Sci. Sports Exerc. 1989, 21, 515–525. [Google Scholar] [CrossRef]
  3. Sazonov, E.S.; Schuckers, S. The Energetics of Obesity: A Review: Monitoring Energy Intake and Energy Expenditure in Humans. IEEE Eng. Med. Biol. Mag. 2010, 29, 31–35. [Google Scholar] [CrossRef] [PubMed]
  4. Hill, J.O.; Wyatt, H.R.; Peters, J.C. The Importance of Energy Balance. Eur. Endocrinol. 2013, 9, 111. [Google Scholar] [CrossRef] [PubMed]
  5. Hill, J.O.; Commerford, R. Physical Activity, Fat Balance, and Energy Balance. Int. J. Sport Nutr. Exerc. Metab. 1996, 6, 80–92. [Google Scholar] [CrossRef] [PubMed]
  6. Abdel-Hamid, T.; Ankel, F.; Battle-Fisher, M.; Gibson, B.; Gonzalez-Parra, G.; Jalali, M.; Kaipainen, K.; Kalupahana, N.; Karanfil, O.; Marathe, A.; et al. Public and Health Professionals’ Misconceptions about the Dynamics of Body Weight Gain/Loss. Syst. Dyn. Rev. 2014, 30, 58–74. [Google Scholar] [CrossRef] [PubMed]
  7. World Health Organization, T. Global Recommendations on Physical Activity for Health; World Health Organization: Geneva, Switzerland, 2010; ISBN 92-4-159997-9. [Google Scholar]
  8. United Nations University; World Health Organization. Human Energy Requirements: Report of a Joint FAO/WHO/UNU Expert Consultation: Rome, 17–24 October 2001; Food & Agriculture Organization: Rome, Italy, 2004; Volume 1, ISBN 92-5-105212-3. [Google Scholar]
  9. Prevention Strategies & Guidelines|Overweight & Obesity|CDC. Available online: https://www.cdc.gov/obesity/resources/strategies-guidelines.html (accessed on 23 May 2023).
  10. Caspersen, C.J.; Powell, K.E.; Christenson, G.M. Physical Activity, Exercise, and Physical Fitness: Definitions and Distinctions for Health-Related Research. Public Health Rep. 1985, 100, 126. [Google Scholar] [PubMed]
  11. Ainsworth, B.E.; Haskell, W.L.; Herrmann, S.D.; Meckes, N.; Bassett Jr, D.R.; Tudor-Locke, C.; Greer, J.L.; Vezina, J.; Whitt-Glover, M.C.; Leon, A.S. 2011 Compendium of Physical Activities: A Second Update of Codes and MET Values. Med. Sci. Sports Exerc. 2011, 43, 1575–1581. [Google Scholar] [CrossRef]
  12. Dawar, N.; Kehtarnavaz, N. Continuous Detection and Recognition of Actions of Interest among Actions of Non-Interest Using a Depth Camera. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 4227–4231. [Google Scholar]
  13. Eum, H.; Yoon, C.; Lee, H.; Park, M. Continuous Human Action Recognition Using Depth-MHI-HOG and a Spotter Model. Sensors 2015, 15, 5197–5227. [Google Scholar] [CrossRef]
  14. Ziaeefard, M.; Bergevin, R. Semantic Human Activity Recognition: A Literature Review. Pattern Recognit. 2015, 48, 2329–2345. [Google Scholar] [CrossRef]
  15. Chaaraoui, A.A.; Padilla-López, J.R.; Ferrández-Pastor, F.J.; Nieto-Hidalgo, M.; Flórez-Revuelta, F. A Vision-Based System for Intelligent Monitoring: Human Behaviour Analysis and Privacy by Context. Sensors 2014, 14, 8895–8925. [Google Scholar] [CrossRef]
  16. Jalal, A.; Kamal, S.; Kim, D. A Depth Video Sensor-Based Life-Logging Human Activity Recognition System for Elderly Care in Smart Indoor Environments. Sensors 2014, 14, 11735–11759. [Google Scholar] [CrossRef] [PubMed]
  17. Zheng, Y.-L.; Ding, X.-R.; Poon, C.C.Y.; Lo, B.P.L.; Zhang, H.; Zhou, X.-L.; Yang, G.-Z.; Zhao, N.; Zhang, Y.-T. Unobtrusive Sensing and Wearable Devices for Health Informatics. IEEE Trans. Biomed. Eng. 2014, 61, 1538–1554. [Google Scholar] [CrossRef] [PubMed]
  18. Dai, J.; Bai, X.; Yang, Z.; Shen, Z.; Xuan, D. PerFallD: A Pervasive Fall Detection System Using Mobile Phones. In Proceedings of the 2010 8th IEEE International Conference on Pervasive Computing and Communications Workshops (PERCOM Workshops), Mannheim, Germany, 29 March–2 April 2010; pp. 292–297. [Google Scholar]
  19. Fontecha, J.; Navarro, F.J.; Hervás, R.; Bravo, J. Elderly Frailty Detection by Using Accelerometer-Enabled Smartphones and Clinical Information Records. Pers. Ubiquitous Comput. 2013, 17, 1073–1083. [Google Scholar] [CrossRef]
  20. Sazonov, E.; Hegde, N.; Browning, R.C.; Melanson, E.L.; Sazonova, N.A. Posture and Activity Recognition and Energy Expenditure Estimation in a Wearable Platform. IEEE J. Biomed. Health Inform. 2015, 19, 1339–1346. [Google Scholar] [CrossRef] [PubMed]
  21. De Bois, M.; Amroun, H.; Ammi, M. Energy Expenditure Estimation through Daily Activity Recognition Using a Smart-Phone. In Proceedings of the 2018 IEEE 4th World Forum on Internet of Things (WF-IoT), Singapore, 5–8 February 2018; pp. 167–172. [Google Scholar]
  22. Rahn, V.X.; Zhou, L.; Klieme, E.; Arnrich, B. Optimal Sensor Placement for Human Activity Recognition with a Minimal Smartphone-IMU Setup. In Proceedings of the SENSORNETS, Online, 9–10 February 2021; pp. 37–48. [Google Scholar]
  23. White, T.; Westgate, K.; Hollidge, S.; Venables, M.; Olivier, P.; Wareham, N.; Brage, S. Estimating Energy Expenditure from Wrist and Thigh Accelerometry in Free-Living Adults: A Doubly Labelled Water Study. Int. J. Obes. 2019, 43, 2333–2342. [Google Scholar] [CrossRef]
  24. wGT3X-BT|ActiGraph Wearable Devices. Available online: https://theactigraph.com/actigraph-wgt3x-bt (accessed on 23 May 2023).
  25. Guediri, A.; Robin, L.; Lacroix, J.; Aubourg, T.; Vuillerme, N.; Mandigout, S. Comparison of Energy Expenditure Assessed Using Wrist-and Hip-Worn ActiGraph GT3X in Free-Living Conditions in Young and Older Adults. Front. Med. 2021, 8, 696968. [Google Scholar] [CrossRef] [PubMed]
  26. Davoudi, A.; Mardini, M.T.; Nelson, D.; Albinali, F.; Ranka, S.; Rashidi, P.; Manini, T.M. The Effect of Sensor Placement and Number on Physical Activity Recognition and Energy Expenditure Estimation in Older Adults: Validation Study. JMIR Mhealth Uhealth 2021, 9, e23681. [Google Scholar] [CrossRef] [PubMed]
  27. Gjoreski, H.; Kiprijanovska, I.; Stankoski, S.; Kalabakov, S.; Broulidakis, J.; Nduka, C.; Gjoreski, M. Head-Ar: Human Activity Recognition with Head-Mounted Imu Using Weighted Ensemble Learning. Act. Behav. Comput. 2021, 204, 153–167. [Google Scholar] [CrossRef] [PubMed]
  28. Hossain, T.; Islam, M.S.; Ahad, M.A.R.; Inoue, S. Human Activity Recognition Using Earable Device. In Proceedings of the Adjunct Proceedings of the 2019 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2019 ACM International Symposium on Wearable Computers, London, UK, 9–13 September 2019; pp. 81–84. [Google Scholar]
  29. Farooq, M.; Sazonov, E. A Novel Wearable Device for Food Intake and Physical Activity Recognition. Sensors 2016, 16, 1067. [Google Scholar] [CrossRef]
  30. Farooq, M.; Sazonov, E. Real Time Monitoring and Recognition of Eating and Physical Activity with a Wearable Device Connected to the Eyeglass. In Proceedings of the 2017 Eleventh International Conference on Sensing Technology (ICST), Sydney, Australia, 4–6 December 2017; pp. 1–6. [Google Scholar]
  31. Why Do People Wear Glasses and Why Do People Need Glasses? Available online: https://www.allaboutvision.com/eyeglasses/faq/why-people-wear-glasses/ (accessed on 8 November 2023).
  32. Eyewear Industry Statistics. Available online: https://www.overnightglasses.com/eyewear-industry-statistics/ (accessed on 8 November 2023).
  33. Demrozi, F.; Pravadelli, G.; Bihorac, A.; Rashidi, P. Human Activity Recognition Using Inertial, Physiological and Environmental Sensors: A Comprehensive Survey. IEEE Access 2020, 8, 210816–210836. [Google Scholar] [CrossRef]
  34. Alvarez-Garcia, J.A.; Cvetković, B.; Luštrek, M. A Survey on Energy Expenditure Estimation Using Wearable Devices. ACM Comput. Surv. (CSUR) 2020, 53, 1–35. [Google Scholar] [CrossRef]
  35. Khatun, M.A.; Yousuf, M.A.; Ahmed, S.; Uddin, M.Z.; Alyami, S.A.; Al-Ashhab, S.; Akhdar, H.F.; Khan, A.; Azad, A.; Moni, M.A. Deep CNN-LSTM with Self-Attention Model for Human Activity Recognition Using Wearable Sensor. IEEE J. Transl. Eng. Health Med. 2022, 10, 1–16. [Google Scholar] [CrossRef] [PubMed]
  36. Trujillo-Guerrero, M.F.; Román-Niemes, S.; Jaén-Vargas, M.; Cadiz, A.; Fonseca, R.; Serrano-Olmedo, J.J. Accuracy Comparison of CNN, LSTM, and Transformer for Activity Recognition Using IMU and Visual Markers. IEEE Access 2023, 11, 106650–106669. [Google Scholar] [CrossRef]
  37. Sun, Y.G.; Kim, S.H.; Lee, S.; Seon, J.; Lee, S.; Kim, C.G.; Kim, J.Y. Performance of End-to-End Model Based on Convolutional LSTM for Human Activity Recognition. J. Web Eng. 2022, 21, 1671–1689. [Google Scholar] [CrossRef]
  38. Choudhury, N.A.; Soni, B. An Adaptive Batch Size Based-CNN-LSTM Framework for Human Activity Recognition in Uncontrolled Environment. IEEE Trans. Ind. Inform. 2023, 19, 10379–10387. [Google Scholar] [CrossRef]
  39. Phukan, N.; Mohine, S.; Mondal, A.; Manikandan, M.S.; Pachori, R.B. Convolutional Neural Network-Based Human Activity Recognition for Edge Fitness and Context-Aware Health Monitoring Devices. IEEE Sens. J. 2022, 22, 21816–21826. [Google Scholar] [CrossRef]
  40. Albeshri, A. SVSL: A Human Activity Recognition Method Using Soft-Voting and Self-Learning. Algorithms 2021, 14, 245. [Google Scholar] [CrossRef]
  41. Zhu, J.; Pande, A.; Mohapatra, P.; Han, J.J. Using Deep Learning for Energy Expenditure Estimation with Wearable Sensors. In Proceedings of the 2015 17th International Conference on E-health Networking, Application & Services (HealthCom), Boston, MA, USA, 14–17 October 2015; pp. 501–506. [Google Scholar]
  42. Rothney, M.P.; Neumann, M.; Béziat, A.; Chen, K.Y. An Artificial Neural Network Model of Energy Expenditure Using Nonintegrated Acceleration Signals. J. Appl. Physiol. 2007, 103, 1419–1427. [Google Scholar] [CrossRef] [PubMed]
  43. Al-Janabi, S.; Salman, A.H. Sensitive Integration of Multilevel Optimization Model in Human Activity Recognition for Smartphone and Smartwatch Applications. Big Data Min. Anal. 2021, 4, 124–138. [Google Scholar] [CrossRef]
  44. Crawshaw, M. Multi-Task Learning with Deep Neural Networks: A Survey. arXiv 2020, arXiv:2009.09796. [Google Scholar]
  45. Saris, W.H.M.; Blair, S.N.; Van Baak, M.A.; Eaton, S.B.; Davies, P.S.W.; Di Pietro, L.; Fogelholm, M.; Rissanen, A.; Schoeller, D.; Swinburn, B. How Much Physical Activity Is Enough to Prevent Unhealthy Weight Gain? Outcome of the IASO 1st Stock Conference and Consensus Statement. Obes. Rev. 2003, 4, 101–114. [Google Scholar] [CrossRef] [PubMed]
  46. Barut, O.; Zhou, L.; Luo, Y. Multitask LSTM Model for Human Activity Recognition and Intensity Estimation Using Wearable Sensor Data. IEEE Internet Things J. 2020, 7, 8760–8768. [Google Scholar] [CrossRef]
  47. Nakamura, K.; Yeung, S.; Alahi, A.; Li, F.-F. Jointly Learning Energy Expenditures and Activities Using Egocentric Multimodal Signals. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; Volume 2017, pp. 6817–6826. [Google Scholar]
  48. Doulah, A.; Ghosh, T.; Hossain, D.; Imtiaz, M.H.; Sazonov, E. “Automatic Ingestion Monitor Version 2”–a Novel Wearable Device for Automatic Food Intake Detection and Passive Capture of Food Images. IEEE J. Biomed. Health Inform. 2020, 25, 568–576. [Google Scholar] [CrossRef] [PubMed]
  49. GT9X Link|ActiGraph Wearable Devices. Available online: https://theactigraph.com/actigraph-link (accessed on 23 May 2023).
  50. COSMED—K5: Wearable Metabolic System for Both Laboratory and Field Testing. Available online: https://www.cosmed.com/en/products/cardio-pulmonary-exercise-test/k5 (accessed on 23 May 2023).
  51. Crouter, S.E.; LaMunion, S.R.; Hibbing, P.R.; Kaplan, A.S.; Bassett, D.R., Jr. Accuracy of the Cosmed K5 Portable Calorimeter. PLoS ONE 2019, 14, e0226290. [Google Scholar] [CrossRef] [PubMed]
  52. Brønd, J.C.; Arvidsson, D. Sampling Frequency Affects the Processing of Actigraph Raw Acceleration Data to Activity Counts. J. Appl. Physiol. 2016, 120, 362–369. [Google Scholar] [CrossRef] [PubMed]
  53. Niazi, A.H.; Yazdansepas, D.; Gay, J.L.; Maier, F.W.; Ramaswamy, L.; Rasheed, K.; Buman, M.P. Statistical Analysis of Window Sizes and Sampling Rates in Human Activity Recognition. In Proceedings of the HEALTHINF, Porto, Portugal, 21–23 February 2017; pp. 319–325. [Google Scholar]
  54. Munguia Tapia, E. Using Machine Learning for Real-Time Activity Recognition and Estimation of Energy Expenditure. Doctoral dissertation, Massachusetts Institute of Technology, Cambridge, MA, USA, 2008. [Google Scholar]
  55. Trost, S.G.; Wong, W.-K.; Pfeiffer, K.A.; Zheng, Y. Artificial Neural Networks to Predict Activity Type and Energy Expenditure in Youth. Med. Sci. Sports Exerc. 2012, 44, 1801. [Google Scholar] [CrossRef]
  56. Ruder, S. An Overview of Multi-Task Learning in Deep Neural Networks. arXiv 2017, arXiv:1706.05098. [Google Scholar]
  57. Altini, M.; Penders, J.; Vullers, R.; Amft, O. Combining Wearable Accelerometer and Physiological Data for Activity and Energy Expenditure Estimation. In Proceedings of the 4th Conference on Wireless Health, Baltimore, MD, USA, 1–3 November 2013; pp. 1–8. [Google Scholar]
Figure 1. Data collection by wearable sensor devices for the Multitasking-CNN model.
Figure 1. Data collection by wearable sensor devices for the Multitasking-CNN model.
Sensors 24 03046 g001
Figure 2. Sample accelerometer and gyroscope raw sensor data (20 s) collected from participant 101 by the AIM-2 device ((a,b): basketball activity; (c,d): running activity).
Figure 2. Sample accelerometer and gyroscope raw sensor data (20 s) collected from participant 101 by the AIM-2 device ((a,b): basketball activity; (c,d): running activity).
Sensors 24 03046 g002
Figure 3. Sensor data processing for Multitasking-CNN model development.
Figure 3. Sensor data processing for Multitasking-CNN model development.
Sensors 24 03046 g003
Figure 4. Sample final output (grayscale image) of sensor signal to image conversion for different participants and activities for Acc+Gyr.
Figure 4. Sample final output (grayscale image) of sensor signal to image conversion for different participants and activities for Acc+Gyr.
Sensors 24 03046 g004
Figure 5. Overall architecture of PAR and EE estimations.
Figure 5. Overall architecture of PAR and EE estimations.
Sensors 24 03046 g005
Figure 6. Multitasking-CNN model.
Figure 6. Multitasking-CNN model.
Sensors 24 03046 g006
Figure 7. AG (Acc) vs. AIM-2 (Acc+Gyr) for EE estimation with progressive training.
Figure 7. AG (Acc) vs. AIM-2 (Acc+Gyr) for EE estimation with progressive training.
Sensors 24 03046 g007
Table 1. Duration of data collection (in minutes) for PAR and EE estimation.
Table 1. Duration of data collection (in minutes) for PAR and EE estimation.
ActivityParticipant ID
101102103104105106
Computer65.566--6
Sweeping5.565.5665
Cycling65.56666
Walking3 mph (0% grade)5.56.56666
4 mph (0% grade)5.54.55.565.55.5
3 mph (5% grade)55.55.565.55.5
Running at 6 mph5.55.55.55.55.55.5
Basketball65.56666
Total duration (in minutes)4544.54647.540.545.5
Table 2. Testing MSE and MAPE for EE estimation using different CNN block combinations (accelerometer sensor only).
Table 2. Testing MSE and MAPE for EE estimation using different CNN block combinations (accelerometer sensor only).
CNN Blocks to Head 2AIM-2 Sensor DataAG Sensor Data
MSEMAPEMSEMAPE
Block-10.820.140.870.12
Block-21.750.181.400.19
Block-30.970.231.580.22
Block-42.780.253.580.32
Block 1 + 40.670.151.030.16
Block 1 + 20.750.140.730.12
Block 1 + 30.650.130.990.15
Block 2 + 31.330.201.980.23
Block 2 + 41.780.211.690.22
Block 1 + 2 + 30.700.140.910.13
Block 2 + 3 + 41.340.201.700.23
Block 1 + 2 + 3 + 40.770.150.880.15
Table 3. Average testing accuracy, precision, recall, and F1 score (testing) for par with two-step progressive training.
Table 3. Average testing accuracy, precision, recall, and F1 score (testing) for par with two-step progressive training.
Average Metric (%)Using AIM-2 Sensor Data Using AG Sensor Data
Acc+GyrAccGyrAcc+GyrAccGyr
Accuracy918895827880
Precision959296837585
Recall908795817781
F1928995827683
Table 4. Testing MSE and MAPE for EE estimation by using two-step progressive training.
Table 4. Testing MSE and MAPE for EE estimation by using two-step progressive training.
Block
Combination
AIM-2 Sensor DataAG Sensor Data
Acc+GyrAccGyrAcc+GyrAccGyr
MSEMAPEMSEMAPEMSEMAPEMSEMAPEMSEMAPEMSEMAPE
Block 1 + 20.590.110.750.142.030.231.380.150.730.121.670.19
Block 1 + 30.660.120.650.131.520.221.020.140.990.151.480.19
Table 5. PAR and EE estimation testing performance for different w1 (accelerometer sensor only).
Table 5. PAR and EE estimation testing performance for different w1 (accelerometer sensor only).
w 1 AIM-2 Sensor DataAG Sensor Data
PAREE EstimationPAREE Estimation
Accuracy (%)Precision
(%)
Recall
(%)
MSEMAPEAccuracy
(%)
Precision
(%)
Recall
(%)
MSEMAPE
0.17279722.690.277571753.021.45
0.27878782.710.277977792.850.31
0.38086802.690.288078802.790.30
0.48086802.740.288078802.850.29
0.58287822.730.287878782.890.31
0.68387832.850.307876782.850.29
0.78489842.890.307878783.160.32
0.88489842.940.317876784.460.36
0.98590852.830.307876782.980.29
Table 6. PAR testing accuracy for AIM-2 and AG devices using one-step hard parameter sharing training.
Table 6. PAR testing accuracy for AIM-2 and AG devices using one-step hard parameter sharing training.
Block - 1 + 2 ,  
w 1 = 0.3   w 2 = 0.7
Average Metric
(%)
AIM-2 Sensor DataAG Sensor Data
Acc+GyrAccGyrAcc+GyrAccGyr
Accuracy858088768069
Precision878691747873
Recall 858088768069
F1868289757971
Table 7. Testing MSE and MAPE for EE estimation by using one-step hard parameter sharing training.
Table 7. Testing MSE and MAPE for EE estimation by using one-step hard parameter sharing training.
Block - 1 + 2 ,  
w 1 = 0.3   w 2 = 0.7
AIM-2 Sensor DataAG Sensor Data
Acc+GyrAccGyrAcc+GyrAccGyr
MSEMAPEMSEMAPEMSEMAPEMSEMAPEMSEMAPEMSEMAPE
3.900.312.690.283.280.282.930.282.790.304.300.32
Table 8. Comparison of studies addressing simultaneous PAR and EE estimations.
Table 8. Comparison of studies addressing simultaneous PAR and EE estimations.
StudyPosition of WearSensor or Device UsedNumber of Activities and ParticipantsMaximum PAR PerformanceMaximum EE PerformancePros and Cons (Compared to Our Experiment)
[46]waistActiGraph GT9X (acc, gyr, and mag)7 activities from 10 adultsF1 score of 0.83 (L1-64) and
accuracy of 0.98 (L1-64)
MAE of 0.55 (L2-64) METs and
MAPE of 17.56% (L2-64)
Cons:
-
EE (MET) calculated from the Friedson adult formula.
-
Requires an additional gadget for monitoring.
Pros:
-
Model validated across various datasets.
[47]chest and wristCell phone (camera and acc) and heart rate sensors24 activities from 10 participantsmean average precision of 78.1Absolute error of 0.696 kcal.m−1Cons:
-
Inconvenience of wearing a cell phone or wearable device on the chest.
-
EE ground truth is taken from a predefined value.
-
Privacy concerns of bystanders due to video recording.
Pros:
-
Data collected in a free-living environment.
-
Potential for higher accuracy with the integration of visual and sensor data.
[20]footSmartShoe (force-sensitive resistors and acc sensor)15 activities from 15 participantsclassification accuracy of 95%Root mean square error (RMSE) of 0.78 kcal/minCons:
-
Inconvenience of wearing shoes even during rest (e.g., supine and sitting).
-
Requires two different models for EE estimation and PAR.
Pros:
-
Reduced model complexity, suitable for real-time recognition.
-
EE ground truth is measured via indirect calorimeter.
[26]wrist, hip, ankle, upper arm, and thighActiGraph GT3X (acc, gyr, and mag)32 activities from 93 participantsbalanced accuracy for individual activity recognition of 0.42 with the waist sensorRMSE of 0.91 (hip or thigh position)Cons:
-
Requires two separate models for EE estimation and PAR.
-
Involves handcrafted feature selection for model development.
-
Requires additional gadgets for monitoring.
Pros:
-
Includes a larger number of activities and participants.
-
Compares results across different body positions.
[57]neck and wristImec’s ECG6 cluster activities from 16 participantsaccuracy of 95%, including the accelerometer and heart rateRMSE of 1.59 kcal/min with an accelerometer and heart rateCons:
-
Discomfort from two gel electrodes on participants’ chests.
-
Requires additional monitoring gadgets.
Pros:
-
Investigate the benefits of three physiological signals, i.e., respiration rate, galvanic skin response, and skin humidity.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hossain, M.B.; LaMunion, S.R.; Crouter, S.E.; Melanson, E.L.; Sazonov, E. A CNN Model for Physical Activity Recognition and Energy Expenditure Estimation from an Eyeglass-Mounted Wearable Sensor. Sensors 2024, 24, 3046. https://doi.org/10.3390/s24103046

AMA Style

Hossain MB, LaMunion SR, Crouter SE, Melanson EL, Sazonov E. A CNN Model for Physical Activity Recognition and Energy Expenditure Estimation from an Eyeglass-Mounted Wearable Sensor. Sensors. 2024; 24(10):3046. https://doi.org/10.3390/s24103046

Chicago/Turabian Style

Hossain, Md Billal, Samuel R. LaMunion, Scott E. Crouter, Edward L. Melanson, and Edward Sazonov. 2024. "A CNN Model for Physical Activity Recognition and Energy Expenditure Estimation from an Eyeglass-Mounted Wearable Sensor" Sensors 24, no. 10: 3046. https://doi.org/10.3390/s24103046

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop