Next Article in Journal
Comparing Video Analysis to Computerized Detection of Limb Position for the Diagnosis of Movement Control during Back Squat Exercise with Overload
Previous Article in Journal
A Review on Traversability Risk Assessments for Autonomous Ground Vehicles: Methods and Metrics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PAR-Net: An Enhanced Dual-Stream CNN–ESN Architecture for Human Physical Activity Recognition

Mixed Reality and Interaction Lab, Department of Software, Sejong University, Seoul 05006, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(6), 1908; https://doi.org/10.3390/s24061908
Submission received: 16 January 2024 / Revised: 5 March 2024 / Accepted: 12 March 2024 / Published: 16 March 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Physical exercise affects many facets of life, including mental health, social interaction, physical fitness, and illness prevention, among many others. Therefore, several AI-driven techniques have been developed in the literature to recognize human physical activities. However, these techniques fail to adequately learn the temporal and spatial features of the data patterns. Additionally, these techniques are unable to fully comprehend complex activity patterns over different periods, emphasizing the need for enhanced architectures to further increase accuracy by learning spatiotemporal dependencies in the data individually. Therefore, in this work, we develop an attention-enhanced dual-stream network (PAR-Net) for physical activity recognition with the ability to extract both spatial and temporal features simultaneously. The PAR-Net integrates convolutional neural networks (CNNs) and echo state networks (ESNs), followed by a self-attention mechanism for optimal feature selection. The dual-stream feature extraction mechanism enables the PAR-Net to learn spatiotemporal dependencies from actual data. Furthermore, the incorporation of a self-attention mechanism makes a substantial contribution by facilitating targeted attention on significant features, hence enhancing the identification of nuanced activity patterns. The PAR-Net was evaluated on two benchmark physical activity recognition datasets and achieved higher performance by surpassing the baselines comparatively. Additionally, a thorough ablation study was conducted to determine the best optimal model for human physical activity recognition.

1. Introduction

Human physical activity recognition (HPAR) has emerged as a crucial area of research with diverse applications in surveillance systems [1], health care systems [2,3], gyms [4], gaming [5], etc. Additional applications of PAR include assisting older people to live freely and safely in the community while utilizing the improved healthcare system. Moreover, vision-based human activity recognition for human–computer interaction (HCI) has attracted the interest of many researchers in recent times. For example, the invention of Kinect enables real-time HCI by recognizing user gestures and body motions to control a character or for gameplay features. However, despite advancements in vision-based HCI, the major obstacle is the automatic recognition of gesture, action, and behavioral context, which is still an open and challenging topic to tackle. Therefore, several researchers utilized Kinect for the evaluation of indoor HPAR for full body tracking and gesture-based detection by an intelligent HCI system [5,6]. Human activity recognition is an essential domain that involves analyzing the gestures and actions that humans perform throughout their daily lives. Researchers use many data modalities to understand and identify these activities, with each modality providing distinct insights into patterns of human motion [7,8,9,10]. RGB video data are one of the modalities that collect visual information by using the colors red, green, and blue and provide a comprehensive perspective of the activities performed by humans [11,12]. They offer a comprehensive depiction of applications’ objects and motions, allowing for meticulous examination using image processing and computer vision methodologies. Although RGB data provide extensive visual information, they often pose difficulties in processing caused by variables such as variations in lighting, obstructions, and distracting background elements, all of which may impact the accuracy of identification. On the other hand, skeletal data offer a different and often more sophisticated method for recognizing activities [13]. They simplify human postures into essential joint coordinates, providing an organized depiction of physical motions [14]. This data type offers a streamlined and concentrated perspective, highlighting the fundamental characteristics necessary for identifying actions. Skeleton data provide a stable and robust representation by recording body joint spatial connections and configurations throughout motions [15]. They are less influenced by external variables like lighting conditions or background clutter. By concentrating on the fundamental elements of joint positions and motions, this abstraction simplifies the recognition process, allowing for a more efficient and precise study of human activities. The advantage of skeletal data over RGB data in certain situations stems from their capacity to condense intricate motions into essential joint locations and trajectories, resulting in a more distinct and less distorted depiction of human activities. The simplified representation of skeletal data allows for more efficient analysis and identification, making them a preferred method in situations where exact activity detection is essential, such as in gesture recognition [16], sports analytics [17], or healthcare applications [18] that need precise movement monitoring. While RGB data provide a wider contextual perspective, skeletal data’s concentrated representation is frequently more appropriate and resilient for the recognition and categorization of human physical activity.
Recently, artificial intelligence-based methods, especially deep learning, have achieved outstanding breakthroughs in human activity recognition. Accurate and real-time recognition of human activities provides valuable insights into understanding human behavior, promoting fitness, and enhancing quality of life. Furthermore, advancements in sensor technology and the availability of large-scale annotated datasets have fueled the development of sophisticated activity recognition systems. Among the various modalities used for PAR, skeleton data have gained significant attention due to their effectiveness, low computational cost, and robustness to occlusions and noise. Skeleton data represent human body movements as a series of 2D or 3D joint positions over time, which can be captured using depth sensors, such as Microsoft Kinect, or through pose estimation techniques [19] from 2D images or videos. The inherent spatiotemporal representation of skeleton data enables a concise yet informative representation of human actions, making them an ideal input for activity recognition models.
Deep learning-based methods have demonstrated remarkable performance for HPAR. However, several challenges are associated with the current literature. The baseline methods utilize CNN or recurrent neural network (RNN) variants for HPAR. However, these methods can only extract spatial or temporal patterns from the data, while HPAR data have spatiotemporal dependencies. To overcome this challenge, the recent literature has developed hybrid models; however, these methods utilize stacked layer phenomena, which extract spatial patterns followed by temporal patterns or vice versa. Furthermore, the output of these methods is directly inputted into fully connected layers for classification. To overcome these challenges, we developed a PAR-Net that extracts both spatial and temporal features simultaneously. The PAR-Net integrates a CNN for spatial patterns and an ESN for temporal dependencies for extraction. The output of both streams is then concatenated, followed by a self-attention mechanism for optimal feature selection. Additionally, we investigate the network’s interpretability, providing insights into how it learns to recognize specific activities from the skeleton data. The main contributions of this research are as follows:
  • The PAR-Net introduces a novel dual-stream paradigm tailored for HPAR, which involves the integration of two distinct streams within the network architecture. The first stream utilized a CNN, which is adept at capturing spatial patterns inherent in the data. Simultaneously, the second stream incorporates an ESN, which is uniquely designed to capture temporal dependencies embedded within the activity sequences. By operating these streams in parallel, the PAR-Net inherently preserves both spatial and temporal characteristics crucial for accurate activity recognition, thereby overcoming the limitations posed by prior models that could only capture spatial or temporal aspects separately or in stacked layer phenomena.
  • The PAR-Net incorporates a crucial self-attention mechanism. This mechanism dynamically highlights relevant spatiotemporal features while suppressing noise and irrelevant information. By allowing the network to focus on critical aspects of the data selectively, the self-attention mechanism significantly enhances the discriminative capabilities of the PAR-Net. This integration empowers the network to learn more salient representations, thereby contributing to improved accuracy and robustness in physical activity recognition tasks.
  • This research extensively examines the PAR-Net and other architectures through thorough ablation studies. Each component, including the dual-stream design, self-attention mechanism, CNN–ESN architecture, and other architecture utilization, is systematically evaluated to understand their individual impact before the PAR-Net selection.
  • To gauge the efficacy and superiority of the PAR-Net, comprehensive comparisons are made against the baseline methods commonly used in physical activity recognition. The evaluation encompasses various metrics, including accuracy, robustness, and computational efficiency, comparing the PAR-Net performance against baselines.
The rest of the paper is structured as follows: Section 2 presents a review of related works in human physical activity recognition and discusses existing approaches using skeleton data. Section 3 details the architecture and components of the PAR-Net. Section 4 presents the experimental setup, dataset descriptions, evaluation metrics, and comparative analysis with state-of-the-art methods. Finally, Section 5 concludes the paper and outlines potential directions for future research in this domain.

2. Literature Review

Over the past eight years, significant advancements have been made in human physical activity recognition using 2D skeleton data based on advanced neural network architectures. For instance, Shi et al. [20] demonstrated the efficacy of CNNs for skeleton-based action recognition. The authors proposed a method to map 2D joint coordinates to heatmaps, effectively transforming the skeleton data into a format suitable for CNN input. Although lacking temporal modeling, this study laid the foundation for CNN-based approaches in the field. The authors [21] proposed a novel LSTM network called global context-aware attention LSTM (GCA-LSTM), which mainly focuses on useful joints using a global context memory cell. They also introduced a recurrent attention mechanism to increase the attention capability of the network and combined fine-grained and coarse-grained attention in a two-stream framework. This approach achieves state-of-the-art accuracy on five different challenging datasets utilized by researchers for skeleton-based action recognition. Liu et al. [22] proposed a skeleton visualization method for view-invariant human action recognition. The method involves a view-invariant transform based on sequences to eliminate the influence of view variations, visualizing transformed skeletons as color images, and applying visual and motion enhancement methods to enhance local patterns. A CNN-based model is then used to extract informative and discriminative features from the RGB images, and decision-level fusion is applied to generate action class scores.
Another research study [23] provided a deep learning-inspired ConvNets-based strategy for activity identification utilizing numerous visual signals. It also presented a novel technique for producing skeletal pictures that reflect motion data. RGB, skeletal data, and depth from RGB-D sensors are employed in the method, which is subsequently trained on ConvNets and merged at the outcome level. Ghazal et al. [24] focused on vision-based human activity recognition (HAR) using skeletal data obtained from standard cameras. The approach extracts motion features from the 2D positions of human skeleton joint data through the OpenPose library. The study used supervised machine learning with different classifiers (K-nearest neighbors (KNN), support vector machine (SVM), naive Bayes (NB), linear discriminant, and feed-forward back-propagation neural network) to recognize four activity classes: standing, walking, sitting, and falling. The experimental results showed that the K-nearest neighbors classifier performed the best among the others. This method offers device-independent HAR for various applications like video surveillance, telecare, ambient intelligence, and robot navigation. Experimental results for deep learning methods like [25], where a CNN and long short-term memory (LSTM) have been adopted to obtain spatial-temporal information, show that the score fusion of CNN and LSTM performs better than LSTM, achieving 87.40% accuracy and ranking 1st in the Large-Scale 3D Human Activity Recognition Challenge in Depth Videos. Li et al. [26] developed actional–structural graph convolutional networks (AS-GCNs) that integrate both actional and structural graph convolutions to capture the dynamics of actions and the spatial relationships between joints. The actional graph convolution captures the temporal dependencies between joints, while the structural graph convolution focuses on modeling the skeletal structure. The AS-GCN demonstrated superior performance on challenging action recognition tasks, emphasizing the importance of considering both spatial and temporal aspects. Another study proposed an intelligent HAR-based system [27] that combines image processing and deep learning techniques using human skeleton information and automatically recognizes human daily activities. This approach has proven promising because of its low computation cost and high accuracy outcomes, making it suitable for embedded systems.
However, effective model training is essential for PAR systems to accurately detect human body parts and the required physical activities. Therefore, Nadeem et al. [28] presented an integrated framework that deals with multidimensional features using a fusion of human body part models and quadratic discriminant analysis. Multilevel features are extracted as displacement parameters, representing the spatiotemporal properties of body parts over time. The recognition engine utilizes a maximum-entropy Markov model to process these characteristics. The experimental findings demonstrated that this method outperformed existing techniques in terms of accuracy for body part identification and physical activity recognition. The body part detection accuracy achieved a rate of 90.91% on the sports action dataset provided by the University of Central Florida. In contrast, the accuracy for activity recognition on UCF YouTube action dataset was found to be 89.09% and the accuracy on IM-DailyRGBEvents dataset was 88.26%, respectively. To capture spatial information and temporal variations for action classification, Tasnim et al. [29] offered a spatio-temporal image formation (STIF) approach for 3D skeletal joints. The transfer learning technique was used by utilizing the pre-trained weights of the models, including MobileNetV2, ResNet18, and DenseNet121, and their proposed method. The method outperformed prior works using UTD-MHAD and MSR-Action3D datasets, including STIF representation, attaining performance accuracy of approximately 98.93%, 98.80%, and 99.65% on UTD-MHAD and 96.00%, 97.08%, and 98.75% on MSR-Action3D datasets for each method, respectively. A dual-stream structured graph convolution network (DS-SGCN) was proposed in [30] to address the issue of skeleton-based action recognition. The DS-SGCN integrates the spatiotemporal joint coordinates and appearance contexts of skeletal joints, allowing for a better representation of discrete joints. The module encodes segregated body parts and dynamic interactions in a spatiotemporal sequence. Structured intra-part graphs represent distinct body parts, while inter-part graphs model dynamic interactions across different body parts. The DS-SGCN acquires intrinsic properties and dynamic interactions of human activity by learning on both inter- and intra-part graphs. After integrating spatial context and coordinate cues, a convolution-filtering process captures the temporal dynamics of human movement. The fusion of two streams of graph convolution responses predicts category information about human action end-to-end. The DS-SGCN technique achieved encouraging performance on five benchmark datasets.
Another method developed in [31] aimed to assess the efficacy of two RNN models, namely the 2BLSTM and 3BGRU, in accurately detecting daily postures and possibly dangerous situations in a home monitoring system. The analysis was conducted using 3D skeletal data obtained from a Kinect V2 device. The RNN models were evaluated using two distinct feature sets: one composed of eight kinematic characteristics that were manually picked using a genetic algorithm, and another composed of 52 ego-centric 3D coordinates for each skeletal joint, together with the subject’s distance from the Kinect V2. In order to enhance the capacity of the 3BGRU model to make generalizations, they used a data augmentation technique to ensure a balanced training dataset and attained the highest level of accuracy of 88%. Cheng et al. [32] proposed a system for recognizing physical exercise from video frames (extracted 3D human skeleton data using VIBE) using deep semantic features and repetitive segmentation algorithms. The system locates and segments the activity into multiple unit actions, improving recognition and time intervals. Experiments on the “NOL-18 Exercise” dataset showed an accuracy of 96.27% with a 0.23% time error. This system could be used in fitness or rehabilitation centers for patient treatment. Another study [33] presented a human tracking model for squat exercises using open-source MediaPipe technology. The model detects and tracks vital body joints, analyzing critical joint motions for abnormal movements. Validated using a squat dataset from ten healthy individuals, the model uses double exponential smoothing to classify movements as normal or abnormal. The model accurately predicted six out of ten subjects, with a mean square error of 56.31% for normal squat and 29.78% mean square error for abnormal squat setting, respectively. This low-cost camera-based squat movement condition detection model effectively detects workout movement abnormalities. Chariar et al. [34] introduced a method for classifying different types of squats and recommending the appropriate version for individuals. The study utilized MediaPipe and a deep learning-based approach to determine whether squatting is performed correctly or incorrectly. A stacked Bidirectional Gated Recurrent Unit (Bi-GRU) with an attention layer was used to consistently and fairly evaluate each user, categorizing squats into seven classes. The model was compared to other advanced models, both with and without the attention layer, and it outperformed the baselines. Recently, Li et al. [35] developed a unique approach to recognizing human skeletal movement in energy-efficient smart homes by fusing several CNNs. Gray value encoding is used in this technique to encode spatiotemporal information for each 3D skeletal sequence. The CNN fusion model uses three SPI and three STSI sequences as model input for skeletal activity recognition, allowing hierarchical learning of spatiotemporal features. Experimental results were carried out on three public datasets, showing that their performance is better than that of state-of-the-art methods. To address the limitations of traditional machine learning and deep learning models for activity recognition, the authors [36] developed a hybrid model that combines CNN and LSTM for activity recognition. They also generated a new dataset containing 12 different classes of human physical activities, collected from 20 participants using the Kinect V2 sensor. Through an extensive study, the CNN-LSTM technique achieved an accuracy of 90.89%, demonstrating its suitability for HAR applications. Furthermore, Luwe et al. [37] presented a hybrid deep learning model called 1D-CNN-BiLSTM for recognizing human activities using data collected from wearable sensors. Using bidirectional long short-term memory (BiLSTM), the model encodes long-range relationships while transforming sensor time series data into high-level representative characteristics.
Researchers have explored various approaches, including multi-modal fusion, attention mechanisms, visualization networks, and domain adaptation, to enhance the accuracy, interpretability, and generalization capabilities of the models. These advancements contribute to human activity recognition systems for diverse real-world applications using different types of input data, which collectively contribute to developing a robust and reliable system and also provide valuable insights into building more efficient recognition systems. In summary, the literature review highlights the evolution of methodologies in human physical activity recognition using skeleton data. However, several challenges are associated with the current literature. For instance, hybrid models in HPAR often employ sequential approaches, limiting their ability to extract both spatial and temporal features concurrently. This compromises accurate recognition as it fails to preserve the intricate relationship between spatial and temporal characteristics. Furthermore, existing methods lack efficient strategies for feature selection, potentially leading to suboptimal representations and impacting recognition accuracy. To overcome these challenges, the PAR-Net introduces a dual-stream paradigm, integrating a CNN for spatial patterns and an ESN for temporal dependencies. This simultaneous extraction of spatial and temporal features preserves the essential characteristics for accurate activity recognition. Additionally, the PAR-Net incorporates a self-attention mechanism, dynamically highlighting salient spatiotemporal features, thereby enhancing discriminative capabilities. This comprehensive approach ensures optimal feature representation and addresses the limitations of prior methods, leading to superior performance in accuracy, robustness, and computational efficiency for physical activity recognition.
The existing literature predominantly focuses on solo architectures for HPAR, overlooking the challenge posed by datasets containing both spatial and temporal features. Solo models struggle to effectively extract both types of features, hindering their performance. While some hybrid models have been proposed, such as combinations of CNN and RNN variants, they often utilize a stacked layer phenomenon for feature extraction. This approach, where one model learns from the original data while the subsequent model learns from the extracted features, faces difficulties in capturing effective temporal dependencies due to the reduction in data dimensionality after each convolutional layer. Moreover, the outputs of these models, whether solo or hybrid, are directly fed to dense layers without optimal feature selection. Consequently, the accuracy attained is insufficient for real-world implementation. In response to these limitations, we developed a PAR-Net, leveraging a dual-stream architecture to extract spatial and temporal features from raw data. These features are then concatenated and subjected to a self-attention mechanism for optimal feature selection, followed by classification using dense layers. Notably, the PAR-Net achieves higher accuracy compared to baseline models, addressing the shortcomings of existing approaches.

3. Proposed Methodology

A PAR-Net was developed to address the inherent challenges in recognizing complex activities from skeleton data. It incorporates various modules, including CNN, ESN, and self-attention, to extract spatial and temporal features effectively, where the integration of the attention module plays a crucial role in feature selection. All these modules are further discussed in the subsequent sections.

3.1. Spatial Feature Extraction

The 1D CNN is an effective architecture for extracting spatial patterns from data [38]. Unlike traditional approaches, which are reliant on handcrafted features, a 1D CNN automatically learns hierarchical representations of spatial features, making it capable of discerning complex patterns. Skeleton-based human physical activity recognition data provide a unique perspective on human movements through joints and key point configurations over time. A 1D CNN uses filters over the spatial dimension for feature extraction that gradually abstracts into high-level feature representation. A CNN layer is followed by a max-pooling layer to reduce the computational burden by focusing on the most salient information in feature maps. To finally classify the resulting features, the output is forwarded to fully connected layers. This architecture excels at learning both short-term and long-term dependencies within the data, making it well-suited for discerning nuanced patterns in activities. The processes of the l t h convolution operation, max pooling operation, and fully connected layer can be mathematically demonstrated below where Equation (1) can be written as:
Y i j l = σ m = 1 M W m , j l · X i = m 1 , j 0 + B j l ,
Here, B j l is the bias term of the j t h feature map, W represents the weights, m is the filter, and σ is the activation function (e.g., ReLU).
p i j l = m a x d i = 0 s Y i t + d i , j l 1
Equation (2) shows the mathematical representation of the max pooling operation, which downsamples the input by selecting the maximum value for the window. Where t is the stride and s is the size of the pooling operation.
d i l = j = 1 W j i l 1   σ X i l 1 + B j l 1 ,
where W j i l 1 is the weights, σ is the activation function, X i l 1 is the input data, and B j l 1 is the bias term in Equation (3).

3.2. Temporal Feature Learning

The ESN is the reservoir computing architecture that has become popular for processing sequential data, which makes it useful for time series analysis and advantageous in physical activity detection compared to other RNN architectures in terms of effectiveness and efficiency. ESNs can be distinguished from the other RNNs by a fixed, randomly generated reservoir of neurons that function as a dynamic memory. The ESNs have a fixed set of parameters and only optimize the output layer weights during training; they are better suited for applications with minimal labeled data because training is less complicated. The intrinsic memory capacity of ESNs is well suited for processing skeleton data. The reservoir of the network extracts temporal features from the sequential input by identifying patterns and relationships. Due to its dynamic properties, the reservoir can accurately simulate the temporal information of human actions and can accurately identify complex activities. To employ ESNs for activity recognition with skeleton data, the input sequences, which describe the spatiotemporal dynamics of human motions, are fed into the reservoir. The input data are converted into a high-dimensional state space by the reservoir. A linear readout layer that maps the high-dimensional representations to the target activity classes is subsequently trained using the reservoir states. Particularly, the only portion of the network that is trained is this readout layer, which streamlines the optimization procedure. ESNs have the benefit of being robust against noise, capturing long-term dependencies, and providing effective training with a small amount of labeled data. Due to the intrinsic memory provided by the fixed reservoir structure, the network can recognize and make use of temporal patterns in the skeletal data. Moreover, ESN training makes it easy, especially in situations where gathering sizable, labeled datasets might be difficult.
The ESN is mainly composed of three layers (input layer, reservoir layer, and output layer), as shown in Figure 1. The input layer and reservoir layer are generated randomly. The reservoir layer consists of N neurons and is considered fading memory, where x ( t ) is the input and h ( t ) combines it at time t . The state of the ESN is updated by Equation (4), as given in [39].
h t = ( W i n x t + W h   t 1 + W b a c k y t 1 + b x ) ,
The activation function utilized in reservoir neurons is represented by ( ) , the weight matrices of the reservoir layer are represented by W i n R M × N , the weight matrices of the input and output feedback are represented by W b a c k R N × Y , and the input bias is denoted by b x R N × 1 . The input y ( t 1 ) at the current instant is desired. Hence, W b a c k y ( t 1 ) can be combined with the W i n u ( t ) item. Equation (4) can be written as shown in Equation (5), where b x can be fused into W i n x ( t ) while only connecting the output rather than connecting the input and other neurons in the reservoir layer.
h t = ( W i n [ 1 : x ( t ) ] + W h   t 1 ) ,
Equation (6) is used to determine the target Y t at time t :
Y t = W o u t h t + b y ,
where the output weights of the ESN are denoted with W o u t R ( N + 1 ) × Y , the output bias is denoted by b y R N × 1 , and b y is fused with W o u t h t according to Equation (6), which can be written as shown in Equation (7):
Y t = W o u t [ 1 ; h t ] ,
ESN implementation has two steps for target prediction, where the first step is specifying W i n , W , and states of ESN. The input weights W i n and W of the reservoir layer are generated randomly and kept constant. The output weights W o u t are optimized, all states are stored in matrix H , and the targeted outputs are stored in a feature vector y * . Finally, the model is connected with a linear regressor where the ridge regressor is used, which computes the regularized least-square problem using Equation (8), where β R + represents the L2 regularization coefficient.
W o u t * = a r g m i n   W R 1 2   | | H W y * | | 2 + β 2 | | W | | 2 = ( H T H + β I ) 1 H T y *

3.3. Self-Attention Mechanism

The self-attention mechanism is a key tool for improving model performance, particularly in hybrid models. Both stream outputs are combined in order to generate a single feature vector, which is then inputted into SAM to provide a representative pattern for the final prediction. Figure 2 illustrates a visual representation of each stream output, concatenation layer, SAM, and fully connected layer. We integrate the SAM to select optimal features from the data by dynamically adjusting the importance of hidden patterns. SAM is used to identify the key patterns in the merged feature vector of the CNN and ESN streams prior to the prediction process. It also analyzes the correlation of hidden characteristics across different timestamps in each dimension. In SAM, the score of hidden features of the g t h time stamp in the p t h dimension can be calculated using Equation (9):
S g ,   p = f s c o W g , p h 1 , p , h 3 , p , h 3 , p , h n , p , ,   p = 1,2 , 3 n ,   g = 1,2 , 3 t
Here, the p t h dimension of the hidden state on the g t h timestamp is represented by h g , p , the weight metric is indicated by W g , p , a function implemented by fully connected layers is represented by f s c o , the number of timestamps is represented by t , and the hidden feature dimension is represented by n . The final layers of our proposed model are fully connected layers, which are used for activity recognition. The SAM output is flattened to a Z l = Z 2 , Z 2 , Z 3 Z n feature vector, where the number of SAM output dimensions is represented by l . The output of the attention module is fed to a dense layer.

3.4. PAR-Net

We proposed a dual-stream model with an attention mechanism called PAR-Net to efficiently recognize physical activities using skeleton data. In the spatial stream, CNN layers are employed to extract spatial features from the input skeleton data. These convolutional layers apply filters to the data, capturing hierarchical spatial patterns. In the temporal stream, the ESN is integrated to model the temporal dynamics of the skeletal sequence. In the CNN stream, the data are processed by two 1D-convalutional layers with 64 and 128 filters and a kernel size of 3 and 1, respectively. This process is followed by a dense layer with 128 units and a flattened layer. Meanwhile, the ESN stream utilizes an Echo State RNN cell consisting of 32 units, which has been configured with a hyperbolic tangent (tanh) activation function. The cell is started with predetermined parameters, including decay, epsilon, and alpha, and is configured to update certain variables throughout the training process. The output of both streams is then combined by a concatenation method and forwarded to a self-attention mechanism. A self-attention mechanism is used to select more prominent features from the combined feature vector. These optimal features are then fed to a dense layer consisting of 64 units using a rectified linear unit (ReLU) activation function. Finally, a SoftMax layer consisting of 5 units (depending on the number of classes) is applied to calculate the model output probabilities for the various activity classes. The model is trained using the Adam optimizer, with a learning rate of 0.0001, and the categorical cross-entropy loss function. During the training process, the model is adjusted to the training data using a batch size of 32 for 100 epochs.

4. Experimental Results

This section delves into a detailed exploration of the evaluation metrics, datasets, and performance comparisons of different models. Specifically, we focus on assessing the efficacy of the PAR-Net in comparison to baseline models. We have used two datasets for the evaluations of our proposed model and the dataset details are given in Table 1. Several evaluation metrics are used to assess the performance of models that aim to recognize or classify human activities based on skeletal joint data captured from devices. We used accuracy, precision, recall, and F1-score for performance and comparative analysis; the details of these metrics are given in [36]. The experiments are conducted on a computational setup composed of a GeForce RTX 3070 GPU with 8 GB of RAM, a Core i7 processor with 32 GB of onboard memory, and the Windows 10 operating system (Nvidia Corporation, Santa Clara, CA, USA). The implementation was carried out using Python V3.7.4, employing the Keras deep learning framework built on TensorFlow.

4.1. Dataset Description

In this study, we used two datasets for the PAR-Net evaluation. The first dataset, the Physical Activity Recognition Dataset (PAR) [36], is composed of 12 distinct activities performed by 20 different individuals aged 25–35 years. Data collection involved the use of Microsoft’s Kinect sensor V2, which is capable of extracting 25 different joints from the human body (refer to Figure 3). The dataset contains the x- and y-axis values of all body joints and stores them as CSV files. Each participant performed an activity for 10 s, resulting in 200 samples for each activity, with 10 repetitions per participant (totaling 120 samples per participant). The Kinect V2 sensor was employed to extract human skeleton joints using the Discrete Gestures Basics WPF SDK. Joint data were captured by the Kinect Body View script and saved as CSV files.
The data are organized in the sequence shown in Figure 3, representing 25 body joints, labeled numerically as indicated in Table 1. The labeled activity data files are consolidated and identified by their respective class numbers. All activity files are then merged into a single training file and transferred to the model using 5 different frame sequences, such as 30, 60, 90, 120, and 150 frame sequence data.
The second dataset used in this study is called the Physical Exercise Recognition Dataset (PER) [40], which is composed of 10 different poses that can be used to distinguish 5 exercise activities. The exercises are jumping jacks, push-ups, sit-ups, pull-ups, and squats. The dataset is composed of 33 different landmarks representing the positions of human body parts. A sequence of poses is provided to preserve the order of frames in each record. The data is collected from 447 videos in which different people are performing exercises, and these videos are collected from the Countix dataset, which provides the YouTube links to several activity videos. The videos are downloaded and preprocessed, where the frames are extracted from the videos and stored. They used the MediaPipe framework to extract the human skeletal data from the video frames, which can efficiently predict the location of 33 landmarks on the human body and face and store it as CSV files. The labels are provided with data in separate CSV files, which contain details about the data and their specific labels. The detailed explanation of both datasets used in this research is shown in Table 1.

4.2. Comparative Analysis of Different Models

The performance of different deep learning models is evaluated for activity recognition on different frame sequences, such as 30, 60, 90, 120, and 150 frames. The data of each frame are stored row-wise in a CSV file, where the information for each frame includes the coordinates of skeletal joints and corresponding labels indicating the specific activity. Therefore, each row of data represents one frame of the activity sequence, with the activity label in the last column. The term "frames" refers to the number of consecutive frames of skeleton data passed to the model as input. For example, when we mention using 30 frames of data, it means that we fed sequences of 30 consecutive frames of data to the model. The accuracy of each model is emphasized during these intervals, offering information about how effective it is in terms of accuracy, as reported in Table 2. The ablation results show a consistent tendency among the models evaluated, where most models show a reasonable level of accuracy at shorter intervals but are unable to sustain it as the frame intervals increase. Certain models, such as the MLP, CNN, LSTM, Bidirectional LSTM, ESN, GRU, BiGRU, CNN-GRU, CNN-LSTM, CNN–ESN, and Dual-Stream CNN–ESN without an attention mechanism, show good accuracy at certain intervals on both datasets but become less accurate at longer frames, suggesting that they are unable to capture long-term temporal dependencies. Nonetheless, the proposed dual-stream model, coupled with the attention mechanism, consistently obtains the best results comparatively. Figure 4 shows the accuracy of our proposed model compared to other models.
Table 3 shows the precision scores generated by various deep-learning models used to recognize activities at different intervals of time: 30, 60, 90, 120, and 150 frames. These models include the MLP, CNN, LSTM, Bidirectional LSTM, ESN, GRU, BiGRU, CNN-GRU, CNN-LSTM, CNN–ESN, and Dual-Stream CNN–ESN without an attention mechanism. However, when the frame intervals increase, they show a reduction in accuracy, suggesting that greater periods cannot be reliably predicted. On the other hand, the PAR-Net consistently performs well with higher precision scores through all frame intervals and also maintains an impressive level of precision, peaking at 94.41% at 120 frames on the PAR dataset, while the peak precision score is 94.88% on the physical exercise recognition dataset. Such high performance highlights the PAR-Net's ability to reliably anticipate a wide range of actions, demonstrating its advantage over other models during evaluation across varying time intervals, as shown in Figure 5. The recall and F1-score metrics are also used to evaluate the performance of different models with diverse frame sequences, as shown in Table 4 and Table 5 respectively. The performance of these models decreases when the sequence increases. However, the PAR-Net performs well at longer intervals, peaking at 120 sequences with a recall value of 94.02% and a 94% F1-score on the physical activity recognition dataset, while the recall score and F1-score on the physical exercise dataset are 94.02% and 93.5%, respectively, as given in Figure 6 and Figure 7.
Overall, the PAR-Net performed better in all evaluated measures (accuracy, precision, recall, and F1-score), demonstrating its ability to capture complex temporal correlations and make accurate predictions over extended periods. Although some models demonstrate proficiency within particular temporal domains, they are unable to preserve stability and uniformity, particularly as the input sequences of the model increase. Therefore, the empirical results emphasize the PAR-Net as an optimal solution among the assessed models, supported by its dual-stream design and integrated attention mechanism. The ability of the PAR-Net to interpret activities over a range of time intervals highlights an optimal solution for activity recognition tasks, providing a thorough understanding of temporal and spatial dynamics in input data.
In order to provide a thorough study of the model’s performance, we have included confusion matrices for both datasets at different frame sequences (30, 60, 90, 120, and 150). The confusion matrices, which provide the counts of true positive, true negative, false positive, and false negative predictions for each activity class, provide insightful information about the classification results. Figure 8 shows the confusion metrics of different frame sequences on the PAR dataset, where the activities in the confusion metrics are represented by the number sequences mentioned in Table 1. Figure 9 represents the confusion metrics of the PER dataset on various frame sequences.

4.3. Stability Analysis Test

Stability analysis involves examining a model’s behavior and performance throughout many training iterations and epochs to determine its consistency and dependability. This research evaluates many aspects of the model’s learning dynamics, with a specific emphasis on its capacity to reach an optimum solution without experiencing overfitting or underfitting. Stability analysis involves examining the consistency and convergence of accuracy trends over consecutive epochs while analyzing the training and validation accuracy graphs. Both training and validation accuracy should ideally show consistent improvement and stability, indicating successful learning without affecting generalization. When analyzing the training and validation loss graphs, stability analysis entails observing the decreasing trend of loss values over epochs to verify that the model is efficiently reducing errors. An important part of stability analysis is recognizing the convergence of validation loss, which signals that the model has achieved its optimum performance on the validation set. Figure 10 and Figure 11 show the training and validation accuracy and loss on the PAR and PER datasets, respectively.

4.4. Statistical Significance Tests

Statistical significance tests, or t-tests, are used in our research to evaluate and guarantee the suggested model’s capacity for generalization. It is a systematic process used in statistical analysis to determine whether the results of the experiment provide strong enough evidence to reject a null hypothesis against an alternative hypothesis. This test aims to determine whether any differences or impacts in the data are statistically significant or just due to random variation. The procedure usually starts by creating two opposing hypotheses: the null hypothesis (H0), stating no impact or difference, and the alternative hypothesis (H1), proposing the existence of an effect. We chose a significance level (α) of 0.05 to determine the threshold of significance for approving or disapproving the null hypothesis. The significance test uses an appropriate test statistic, t-statistics, to produce either the p-value or critical value. The p-value indicates the likelihood of observing the data or more extreme data if the null hypothesis is true, but the critical value indicates the threshold at which the null hypothesis is rejected. The null hypothesis is determined by comparing the estimated p-value to the specified significance level or determining whether the test statistic is inside the rejection zone. When the p-value is less than α or when the test statistic exceeds the critical value, the null hypothesis is rejected, suggesting a statistically significant difference. If the p-value is greater than α or if the test statistic is within the non-rejection zone, the null hypothesis is not rejected since there is not enough evidence to support its rejection. The model is considered accurate if the t-statistic value is low, and the p-value is acquired through the comparison of the PAR-Net with another baseline model. The significance test provides a structured method to make meaningful inferences from data, facilitating informed decision-making and inference in scientific studies. Table 6 shows the significant test results comparison of the PAR-Net with other baseline models.
The PAR-Net holds significant potential for real-world applications, particularly in healthcare, sports training, and smart home systems. In healthcare, it enables remote monitoring and customized therapies by tracking patients’ physical activity levels, which is especially beneficial for long-term health issues and rehabilitation. Athletes can benefit from improved training and performance monitoring through insights into movement patterns and workload management, reducing the risk of injuries. Integrating the model into smart home devices enhances comfort and energy efficiency by adjusting environmental parameters based on inhabitants’ activities and preferences. However, challenges arise from reliance on precise data collection devices and fluctuations in data integrity and external circumstances, necessitating improvements in data collection methods, noise handling, and validation through longitudinal studies. Investigating hybrid methodologies combining skeletal data with other modalities may enhance adaptability in various scenarios.

5. Conclusions and Future Work

This paper proposes a dual-stream architecture followed by an attention mechanism that can effectively recognize human physical activity. Using skeletal data and sophisticated models has been a key strategy in physical activity identification. By putting forth a novel solution, the PAR-Net performance is evaluated on two datasets aimed at transforming this field and efficiently recognizing human physical activity. The main goal is to recognize complex activity patterns across varied time intervals, broken down into frame sequences of 30, 60, 90, 120, and 150 sequences in both datasets, and compare the results of different methods. After a thorough analysis, the proposed model performs better than existing models. The PAR-Net demonstrated its superiority in properly identifying physical activities by outperforming other models over a wide range of temporal sequences.

Limitations and Future Direction

The PAR-Net achieved higher performance; however, several limitations are associated with its current implementation and areas for future exploration. The PAR-Net's reliance solely on skeleton data disregards the potential information collected from other modalities like RGB images or depth maps. Incorporating multimodal data sources could enrich feature representation and boost the model's robustness. Furthermore, the model focuses on single-person skeletal data, which restricts its applicability in scenarios involving multiple individuals. Future research should prioritize developing models capable of concurrently processing multi-person skeletal data to effectively address real-world interactions. The PAR-Net uses multiple architectures to extract spatial and temporal features, which makes the model computationally expensive. Future research should explore the development of a unified architecture capable of extracting both spatial and temporal features efficiently for human physical activity recognition. This could involve the creation of a single architecture tailored to handle both types of features, potentially reducing computational complexity while maintaining performance. Moreover, real-time deployment considerations require optimization of the model architecture and inference techniques for efficient real-time performance. Furthermore, the absence of a comprehensive graphical user interface (GUI) hinders the model's accessibility and usability, highlighting the need for an intuitive interface design. By addressing these limitations and exploring potential avenues for enhancement, we aim to advance the efficacy and applicability of human physical activity recognition systems.

Author Contributions

Conceptualization, I.U.K.; Methodology, I.U.K.; Software, J.W.L.; Validation, I.U.K. and J.W.L.; Formal Analysis, I.U.K. and J.W.L.; Investigation, J.W.L.; Resources, J.W.L.; Data Curation, I.U.K. and J.W.L.; Writing—Original Draft Preparation, I.U.K.; Writing—Review & Editing, I.U.K. and J.W.L.; Visualization, I.U.K.; Supervision, J.W.L.; Project Administration, J.W.L.; Funding Acquisition, J.W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Ministry of Science and ICT (MSIT), Korea, under the Information Technology Research Center support program (IITP-2023-RS-2023-00156354) supervised by the Institute for Information and Communications Technology Planning and Evaluation (IITP). It was also supported by the IITP grant under the metaverse support program to nurture the best talents (IITP-2023-RS-2023-00254529), funded by the Korean government (MSIT).

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

Data are contained in this article.

Acknowledgments

This research was supported by the Ministry of Science and ICT (MSIT), Korea, under the Information Technology Research Center support program (IITP-2023-RS-2023-00156354) supervised by the Institute for Information and Communications Technology Planning and Evaluation (IITP). It was also supported by the IITP grant under the metaverse support program to nurture the best talents (IITP-2023-RS-2023-00254529), funded by the Korean government (MSIT).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Taha, A.; Zayed, H.H.; Khalifa, M.; El-Horbaty, E.-S.M. Human activity recognition for surveillance applications. In Proceedings of the 7th International Conference on Information Technology, Amman, Jordan, 12–15 May 2015. [Google Scholar]
  2. Mousse, M.A.; Motamed, C.; Ezin, E.C. Percentage of human-occupied areas for fall detection from two views. Vis. Comput. 2017, 33, 1529–1540. [Google Scholar] [CrossRef]
  3. Wang, Y.; Cang, S.; Yu, H. A survey on wearable sensor modality centred human activity recognition in health care. Expert Syst. Appl. 2019, 137, 167–190. [Google Scholar] [CrossRef]
  4. Lin, B.-S.; Wang, L.-Y.; Hwang, Y.-T.; Chiang, P.-Y.; Chou, W.-J. Depth-camera-based system for estimating energy expenditure of physical activities in gyms. IEEE J. Biomed. Health Inform. 2018, 23, 1086–1095. [Google Scholar] [CrossRef]
  5. Obdržálek, Š.; Kurillo, G.; Ofli, F.; Bajcsy, R.; Seto, E.; Jimison, H.; Pavel, M. Accuracy and robustness of Kinect pose estimation in the context of coaching of elderly population. In Proceedings of the 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, San Diego, CA, USA, 28 August–1 September 2012. [Google Scholar]
  6. Aoki, H.; Miyazaki, M.; Nakamura, H.; Furukawa, R.; Sagawa, R.; Kawasaki, H. Non-contact respiration measurement using structured light 3-D sensor. In Proceedings of the 2012 Proceedings of SICE Annual Conference (SICE), Akita, Japan, 20–23 August 2012. [Google Scholar]
  7. Batchuluun, G.; Kang, J.K.; Nguyen, D.T.; Pham, T.D.; Arsalan, M.; Park, K.R. Action recognition from thermal videos using joint and skeleton information. IEEE Access 2021, 9, 11716–11733. [Google Scholar] [CrossRef]
  8. Bulbul, M.F.; Tabussum, S.; Ali, H.; Zheng, W.; Lee, M.Y.; Ullah, A. Exploring 3D human action recognition using STACOG on multi-view depth motion maps sequences. Sensors 2021, 21, 3642. [Google Scholar] [CrossRef]
  9. Sun, Z.; Ke, Q.; Rahmani, H.; Bennamoun, M.; Wang, G.; Liu, J. Human action recognition from various data modalities: A review. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 3200–3225. [Google Scholar] [CrossRef]
  10. Yadav, S.K.; Tiwari, K.; Pandey, H.M.; Akbar, S.A. A review of multimodal human activity recognition with special emphasis on classification, applications, challenges and future directions. Knowl.-Based Syst. 2021, 223, 106970. [Google Scholar] [CrossRef]
  11. Ullah, A.; Muhammad, K.; Ding, W.; Palade, V.; Haq, I.U.; Baik, S.W. Efficient activity recognition using lightweight CNN and DS-GRU network for surveillance applications. Appl. Soft Comput. 2021, 103, 107102. [Google Scholar] [CrossRef]
  12. Haroon, U.; Ullah, A.; Hussain, T.; Ullah, W.; Sajjad, M.; Muhammad, K.; Lee, M.Y.; Baik, S.W. A multi-stream sequence learning framework for human interaction recognition. IEEE Trans. Human-Mach. Syst. 2022, 52, 435–444. [Google Scholar] [CrossRef]
  13. Pham, Q.-T.; Nguyen, D.-A.; Nguyen, T.-T.; Nguyen, T.N.; Nguyen, D.-T.; Pham, D.-T.; Tran, T.H.; Le, T.-L.; Vu, H. A study on skeleton-based action recognition and its application to physical exercise recognition. In Proceedings of the 11th International Symposium on Information and Communication Technology, Hanoi, Vietnam, 1–3 December 2022. [Google Scholar]
  14. Sardari, S.; Sharifzadeh, S.; Daneshkhah, A.; Nakisa, B.; Loke, S.W.; Palade, V.; Duncan, M.J. Artificial Intelligence for skeleton-based physical rehabilitation action evaluation: A systematic review. Comput. Biol. Med. 2023, 158, 106835. [Google Scholar] [CrossRef] [PubMed]
  15. Wu, L.; Zhang, C.; Zou, Y. SpatioTemporal focus for skeleton-based action recognition. Pattern Recognit. 2023, 136, 109231. [Google Scholar] [CrossRef]
  16. Nunez, J.C.; Cabido, R.; Pantrigo, J.J.; Montemayor, A.S.; Velez, J.F. Convolutional neural networks and long short-term memory for skeleton-based human activity and hand gesture recognition. Pattern Recognit. 2018, 76, 80–94. [Google Scholar] [CrossRef]
  17. Elaoud, A.; Barhoumi, W.; Zagrouba, E.; Agrebi, B. Skeleton-based comparison of throwing motion for handball players. J. Ambient Intell. Humaniz. Comput. 2020, 11, 419–431. [Google Scholar] [CrossRef]
  18. Lincke, A.; Fagerström, C.; Ekstedt, M.; Löwe, W.; Backåberg, S. A comparative study of the 2D-and 3D-based skeleton avatar technology for assessing physical activity and functioning among healthy older adults. Health Inform. J. 2023, 29, 14604582231214589. [Google Scholar] [CrossRef]
  19. Martınez, G.H. Openpose: Whole-Body Pose Estimation. Ph.D. Thesis, Carnegie Mellon University, Pittsburgh, PA, USA, 2019. [Google Scholar]
  20. Du, Y.; Fu, Y.; Wang, L. Skeleton based action recognition with convolutional neural network. In Proceedings of the 2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR), Kuala Lumpur, Malaysia, 3–6 November 2015. [Google Scholar]
  21. Liu, J.; Wang, G.; Duan, L.-Y.; Abdiyeva, K.; Kot, A.C. Skeleton-based human action recognition with global context-aware attention LSTM networks. IEEE Trans. Image Process. 2017, 27, 1586–1599. [Google Scholar] [CrossRef]
  22. Liu, M.; Liu, H.; Chen, C. Enhanced skeleton visualization for view invariant human action recognition. Pattern Recognit. 2017, 68, 346–362. [Google Scholar] [CrossRef]
  23. Khaire, P.; Kumar, P.; Imran, J. Combining CNN streams of RGB-D and skeletal data for human activity recognition. Pattern Recognit. Lett. 2018, 115, 107–116. [Google Scholar] [CrossRef]
  24. Ghazal, S.; Khan, U.S.; Mubasher Saleem, M.; Rashid, N.; Iqbal, J. Human activity recognition using 2D skeleton data and supervised machine learning. IET Image Process. 2019, 13, 2572–2578. [Google Scholar] [CrossRef]
  25. Li, C.; Wang, P.; Wang, S.; Hou, Y.; Li, W. Skeleton-based action recognition using LSTM and CNN. In Proceedings of the 2017 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), Hong Kong, China, 10–14 July 2017. [Google Scholar]
  26. Li, M.; Chen, S.; Chen, X.; Zhang, Y.; Wang, Y.; Tian, Q. Actional-structural graph convolutional networks for skeleton-based action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  27. Phyo, C.N.; Zin, T.T.; Tin, P. Deep learning for recognizing human activities using motions of skeletal joints. IEEE Trans. Consum. Electron. 2019, 65, 243–252. [Google Scholar] [CrossRef]
  28. Nadeem, A.; Jalal, A.; Kim, K. Accurate physical activity recognition using multidimensional features and Markov model for smart health fitness. Symmetry 2020, 12, 1766. [Google Scholar] [CrossRef]
  29. Tasnim, N.; Islam, M.K.; Baek, J.-H. Deep learning based human activity recognition using spatio-temporal image formation of skeleton joints. Appl. Sci. 2021, 11, 2675. [Google Scholar] [CrossRef]
  30. Xu, C.; Liu, R.; Zhang, T.; Cui, Z.; Yang, J.; Hu, C. Dual-stream structured graph convolution network for skeleton-based action recognition. ACM Trans. Multimedia Comput. Commun. Appl. 2021, 17, 1–22. [Google Scholar] [CrossRef]
  31. Guerra, B.M.V.; Ramat, S.; Beltrami, G.; Schmid, M. Recurrent Network Solutions for Human Posture Recognition Based on Kinect Skeletal Data. Sensors 2023, 23, 5260. [Google Scholar] [CrossRef] [PubMed]
  32. Cheng, S.-H.; Sarwar, M.A.; Daraghmi, Y.-A.; İk, T.-U.; Li, Y.-L. Periodic physical activity information segmentation, counting and recognition from video. IEEE Access 2023, 11, 23019–23031. [Google Scholar] [CrossRef]
  33. Muhammad, N.H.; Hassan, M.F.A.; Ibrahim, N.; Zin, Z.M. Mono Camera-based Human Skeletal Tracking for Squat Exercise Abnormality Detection using Double Exponential Smoothing. Int. J. Adv. Comput. Sci. Appl. 2022, 13, 65–70. [Google Scholar]
  34. Chariar, M.; Rao, S.; Irani, A.; Suresh, S.; Asha, C.S. AI Trainer: Autoencoder Based Approach for Squat Analysis and Correction. IEEE Access 2023, 11, 107135–107149. [Google Scholar] [CrossRef]
  35. Li, M.; Sun, Q. 3D skeletal human action recognition using a CNN fusion model. Math. Probl. Eng. 2021, 2021, 6650632. [Google Scholar] [CrossRef]
  36. Khan, I.U.; Afzal, S.; Lee, J.W. Human activity recognition via hybrid deep learning based model. Sensors 2022, 22, 323. [Google Scholar] [CrossRef]
  37. Luwe, Y.J.; Lee, C.P.; Lim, K.M. Wearable sensor-based human activity recognition with hybrid deep learning model. Informatics 2022, 9, 56. [Google Scholar] [CrossRef]
  38. Kiranyaz, S.; Avci, O.; Abdeljaber, O.; Ince, T.; Gabbouj, M.; Inman, D.J. 1D convolutional neural networks and applications: A survey. Mech. Syst. Signal Process. 2021, 151, 107398. [Google Scholar] [CrossRef]
  39. Liu, Y.; Shan, L.; Yu, D.; Zeng, L.; Yang, M. An echo state network with attention mechanism for production prediction in reservoirs. J. Pet. Sci. Eng. 2022, 209, 109920. [Google Scholar] [CrossRef]
  40. Tuameh, M. Physical Exercise Recognition dataset. Available online: https://www.kaggle.com/datasets/muhannadtuameh/exercise-recognition-time-series (accessed on 5 October 2023).
Figure 1. The internal architecture of the ESN.
Figure 1. The internal architecture of the ESN.
Sensors 24 01908 g001
Figure 2. Main framework of the PAR-Net.
Figure 2. Main framework of the PAR-Net.
Sensors 24 01908 g002
Figure 3. The human body skeleton joint extracted through Kinect.
Figure 3. The human body skeleton joint extracted through Kinect.
Sensors 24 01908 g003
Figure 4. The accuracy of different models on the PAR and PER datasets.
Figure 4. The accuracy of different models on the PAR and PER datasets.
Sensors 24 01908 g004
Figure 5. The precision of different models on the PAR and PER datasets.
Figure 5. The precision of different models on the PAR and PER datasets.
Sensors 24 01908 g005
Figure 6. Recall comparisons of different models on the PAR and PER datasets.
Figure 6. Recall comparisons of different models on the PAR and PER datasets.
Sensors 24 01908 g006
Figure 7. F1-score comparison of different models on the PAR and PER datasets.
Figure 7. F1-score comparison of different models on the PAR and PER datasets.
Sensors 24 01908 g007
Figure 8. The confusion metrics of the PAR-Net on various sequences on the PAR dataset.
Figure 8. The confusion metrics of the PAR-Net on various sequences on the PAR dataset.
Sensors 24 01908 g008
Figure 9. The confusion metrics of the PAR-Net on different sequences on the PER dataset.
Figure 9. The confusion metrics of the PAR-Net on different sequences on the PER dataset.
Sensors 24 01908 g009
Figure 10. The training and validation accuracy and loss graphs on the PAR dataset.
Figure 10. The training and validation accuracy and loss graphs on the PAR dataset.
Sensors 24 01908 g010
Figure 11. The training and validation accuracy and loss graphs on the PER dataset.
Figure 11. The training and validation accuracy and loss graphs on the PER dataset.
Sensors 24 01908 g011
Table 1. Types of activities and skeleton points in each dataset.
Table 1. Types of activities and skeleton points in each dataset.
DatasetsSkeleton Joints DataActivities
PAR Dataset1. Head
2. Neck
3. Spine Shoulder
4. Spin Mid
5. Spine Base
6. Shoulder Right
7. Shoulder Left
8. Hip Right
9. Hip Left
10. Elbow Right
11. Wrist Right
12. Hand Right
13. Hand Tip Right
14. Thumb Right
15. Elbow Left
16. Wrist Left
17. Hand Left
18. Hand Tip Left
19. Thumb Left
20. Knee Right
21. Ankle Right
22. Foot Right
23. Knee Left
24. Ankle Left
25. Foot Left
1. Overhead Arm Raise
2. Front Arm Raise
3. Arm Curl
4. Chair Stand
5. Balance Walk
6. Side Leg Raise (Right, Left)
7. Shoulder
8. Chest
9. Leg Raise (Forward, Backward)
10. Arm Circle
11. Side Twist (Right, Left)
12. Squats
PER Dataset0. nose
1. left_eye_inner
2. left_eye
3. left_eye_outer
4. right_eye_inner
5. right_eye
6. right_eye_outer
7. left_ear
8. right_ear
9. mouth_left
10. mouth_right
11. left_shoulder
12. right_shoulder
13. left_elbow
14. right_elbow
15. left wrist
16. right_wrist
17. left_pinky
18. right_pinky
19. left_index
20. right_index
21. left_thumb
22. right_thumb
23. left_hip
24.right_hip
25. left_knee
26. right_knee
27. left_ankle
28. right_ankle
29. left_heel
30. right_heel
31. left_foot_index
32. right_foot_index
1. Jumping Jack
2. Push-ups
3. Pull-ups
4. Sit-ups
5. Squats
Table 2. Accuracy comparison of models across different frames.
Table 2. Accuracy comparison of models across different frames.
PAR Dataset
No.Model Name306090120150
1MLP85.4583.6483.4787.0571.51
2CNN88.8888.2287.6583.7475.47
3LSTM83.3180.6474.6982.9266.09
4BiLSTM90.1585.3989.3082.0266.26
5ESN85.5590.7382.1792.0181.66
6GRU87.9686.3283.5787.7480.63
7BiGRU86.5792.8385.8983.7473.49
8CNN-GRU89.2588.6388.4887.6082.96
9CNN-LSTM [36]90.8988.9890.4487.9476.50
10CNN–ESN83.7687.0178.5691.6675.81
11Dual-Stream CNN–ESN89.1090.1587.0390.2882.78
12PAR-Net93.6992.3392.8293.9387.09
PER Dataset
1MLP88.2187.5084.8285.7188.39
2CNN87.1492.1487.5086.4290.17
3LSTM85.5387.5082.1481.4275.89
4BiLSTM91.4288.5777.6783.5784.82
5ESN86.6787.2584.8283.5281.77
6GRU88.5685.9082.7083.7579.35
7BiGRU87.6689.2086.5083.7881.89
8CNN-GRU90.5788.5587.9990.3382.66
9CNN-LSTM89.6490.0085.0082.1491.96
10CNN–ESN90.1787.6581.5389.5478.82
11Dual-Stream CNN–ESN90.5089.5488.7887.4783.78
12PAR-Net92.8589.6492.5188.5793.75
Table 3. Precision comparison of models across different frames.
Table 3. Precision comparison of models across different frames.
PAR Dataset
No.Model Name306090120150
1MLP86.1884.3785.1288.5474.97
2CNN89.2088.4888.3783.9378.04
3LSTM83.9482.5174.9584.0464.01
4BiLSTM90.7485.9089.6282.5270.35
5ESN85.8190.7782.3792.0381.10
6GRU87.9186.9584.0287.8981.51
7BiGRU86.5892.9585.8083.3073.48
8CNN GRU89.4989.2888.7988.6285.87
9CNN-LSTM [36]91.1189.3191.1388.8276.13
10CNN–ESN83.9287.2378.5691.7876.85
11Dual-Stream CNN–ESN89.2290.2087.6290.5284.02
12PAR-Net93.8492.5393.4894.4187.39
PER Dataset
1MLP88.0886.6583.6187.2987.19
2CNN86.9091.5386.5587.6189.37
3LSTM85.9986.5980.4881.0875.53
4BiLSTM91.2787.8576.8583.2283.68
5ESN86.0786.0783.3385.1181.55
6GRU88.1085.6082.2082.5580.10
7BiGRU86.5689.5586.2182.9080.98
8CNN GRU90.1088.2987.5690.2581.25
9CNN-LSTM89.4390.2385.0982.5891.51
10CNN–ESN90.8587.3981.2289.2178.22
11Dual-Stream CNN–ESN90.9089.7089.5486.9583.66
12PAR-Net92.3188.4092.2389.5794.88
Table 4. Recall comparison of models across different frames.
Table 4. Recall comparison of models across different frames.
PAR Dataset
No.Model Name306090120150
1MLP85.3983.4383.5886.8671.92
2CNN88.8688.0787.7783.5075.36
3LSTM83.2481.2374.1582.8465.89
4BiLSTM90.0585.2489.4182.1167.16
5ESN85.2290.6481.7392.0380.77
6GRU87.7086.6383.6487.7880.24
7BiGRU86.4692.8586.0183.1972.87
8CNN GRU89.2289.0988.6187.5882.32
9CNN-LSTM [36]90.8488.7990.5688.1075.82
10CNN–ESN83.5786.9778.1591.6474.66
11Dual-Stream CNN–ESN88.9989.9286.7190.3482.75
12PAR-Net93.7892.2592.7094.0287.29
PER Dataset
1MLP87.9985.6082.7282.7587.63
2CNN86.9391.7886.1383.8289.48
3LSTM85.1586.7578.7680.0372.39
4BiLSTM91.4687.7174.0880.3484.05
5ESN88.9890.1387.7283.3381.03
6GRU88.9085.1582.8784.6579.55
7BiGRU88.1089.2386.7884.5582.10
8CNN GRU90.5587.9988.1290.7483.14
9CNN-LSTM89.4088.6584.5977.6191.06
10CNN–ESN90.8987.1581.3689.1478.87
11Dual-Stream CNN–ESN90.3589.1287.8588.2083.22
12PAR-Net91.7589.5690.2586.8992.68
Table 5. F1-score comparison of models across different frames.
Table 5. F1-score comparison of models across different frames.
PAR Dataset
No.Model Name306090120150
1MLP85.3783.3283.1286.3670.61
2CNN88.8288.1087.4683.3774.74
3LSTM83.0880.6873.9882.4763.87
4BiLSTM90.1585.4389.4381.7664.45
5ESN85.5590.6681.5391.9680.66
6GRU87.6986.5783.2987.7679.91
7BiGRU86.3992.8685.8283.0372.88
8CNN GRU89.1788.7488.5987.4480.96
9CNN-LSTM [36]90.8588.8790.2987.8075.13
10CNN–ESN83.5686.9478.8991.6374.81
11Dual-Stream CNN–ESN88.9789.9286.5890.2282.57
12PAR-Net93.7392.1892.9094.0087.18
PER Dataset
1MLP87.8285.6782.5282.7687.28
2CNN86.6491.6285.9384.2289.24
3LSTM84.9986.6177.4779.7771.12
4BiLSTM91.3387.7470.4679.4383.44
5ESN87.5088.0585.4784.2182.10
6GRU88.7086.1082.1083.1079.87
7BiGRU87.1089.8986.8983.6581.25
8CNN GRU90.2288.2587.8890.3582.54
9CNN-LSTM89.3489.0484.2277.2991.18
10CNN–ESN90.6687.5881.8789.5578.55
11Dual-Stream CNN–ESN90.8888.9088.2287.7783.50
12PAR-Net91.9488.5490.7087.5193.50
Table 6. Significance test results for comparative analysis of the PAR-Net with other baselines using p-values.
Table 6. Significance test results for comparative analysis of the PAR-Net with other baselines using p-values.
No.Model Namep-Value
1PAR-Net vs. MLP0.015
2PAR-Net vs. CNN0.029
3PAR-Net vs. LSTM0.025
4PAR-Net vs. BiLSTM0.030
5PAR-Net vs. ESN0.029
6PAR-Net vs. GRU0.028
7PAR-Net vs. BiGRU0.029
8PAR-Net vs. CNN-GRU0.033
9PAR-Net vs. CNN-LSTM0.034
10PAR-Net vs. CNN–ESN0.036
11PAR-Net vs. Dual-Stream CNN–ESN0.041
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Khan, I.U.; Lee, J.W. PAR-Net: An Enhanced Dual-Stream CNN–ESN Architecture for Human Physical Activity Recognition. Sensors 2024, 24, 1908. https://doi.org/10.3390/s24061908

AMA Style

Khan IU, Lee JW. PAR-Net: An Enhanced Dual-Stream CNN–ESN Architecture for Human Physical Activity Recognition. Sensors. 2024; 24(6):1908. https://doi.org/10.3390/s24061908

Chicago/Turabian Style

Khan, Imran Ullah, and Jong Weon Lee. 2024. "PAR-Net: An Enhanced Dual-Stream CNN–ESN Architecture for Human Physical Activity Recognition" Sensors 24, no. 6: 1908. https://doi.org/10.3390/s24061908

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop