Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (45)

Search Parameters:
Keywords = WISDM

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
36 pages, 8122 KB  
Article
Human Activity Recognition via Attention-Augmented TCN-BiGRU Fusion
by Ji-Long He, Jian-Hong Wang, Chih-Min Lo and Zhaodi Jiang
Sensors 2025, 25(18), 5765; https://doi.org/10.3390/s25185765 - 16 Sep 2025
Viewed by 694
Abstract
With the widespread application of wearable sensors in health monitoring and human–computer interaction, deep learning-based human activity recognition (HAR) research faces challenges such as the effective extraction of multi-scale temporal features and the enhancement of robustness against noise in multi-source data. This study [...] Read more.
With the widespread application of wearable sensors in health monitoring and human–computer interaction, deep learning-based human activity recognition (HAR) research faces challenges such as the effective extraction of multi-scale temporal features and the enhancement of robustness against noise in multi-source data. This study proposes the TGA-HAR (TCN-GRU-Attention-HAR) model. The TGA-HAR model integrates Temporal Convolutional Neural Networks and Recurrent Neural Networks by constructing a hierarchical feature abstraction architecture through cascading Temporal Convolutional Network (TCN) and Bidirectional Gated Recurrent Unit (BiGRU) layers for complex activity recognition. This study utilizes TCN layers with dilated convolution kernels to extract multi-order temporal features. This study utilizes BiGRU layers to capture bidirectional temporal contextual correlation information. To further optimize feature representation, the TGA-HAR model introduces residual connections to enhance the stability of gradient propagation and employs an adaptive weighted attention mechanism to strengthen feature representation. The experimental results of this study demonstrate that the model achieved test accuracies of 99.37% on the WISDM dataset, 95.36% on the USC-HAD dataset, and 96.96% on the PAMAP2 dataset. Furthermore, we conducted tests on datasets collected in real-world scenarios. This method provides a highly robust solution for complex human activity recognition tasks. Full article
(This article belongs to the Section Biomedical Sensors)
Show Figures

Figure 1

15 pages, 1304 KB  
Article
Conv-ScaleNet: A Multiscale Convolutional Model for Federated Human Activity Recognition
by Xian Wu Ting, Ying Han Pang, Zheng You Lim, Shih Yin Ooi and Fu San Hiew
AI 2025, 6(9), 218; https://doi.org/10.3390/ai6090218 - 8 Sep 2025
Viewed by 526
Abstract
Background: Artificial Intelligence (AI) techniques have been extensively deployed in sensor-based Human Activity Recognition (HAR) systems. Recent advances in deep learning, especially Convolutional Neural Networks (CNNs), have advanced HAR by enabling automatic feature extraction from raw sensor data. However, these models often struggle [...] Read more.
Background: Artificial Intelligence (AI) techniques have been extensively deployed in sensor-based Human Activity Recognition (HAR) systems. Recent advances in deep learning, especially Convolutional Neural Networks (CNNs), have advanced HAR by enabling automatic feature extraction from raw sensor data. However, these models often struggle to capture multiscale patterns in human activity, limiting recognition accuracy. Additionally, traditional centralized learning approaches raise data privacy concerns, as personal sensor data must be transmitted to a central server, increasing the risk of privacy breaches. Methods: To address these challenges, this paper introduces Conv-ScaleNet, a CNN-based model designed for multiscale feature learning and compatibility with federated learning (FL) environments. Conv-ScaleNet integrates a Pyramid Pooling Module to extract both fine-grained and coarse-grained features and employs sequential Global Average Pooling layers to progressively capture abstract global representations from inertial sensor data. The model supports federated learning by training locally on user devices, sharing only model updates rather than raw data, thus preserving user privacy. Results: Experimental results demonstrate that the proposed Conv-ScaleNet achieves approximately 98% and 96% F1-scores on the WISDM and UCI-HAR datasets, respectively, confirming its competitiveness in FL environments for activity recognition. Conclusions: The proposed Conv-ScaleNet model addresses key limitations of existing HAR systems by combining multiscale feature learning with privacy-preserving training. Its strong performance, data protection capability, and adaptability to decentralized environments make it a robust and scalable solution for real-world HAR applications. Full article
Show Figures

Figure 1

29 pages, 7847 KB  
Article
Depthwise-Separable U-Net for Wearable Sensor-Based Human Activity Recognition
by Yoo-Kyung Lee, Chang-Sik Son and Won-Seok Kang
Appl. Sci. 2025, 15(16), 9134; https://doi.org/10.3390/app15169134 - 19 Aug 2025
Viewed by 628
Abstract
In wearable sensor-based human activity recognition (HAR), the traditional sliding window method encounters the challenge of multiclass windows in which multiple actions are combined within a single window. To address this problem, an approach that predicts activities at each point in time within [...] Read more.
In wearable sensor-based human activity recognition (HAR), the traditional sliding window method encounters the challenge of multiclass windows in which multiple actions are combined within a single window. To address this problem, an approach that predicts activities at each point in time within a sequence has been proposed, and U-Net-based models have proven to be effective owing to their excellent space-time feature restoration capabilities. However, these models have limitations in that they are prone to overfitting owing to their large number of parameters and are not suitable for deployment. In this study, a lightweight U-Net was designed by replacing all standard U-Net convolutions with depthwise separable convolutions to implement dense prediction. Compared with existing U-Net-based models, the proposed model reduces the number of parameters by 57–89%. When evaluated on three benchmark datasets (MHEALTH, PAMAP2, and WISDM) using subject-independent splits, the performance of the proposed model was equal to or superior to that of all comparison models. Notably, on the MHEALTH dataset, which was collected in an uncontrolled environment, the proposed model improved accuracy by 7.89%, demonstrating its applicability to real-world wearable HAR systems. Full article
Show Figures

Figure 1

14 pages, 1992 KB  
Article
G-CTRNN: A Trainable Low-Power Continuous-Time Neural Network for Human Activity Recognition in Healthcare Applications
by Abdallah Alzubi, David Lin, Johan Reimann and Fadi Alsaleem
Appl. Sci. 2025, 15(13), 7508; https://doi.org/10.3390/app15137508 - 4 Jul 2025
Viewed by 3003
Abstract
Continuous-time Recurrent Neural Networks (CTRNNs) are well-suited for modeling temporal dynamics in low-power neuromorphic and analog computing systems, making them promising candidates for edge-based human activity recognition (HAR) in healthcare. However, training CTRNNs remains challenging due to their continuous-time nature and the need [...] Read more.
Continuous-time Recurrent Neural Networks (CTRNNs) are well-suited for modeling temporal dynamics in low-power neuromorphic and analog computing systems, making them promising candidates for edge-based human activity recognition (HAR) in healthcare. However, training CTRNNs remains challenging due to their continuous-time nature and the need to respect physical hardware constraints. In this work, we propose G-CTRNN, a novel gradient-based training framework for analog-friendly CTRNNs designed for embedded healthcare applications. Our method extends Backpropagation Through Time (BPTT) to continuous domains using TensorFlow’s automatic differentiation, while enforcing constraints on time constants and synaptic weights to ensure hardware compatibility. We validate G-CTRNN on the WISDM human activity dataset, which simulates realistic wearable sensor data for healthcare monitoring. Compared to conventional RNNs, G-CTRNN achieves superior classification accuracy with fewer parameters and greater stability—enabling continuous, real-time HAR on low-power platforms such as MEMS computing networks. The proposed framework provides a pathway toward on-device AI for remote patient monitoring, elderly care, and personalized healthcare in resource-constrained environments. Full article
(This article belongs to the Special Issue Human Activity Recognition (HAR) in Healthcare, 3rd Edition)
Show Figures

Figure 1

27 pages, 10754 KB  
Article
Efficient and Explainable Human Activity Recognition Using Deep Residual Network with Squeeze-and-Excitation Mechanism
by Sakorn Mekruksavanich and Anuchit Jitpattanakul
Appl. Syst. Innov. 2025, 8(3), 57; https://doi.org/10.3390/asi8030057 - 24 Apr 2025
Cited by 1 | Viewed by 1956
Abstract
Wearable sensors for human activity recognition (HAR) have gained significant attention across multiple domains, such as personal health monitoring and intelligent home systems. Despite notable advancements in deep learning for HAR, understanding the decision-making process of complex models remains challenging. This study introduces [...] Read more.
Wearable sensors for human activity recognition (HAR) have gained significant attention across multiple domains, such as personal health monitoring and intelligent home systems. Despite notable advancements in deep learning for HAR, understanding the decision-making process of complex models remains challenging. This study introduces an advanced deep residual network integrated with a squeeze-and-excitation (SE) mechanism to improve recognition accuracy and model interpretability. The proposed model, ConvResBiGRU-SE, was tested using the UCI-HAR and WISDM datasets. It achieved remarkable accuracies of 99.18% and 98.78%, respectively, surpassing existing state-of-the-art methods. The SE mechanism enhanced the model’s ability to focus on essential features, while gradient-weighted class activation mapping (Grad-CAM) increased interpretability by highlighting essential sensory data influencing predictions. Additionally, ablation experiments validated the contribution of each component to the model’s overall performance. This research advances HAR technology by offering a more transparent and efficient recognition system. The enhanced transparency and predictive accuracy may increase user trust and facilitate smoother integration into real-world applications. Full article
(This article belongs to the Special Issue Smart Sensors and Devices: Recent Advances and Applications Volume II)
Show Figures

Figure 1

22 pages, 1052 KB  
Article
Enhancing Sensor-Based Human Physical Activity Recognition Using Deep Neural Networks
by Minyar Sassi Hidri, Adel Hidri, Suleiman Ali Alsaif, Muteeb Alahmari and Eman AlShehri
J. Sens. Actuator Netw. 2025, 14(2), 42; https://doi.org/10.3390/jsan14020042 - 14 Apr 2025
Cited by 2 | Viewed by 1712
Abstract
Human activity recognition (HAR) is the task of classifying sequences of data into defined movements. Taking advantage of deep learning (DL) methods, this research investigates and optimizes neural network architectures to effectively classify physical activities from smartphone accelerometer data. Unlike traditional machine learning [...] Read more.
Human activity recognition (HAR) is the task of classifying sequences of data into defined movements. Taking advantage of deep learning (DL) methods, this research investigates and optimizes neural network architectures to effectively classify physical activities from smartphone accelerometer data. Unlike traditional machine learning (ML) methods employing manually crafted features, our approach employs automated feature learning with three deep learning architectures: Convolutional Neural Networks (CNN), CNN-based autoencoders, and Long Short-Term Memory Recurrent Neural Networks (LSTM RNN). The contribution of this work is primarily in optimizing LSTM RNN to leverage the most out of temporal relationships between sensor data, significantly improving classification accuracy. Experimental outcomes for the WISDM dataset show that the proposed LSTM RNN model achieves 96.1% accuracy, outperforming CNN-based approaches and current ML-based methods. Compared to current works, our optimized frameworks achieve up to 6.4% higher classification performance, which means that they are more appropriate for real-time HAR. Full article
Show Figures

Figure 1

23 pages, 1479 KB  
Article
A Multi-Agent and Attention-Aware Enhanced CNN-BiLSTM Model for Human Activity Recognition for Enhanced Disability Assistance
by Mst Alema Khatun, Mohammad Abu Yousuf, Taskin Noor Turna, AKM Azad, Salem A. Alyami and Mohammad Ali Moni
Diagnostics 2025, 15(5), 537; https://doi.org/10.3390/diagnostics15050537 - 22 Feb 2025
Cited by 2 | Viewed by 2705
Abstract
Background: Artificial intelligence (AI)-based automated human activity recognition (HAR) is essential in enhancing assistive technologies for disabled individuals, focusing on fall detection, tracking rehabilitation progress, and analyzing personalized movement patterns. It also significantly manages and grows multiple industries, such as surveillance, sports, and [...] Read more.
Background: Artificial intelligence (AI)-based automated human activity recognition (HAR) is essential in enhancing assistive technologies for disabled individuals, focusing on fall detection, tracking rehabilitation progress, and analyzing personalized movement patterns. It also significantly manages and grows multiple industries, such as surveillance, sports, and diagnosis. Methods: This paper proposes a novel strategy using a three-stage feature ensemble combining deep learning (DL) and machine learning (ML) for accurate and automatic classification of activity recognition. We develop a unique activity detection approach in this study by enhancing the state-of-the-art convolutional neural network (CNN) and bi-directional long short-term memory (BiLSTM) models with selective ML classifiers and an attention mechanism. Thus, we developed an ensemble activity recognition model, namely “Attention-CNN-BiLSTM with selective ML”. Results: Out of the nine ML models and four DL models, the top performers are selected and combined in three stages for feature extraction. The effectiveness of this three-stage ensemble strategy is evaluated utilizing various performance metrics and through three distinct experiments. Utilizing the publicly available datasets (i.e., the UCI-HAR dataset and WISDM), our approach has shown superior predictive accuracy (98.75% and 99.58%, respectively). When compared with other methods, namely CNN, LSTM, CNN-BiLSTM, and Attention-CNN-BiLSTM, our approach surpasses them in terms of effectiveness, accuracy, and practicability. Conclusions: We hope that this comprehensive activity recognition system may be augmented with an advanced disability monitoring and diagnosis system to facilitate predictive assistance and personalized rehabilitation strategies. Full article
(This article belongs to the Special Issue AI and Digital Health for Disease Diagnosis and Monitoring)
Show Figures

Figure 1

20 pages, 691 KB  
Article
DiscHAR: A Discrete Approach to Enhance Human Activity Recognition in Cyber Physical Systems: Smart Homes
by Ishrat Fatima, Asma Ahmad Farhan, Maria Tamoor, Shafiq ur Rehman, Hisham Abdulrahman Alhulayyil and Fawaz Tariq
Computers 2024, 13(11), 300; https://doi.org/10.3390/computers13110300 - 19 Nov 2024
Cited by 2 | Viewed by 1381
Abstract
The main challenges in smart home systems and cyber-physical systems come from not having enough data and unclear interpretation; thus, there is still a lot to be done in this field. In this work, we propose a practical approach called Discrete Human Activity [...] Read more.
The main challenges in smart home systems and cyber-physical systems come from not having enough data and unclear interpretation; thus, there is still a lot to be done in this field. In this work, we propose a practical approach called Discrete Human Activity Recognition (DiscHAR) based on prior research to enhance Human Activity Recognition (HAR). Our goal is to generate diverse data to build better models for activity classification. To tackle overfitting, which often occurs with small datasets, we generate data and convert them into discrete forms, improving classification accuracy. Our methodology includes advanced techniques like the R-Frame method for sampling and the Mixed-up approach for data generation. We apply K-means vector quantization to categorize the data, and through the elbow method, we determine the optimal number of clusters. The discrete sequences are converted into one-hot encoded vectors and fed into a CNN model to ensure precise recognition of human activities. Evaluations on the OPP79, PAMAP2, and WISDM datasets show that our approach outperforms existing models, achieving 89% accuracy for OPP79, 93.24% for PAMAP2, and 100% for WISDM. These results demonstrate the model’s effectiveness in identifying complex activities captured by wearable devices. Our work combines theory and practice to address ongoing challenges in this field, aiming to improve the reliability and performance of activity recognition systems in dynamic environments. Full article
Show Figures

Figure 1

22 pages, 577 KB  
Article
Efficient Human Activity Recognition on Wearable Devices Using Knowledge Distillation Techniques
by Paulo H. N. Gonçalves, Hendrio Bragança and Eduardo Souto
Electronics 2024, 13(18), 3612; https://doi.org/10.3390/electronics13183612 - 11 Sep 2024
Cited by 4 | Viewed by 2610
Abstract
Mobile and wearable devices have revolutionized the field of continuous user activity monitoring. However, analyzing the vast and intricate data captured by the sensors of these devices poses significant challenges. Deep neural networks have shown remarkable accuracy in Human Activity Recognition (HAR), but [...] Read more.
Mobile and wearable devices have revolutionized the field of continuous user activity monitoring. However, analyzing the vast and intricate data captured by the sensors of these devices poses significant challenges. Deep neural networks have shown remarkable accuracy in Human Activity Recognition (HAR), but their application on mobile and wearable devices is constrained by limited computational resources. To address this limitation, we propose a novel method called Knowledge Distillation for Human Activity Recognition (KD-HAR) that leverages the knowledge distillation technique to compress deep neural network models for HAR using inertial sensor data. Our approach transfers the acquired knowledge from high-complexity teacher models (state-of-the-art models) to student models with reduced complexity. This compression strategy allows us to maintain performance while keeping computational costs low. To assess the compression capabilities of our approach, we evaluate it using two popular databases (UCI-HAR and WISDM) comprising inertial sensor data from smartphones. Our results demonstrate that our method achieves competitive accuracy, even at compression rates ranging from 18 to 42 times the number of parameters compared to the original teacher model. Full article
Show Figures

Figure 1

14 pages, 1280 KB  
Article
Multihead-Res-SE Residual Network with Attention for Human Activity Recognition
by Hongbo Kang, Tailong Lv, Chunjie Yang and Wenqing Wang
Electronics 2024, 13(17), 3407; https://doi.org/10.3390/electronics13173407 - 27 Aug 2024
Cited by 2 | Viewed by 2805
Abstract
Human activity recognition (HAR) typically uses wearable sensors to identify and analyze the time-series data they collect, enabling recognition of specific actions. As such, HAR is increasingly applied in human–computer interaction, healthcare, and other fields, making accurate and efficient recognition of various human [...] Read more.
Human activity recognition (HAR) typically uses wearable sensors to identify and analyze the time-series data they collect, enabling recognition of specific actions. As such, HAR is increasingly applied in human–computer interaction, healthcare, and other fields, making accurate and efficient recognition of various human activities. In recent years, deep learning methods have been extensively applied in sensor-based HAR, yielding remarkable results. However, complex HAR research, which involves specific human behaviors in varied contexts, still faces several challenges. To solve these problems, we propose a multi-head neural network based on the attention mechanism. This framework contains three convolutional heads, with each head designed using one-dimensional CNN to extract features from sensory data. The model uses a channel attention module (squeeze–excitation module) to enhance the representational capabilities of convolutional neural networks. We conducted experiments on two publicly available benchmark datasets, UCI-HAR and WISDM, to evaluate our model. The results were satisfactory, with overall recognition accuracies of 96.72% and 97.73% on their respective datasets. The experimental results demonstrate the effectiveness of the network structure for the HAR, which ensures a higher level of accuracy. Full article
(This article belongs to the Special Issue Deep Learning-Based Object Detection/Classification)
Show Figures

Figure 1

34 pages, 2564 KB  
Article
Achieving More with Less: A Lightweight Deep Learning Solution for Advanced Human Activity Recognition (HAR)
by Sarab AlMuhaideb, Lama AlAbdulkarim, Deemah Mohammed AlShahrani, Hessah AlDhubaib and Dalal Emad AlSadoun
Sensors 2024, 24(16), 5436; https://doi.org/10.3390/s24165436 - 22 Aug 2024
Cited by 8 | Viewed by 3382
Abstract
Human activity recognition (HAR) is a crucial task in various applications, including healthcare, fitness, and the military. Deep learning models have revolutionized HAR, however, their computational complexity, particularly those involving BiLSTMs, poses significant challenges for deployment on resource-constrained devices like smartphones. While BiLSTMs [...] Read more.
Human activity recognition (HAR) is a crucial task in various applications, including healthcare, fitness, and the military. Deep learning models have revolutionized HAR, however, their computational complexity, particularly those involving BiLSTMs, poses significant challenges for deployment on resource-constrained devices like smartphones. While BiLSTMs effectively capture long-term dependencies by processing inputs bidirectionally, their high parameter count and computational demands hinder practical applications in real-time HAR. This study investigates the approximation of the computationally intensive BiLSTM component in a HAR model by using a combination of alternative model components and data flipping augmentation. The proposed modifications to an existing hybrid model architecture replace the BiLSTM with standard and residual LSTM, along with convolutional networks, supplemented by data flipping augmentation to replicate the context awareness typically provided by BiLSTM networks. The results demonstrate that the residual LSTM (ResLSTM) model achieves superior performance while maintaining a lower computational complexity compared to the traditional BiLSTM model. Specifically, on the UCI-HAR dataset, the ResLSTM model attains an accuracy of 96.34% with 576,702 parameters, outperforming the BiLSTM model’s accuracy of 95.22% with 849,534 parameters. On the WISDM dataset, the ResLSTM achieves an accuracy of 97.20% with 192,238 parameters, compared to the BiLSTM’s 97.23% accuracy with 283,182 parameters, demonstrating a more efficient architecture with minimal performance trade-off. For the KU-HAR dataset, the ResLSTM model achieves an accuracy of 97.05% with 386,038 parameters, showing comparable performance to the BiLSTM model’s 98.63% accuracy with 569,462 parameters, but with significantly fewer parameters. Full article
(This article belongs to the Special Issue Intelligent Wearable Sensor-Based Gait and Movement Analysis)
Show Figures

Figure 1

16 pages, 1160 KB  
Article
BSTCA-HAR: Human Activity Recognition Model Based on Wearable Mobile Sensors
by Yan Yuan, Lidong Huang, Xuewen Tan, Fanchang Yang and Shiwei Yang
Appl. Sci. 2024, 14(16), 6981; https://doi.org/10.3390/app14166981 - 9 Aug 2024
Cited by 2 | Viewed by 1682
Abstract
Sensor-based human activity recognition has been widely used in various fields; however, there are still challenges involving recognition of daily complex human activities using sensors. In order to solve the problem of timeliness and homogeneity of recognition functions in human activity recognition models, [...] Read more.
Sensor-based human activity recognition has been widely used in various fields; however, there are still challenges involving recognition of daily complex human activities using sensors. In order to solve the problem of timeliness and homogeneity of recognition functions in human activity recognition models, we propose a human activity recognition model called ’BSTCA-HAR’ based on a long short-term memory (LSTM) network. The approach proposed in this paper combines an attention mechanism and a temporal convolutional network (TCN). The learning and prediction units in the model can efficiently learn important action data while capturing long time-dependent information as well as features at different time scales. Our series of experiments on three public datasets (WISDM, UCI-HAR, and ISLD) with different data features confirm the feasibility of the proposed method. This method excels in dynamically capturing action features while maintaining a low number of parameters and achieving a remarkable average accuracy of 93%, proving that the model has good recognition performance. Full article
Show Figures

Figure 1

18 pages, 5005 KB  
Article
SensorNet: An Adaptive Attention Convolutional Neural Network for Sensor Feature Learning
by Jiaqi Ge, Gaochao Xu, Jianchao Lu, Xu Xu, Long Li and Xiangyu Meng
Sensors 2024, 24(11), 3274; https://doi.org/10.3390/s24113274 - 21 May 2024
Cited by 1 | Viewed by 1701
Abstract
This work develops a generalizable neural network, SENSORNET, for sensor feature learning across various applications. The primary challenge addressed is the poor portability of pretrained neural networks to new applications with limited sensor data. To solve this challenge, we design SensorNet, which [...] Read more.
This work develops a generalizable neural network, SENSORNET, for sensor feature learning across various applications. The primary challenge addressed is the poor portability of pretrained neural networks to new applications with limited sensor data. To solve this challenge, we design SensorNet, which integrates the flexibility of self-attention with multi-scale feature locality of convolution. Moreover, we invent patch-wise self-attention with stacked multi-heads to enrich the sensor feature representation. SensorNet is generalizable to pervasive applications with any number of sensor inputs, and is much smaller than the state-of-the-art self-attention and convolution hybrid baseline (0.83 M vs. 3.87 M parameters) with similar performance. The experimental results show that SensorNet is able to achieve state-of-the-art performance compared with the top five models on a competition activity recognition dataset (SHL’18). Moreover, pretrained SensorNet in a large inertial measurement unit (IMU) dataset can be fine-tuned to achieve the best accuracy on a much smaller IMU dataset (up to 5% improvement in WISDM) and to achieve the state-of-the-art performance on an EEG dataset (SLEEP-EDF-20), showing the strong generalizability of our approach. Full article
(This article belongs to the Special Issue Smart Sensor Integration in Wearables)
Show Figures

Figure 1

19 pages, 7191 KB  
Article
Quantitative Analysis of Mother Wavelet Function Selection for Wearable Sensors-Based Human Activity Recognition
by Heba Nematallah and Sreeraman Rajan
Sensors 2024, 24(7), 2119; https://doi.org/10.3390/s24072119 - 26 Mar 2024
Cited by 6 | Viewed by 2409
Abstract
Recent advancements in the Internet of Things (IoT) wearable devices such as wearable inertial sensors have increased the demand for precise human activity recognition (HAR) with minimal computational resources. The wavelet transform, which offers excellent time-frequency localization characteristics, is well suited for HAR [...] Read more.
Recent advancements in the Internet of Things (IoT) wearable devices such as wearable inertial sensors have increased the demand for precise human activity recognition (HAR) with minimal computational resources. The wavelet transform, which offers excellent time-frequency localization characteristics, is well suited for HAR recognition systems. Selecting a mother wavelet function in wavelet analysis is critical, as optimal selection improves the recognition performance. The activity time signals data have different periodic patterns that can discriminate activities from each other. Therefore, selecting a mother wavelet function that closely resembles the shape of the recognized activity’s sensor (inertial) signals significantly impacts recognition performance. This study uses an optimal mother wavelet selection method that combines wavelet packet transform with the energy-to-Shannon-entropy ratio and two classification algorithms: decision tree (DT) and support vector machines (SVM). We examined six different mother wavelet families with different numbers of vanishing points. Our experiments were performed on eight publicly available ADL datasets: MHEALTH, WISDM Activity Prediction, HARTH, HARsense, DaLiAc, PAMAP2, REALDISP, and HAR70+. The analysis demonstrated in this paper can be used as a guideline for optimal mother wavelet selection for human activity recognition. Full article
(This article belongs to the Special Issue Wearable Sensors for Behavioral and Physiological Monitoring)
Show Figures

Figure 1

13 pages, 1048 KB  
Article
Using an Interpretable Amino Acid-Based Machine Learning Method to Enhance the Diagnosis of Major Depressive Disorder
by Cyrus Su Hui Ho, Trevor Wei Kiat Tan, Howard Cai Hao Khoe, Yee Ling Chan, Gabrielle Wann Nii Tay and Tong Boon Tang
J. Clin. Med. 2024, 13(5), 1222; https://doi.org/10.3390/jcm13051222 - 21 Feb 2024
Cited by 5 | Viewed by 2271
Abstract
Background: Major depressive disorder (MDD) is a leading cause of disability worldwide. At present, however, there are no established biomarkers that have been validated for diagnosing and treating MDD. This study sought to assess the diagnostic and predictive potential of the differences in [...] Read more.
Background: Major depressive disorder (MDD) is a leading cause of disability worldwide. At present, however, there are no established biomarkers that have been validated for diagnosing and treating MDD. This study sought to assess the diagnostic and predictive potential of the differences in serum amino acid concentration levels between MDD patients and healthy controls (HCs), integrating them into interpretable machine learning models. Methods: In total, 70 MDD patients and 70 HCs matched in age, gender, and ethnicity were recruited for the study. Serum amino acid profiling was conducted by means of chromatography-mass spectrometry. A total of 21 metabolites were analysed, with 17 from a preset amino acid panel and the remaining 4 from a preset kynurenine panel. Logistic regression was applied to differentiate MDD patients from HCs. Results: The best-performing model utilised both feature selection and hyperparameter optimisation and yielded a moderate area under the receiver operating curve (AUC) classification value of 0.76 on the testing data. The top five metabolites identified as potential biomarkers for MDD were 3-hydroxy-kynurenine, valine, kynurenine, glutamic acid, and xanthurenic acid. Conclusions: Our study highlights the potential of using an interpretable machine learning analysis model based on amino acids to aid and increase the diagnostic accuracy of MDD in clinical practice. Full article
(This article belongs to the Section Mental Health)
Show Figures

Figure 1

Back to TopTop