Non-Contact Cross-Person Activity Recognition by Deep Metric Ensemble Learning
Abstract
:1. Introduction
1.1. Conventional Wi-Fi-Based HAR Methods
1.2. Our Proposal
- 1.
- We used the Espressif 32 (ESP32) CSI tool [34] to collect and open a CSI dataset on the mentioned seven activities in an indoor environment.
- 2.
- A new SE-ABLSTM-C method is proposed to improve the generalization ability of CPAR, obtaining higher accuracies than those of conventional HAR methods.
2. Preliminaries
2.1. Dataset Collection
2.2. Problem and Treatment by Ensemble Learning
3. Proposed CPAR Method
3.1. System Framework
3.1.1. CSI Data Preparation
- By the 3.2 s-time window with 0.8 s-forward sliding, the data is sliced into samples with 200-length. Since the observation time of each round is 16.0 s in our experiments, the number of samples becomes for each subcarrier.
- The sliced samples are labeled as in turn, corresponding to the seven specific activities, i.e., waving, clapping, walking, lying down, sitting down, falling, and picking up.
3.1.2. Feature Extraction by ABLSTM Network
3.1.3. Ensemble and Classifiers Training
3.1.4. CPAR on Unknown Persons
3.2. Snapshot Ensemble Learning
Algorithm 1: SE-ABLSTM: Snapshot Ensemble-used ABLSTM | |||||
Input: Training dataset | |||||
Output: Base-classifiers | |||||
1 | Construct ABLSTM [45] with softmax loss ; | ||||
2 | Set parameters: Initial LR , total number of epochs T, and number of cycles M; | ||||
3 | for do | ||||
4 | Obtain current model weight, i.e., Wm(1) = Wcur; | ||||
5 | for do | ||||
6 | |||||
7 | ; | ||||
8 | ; | ||||
9 | ; | ||||
10 | ; | ||||
11 | end | ||||
12 | Obtain hm with trained Wm; | ||||
13 | end |
3.3. Center Loss Belonging to Metric Learning
4. Experimental Results
4.1. Parameter and Task Setting
4.1.1. Task I
4.1.2. Task II
4.2. Evaluation of Performance in Task I
4.3. Evaluation of Generalization in Task II
4.3.1. Contrastive Losses
4.3.2. Ablation Study
4.3.3. HAR Methods on CPAR
4.4. Limitations
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Ohtsuki, T. A smart city based on ambient intelligence (invited paper). IEICE Trans. Commun. 2017, E100.B, 1547–1553. [Google Scholar] [CrossRef]
- Xiao, Z.; Yu, H.; Yang, Y.; Gu, J.; Li, Y.; Zhuang, F.; Yu, D.; Ren, Z. HarMI: Human activity recognition via multi-modality incremental learning. IEEE J. Biomed. Health Inform. 2022, 26, 939–951. [Google Scholar]
- Basset, M.A.; Hawash, H.; Chang, V.; Chakrabortty, R.K.; Ryan, M.J. Deep learning for heterogeneous human activity recognition in complex IoT applications. IEEE Internet Things J. 2022, 9, 5653–5665. [Google Scholar] [CrossRef]
- Keidar, D.; Yaron, D.; Goldstein, E.; Shachar, Y.; Blass, A.; Charbinsky, L.; Aharony, I.; Lifshitz, L.; Lumelsky, D.; Neeman, Z.; et al. COVID-19 classification of X-ray images using deep neural networks. Eur. Radiol. 2021, 31, 9654–9663. [Google Scholar] [CrossRef] [PubMed]
- Semwal, V.B.; Gupta, A.; Lalwani, P. An optimized hybrid deep learning model using ensemble learning approach for human walking activities recognition. J. Supercomput. 2021, 77, 12256–12279. [Google Scholar] [CrossRef]
- Semwal, V.B.; Lalwani, P.; Mishra, M.K.; Bijalwan, V.; Chadha, J.S. An optimized feature selection using bio-geography optimization technique for human walking activities recognition. Computing 2021, 103, 2893–2914. [Google Scholar] [CrossRef]
- Semwal, V.B.; Jain, R.; Maheshwari, P.; Khatwani, S. Gait reference trajectory generation at different walking speeds using LSTM and CNN. Multimed. Tools Appl. 2023, 82, 33401–33419. [Google Scholar] [CrossRef]
- Semwal, V.B.; Kim, Y.; Bijalwan, V.; Verma, A.; Singh, G.; Gaud, N.; Baek, H.; Khan, A.M. Development of the LSTM model and universal polynomial equation for all the sub-phases of human gait. IEEE Sens. J. 2023, 23, 15892–15900. [Google Scholar] [CrossRef]
- Oscar, L.; Miguel, L. A survey on human activity recognition using wearable sensors. IEEE Commun. Surv. Tutor. 2013, 15, 1192–1209. [Google Scholar]
- Wang, X.; Yu, H.; Kold, S.; Rahbek, O.; Bai, S. Wearable sensors for activity monitoring and motion control: A review. Biomim. Intell. Robot. 2023, 3, 100089. [Google Scholar] [CrossRef]
- Patil, P.; Kumar, K.S.; Gaud, N.; Semwal, V.B. Clinical human gait classification extreme learning machine approach. In Proceedings of the 1st International Conference on Advances in Science, Engineering and Robotics Technology (ICASERT), Dhaka, Bangladesh, 3–5 May 2019. [Google Scholar]
- Dua, N.; Singh, S.N.; Semwal, V.B. Multi-input CNN-GRU based human activity recognition using wearable sensors. Computing 2021, 103, 1461–1478. [Google Scholar] [CrossRef]
- Semwal, V.B.; Gaud, N.; Lalwani, P.; Bijalwan, V.; Alok, A.K. Pattern identification of different human joints for different human walking styles using inertial measurement unit (IMU) sensor. Artif. Intell. Rev. 2022, 55, 1149–1169. [Google Scholar] [CrossRef]
- Challa, S.K.; Kumar, A.; Semwal, V.B. A multibranch CNN-BiLSTM model for human activity recognition using wearable sensor data. Vis. Comput. 2022, 103, 4095–4109. [Google Scholar] [CrossRef]
- Chen, W.; Lyu, M.; Ding, X.; Wang, J.; Zhang, J. Electromyography-controlled lower extremity exoskeleton to provide wearers flexibility in walking. Biomed. Signal Process. Control. 2023, 79, 104096. [Google Scholar] [CrossRef]
- Dua, N.; Singh, S.N.; Semwal, V.B.; Challa, S.K. Inception inspired CNN-GRU hybrid network for human activity recognition. Multimed. Tools Appl. 2023, 82, 5369–5403. [Google Scholar] [CrossRef]
- Liu, C.; Downey, R.J.; Salminen, J.S.; Arvelo, R.S.; Richer, N.; Pliner, E.M.; Hwang, J.; Cruz-Almeida, Y.; Manini, T.M.; Hass, C.J.; et al. Electrical brain activity during human walking with parametric variations in terrain unevenness and walking speed. Imaging Neurosci. 2023, 2, 1–33. [Google Scholar] [CrossRef]
- Djamila, B.; Bini, N.; Mohammad, S.; Abdenour, H. Vision-based human activity recognition: A survey. Multimed. Tools Appl. 2020, 79, 30509–30555. [Google Scholar]
- Zhao, D.; Li, H.; Yan, S. Spatial-temporal synchronous transformer for skeleton-based hand gesture recognition. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 1403–1412. [Google Scholar] [CrossRef]
- Muthukumar, K.A.; Bouazizi, M.; Ohtsuki, T. An infrared array sensor-based approach for activity detection, combining low-cost technology with advanced deep learning techniques. Sensors 2022, 22, 3898. [Google Scholar] [CrossRef]
- Challa, S.K.; Kumar, A.; Semwal, V.B.; Dua, N. An optimized-LSTM and RGB-D sensor-based human gait trajectory generator for bipedal robot walking. IEEE Sens. J. 2022, 22, 24352–24363. [Google Scholar] [CrossRef]
- Habaebi, M.H.; Ali, M.M.; Hassan, M.M.; Shoib, M.S.; Zahrudin, A.A.; Kamarulzaman, A.A.; Azhan, W.S.W.; Islam, M.R. Development of physical intrusion detection system using Wi-Fi/ZigBee RF signals. Sensors 2015, 76, 547–552. [Google Scholar] [CrossRef]
- Arshad, S.; Feng, C.; Liu, Y.; Hu, Y.; Yu, R.; Zhou, S.; Li, H. Wi-chase: A WiFi based human activity recognition system for sensorless environments. In Proceedings of the IEEE 18th International Symposium on A World of Wireless, Mobile and Multimedia Networks (WoWMoM), Macau, China, 12–15 June 2017. [Google Scholar]
- Yousefi, S.; Narui, H.; Dayal, S.; Ermon, S.; Valaee, S. A survey on behavior recognition using WiFi channel state information. IEEE Commun. Mag. 2017, 55, 98–104. [Google Scholar] [CrossRef]
- Ding, X.; Jiang, T.; Zhong, Y.; Wu, S.; Yang, J.; Xue, W. Improving WiFi-based human activity recognition with adaptive initial state via one-shot learning. In Proceedings of the IEEE Wireless Communications and Networking Conference (WCNC), Nanjing, China, 29 March–1 April 2021. [Google Scholar]
- Zhang, Y.; Wang, X.; Wang, Y.; Chen, H. Human activity recognition across scenes and categories based on CSI. IEEE Trans. Mob. Comput. 2022, 21, 2411–2420. [Google Scholar] [CrossRef]
- Forbes, G.; Massie, S.; Craw, S. WiFi-based human activity recognition using Raspberry Pi. In Proceedings of the IEEE 32nd International Conference on Tools with Artificial Intelligence (ICTAI), Baltimore, MD, USA, 9–11 November 2020. [Google Scholar]
- Moshiri, F.; Reza, S.; Mohammad, N.; Ghorashi, S.A. A CSI-based human activity recognition using deep learning. Sensors 2021, 21, 7225. [Google Scholar] [CrossRef] [PubMed]
- Nakamura, T.; Bouazizi, M.; Yamamoto, K.; Ohtsuki, T. Wi-Fi-based fall detection using spectrogram image of channel state information. IEEE Internet Things J. 2022, 9, 17220–17234. [Google Scholar] [CrossRef]
- Islam, M.S.; Jannat, M.K.A.; Hossain, M.N.; Kim, W.-S.; Lee, S.-W.; Yang, S.-H. STC-NLSTMNet: An improved human activity recognition method using convolutional neural network with NLSTM from WiFi CSI. Sensors 2023, 23, 356. [Google Scholar] [CrossRef]
- Bouazizi, M.; Ye, C.; Ohtsuki, T. 2D LIDAR-based approach for activity identification and fall detection. IEEE Internet Things J. 2022, 9, 10872–10890. [Google Scholar] [CrossRef]
- Halperin, D.; Hu, W.; Sheth, A.; Wetherall, D. Tool release: Gathering 802.11n traces with channel state information. ACM SIGCOMM Comput. Commun. Rev. 2011, 41, 53. [Google Scholar] [CrossRef]
- Wang, W.; Liu, A.X.; Shahzad, M.; Ling, K.; Lu, S. Understanding and modeling of wifi signal based human activity recognition. In Proceedings of the 21st Annual International Conference on Mobile Computing and Networking, New York, NY, USA, 7–11 September 2015. [Google Scholar]
- Hernandez, S.M.; Bulut, E. Lightweight and standalone IoT based WiFi sensing for active repositioning and mobility. In Proceedings of the IEEE 21st International Symposium on “A World of Wireless, Mobile and Multimedia Networks” (WoWMoM), Cork, Ireland, 31 August–3 September 2020. [Google Scholar]
- Souvik, S.; Jeongkeun, L.; Kyu-Han, K.; Paul, C. Avoiding multipath to revive inbuilding WiFi localization. In Proceedings of the 11th Annual International Conference on Mobile Systems, Applications, and Services, New York, NY, USA, 25–28 June 2013. [Google Scholar]
- Cristianini, N.; Shawe-Taylor, J. An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods; Cambridge University Press: Cambridge, UK, 2000. [Google Scholar]
- Breiman, L. Random Forests. Mach. Learn. 2021, 45, 5–32. [Google Scholar] [CrossRef]
- Bengio, Y.; Courville, A.; Vincent, P. Representation learning: A review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1798–1828. [Google Scholar] [CrossRef]
- Wang, J.; Zhang, X.; Gao, Q.; Yue, H.; Wang, H. Device-free wireless localization and activity recognition: A deep learning approach. IEEE Trans. Veh. Technol. Mach. Learn. 2017, 66, 6258–6267. [Google Scholar] [CrossRef]
- Hou, C.; Liu, G.; Tian, Q.; Zhou, Z.; Hua, L.; Lin, Y. Multi-signal modulation classification using sliding window detection and complex convolutional network in frequency domain. IEEE Internet Things J. 2022, 9, 19438–19449. [Google Scholar] [CrossRef]
- Wang, Y.; Gui, G.; Lin, Y.; Wu, H.-C.; Yuen, C.; Adachi, F. Few-shot specific emitter identification via deep metric ensemble learning. IEEE Internet Things J. 2022, 9, 24980–24994. [Google Scholar] [CrossRef]
- Jiang, W.; Miao, C.; Ma, F.; Yao, S.; Wang, Y.; Yuan, Y.; Xue, H.; Song, C.; Ma, X.; Koutsonikolas, D.; et al. Towards environment independent device free human activity recognition. In Proceedings of the 24th Annual International Conference on Mobile Computing and Networking, New York, NY, USA, 29 October–2 November 2018. [Google Scholar]
- Yang, J.; Chen, X.; Zou, H.; Wang, D.; Xu, Q.; Xie, L. EfficientFi: Towards large-scale lightweight wifi sensing via csi compression. IEEE Internet Things J. 2022, 9, 13086–13095. [Google Scholar] [CrossRef]
- Zhang, J.; Wu, F.; Wei, B.; Zhang, Q.; Huang, H.; Shah, S.W.; Cheng, J. Data augmentation and dense-LSTM for human activity recognition using WiFi signal. IEEE Internet Things J. 2021, 8, 4628–4641. [Google Scholar] [CrossRef]
- Chen, Z.; Zhang, L.; Jiang, C.; Cao, Z.; Cui, W. WiFi CSI based passive human activity recognition using attention based BLSTM. IEEE Trans. Mob. Comput. 2019, 18, 2714–2724. [Google Scholar] [CrossRef]
- Graves, A.; Mohamed, A.-R.; Hinton, G. Speech recognition with deep recurrent neural networks. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013. [Google Scholar]
- Gers, F.A.; Schmidhuber, J.; Cummins, F. Learning to forget: Continual prediction with LSTM. Neural Comput. 2000, 12, 2451–2471. [Google Scholar] [CrossRef]
- Gers, F.A.; Schraudolph, N.N.; Schmidhuber, J. Learning precise timing with LSTM recurrent networks. J. Mach. Learn. Res. 2002, 3, 115–143. [Google Scholar]
- Li, B.; Cui, W.; Wang, W.; Zhang, L.; Chen, Z.; Wu, M. Two-stream convolution augmented transformer for human activity recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021. [Google Scholar]
- Graves, A.; Schmidhuber, J. Framewise phoneme classification with bidirectional LSTM and other neural network architectures. Neural Netw. 2005, 18, 5–6. [Google Scholar] [CrossRef]
- Cui, W.; Li, B.; Zhang, L.; Chen, Z. Device-free single-user activity recognition using diversified deep ensemble learning. Appl. Soft Comput. 2021, 102, 107066. [Google Scholar] [CrossRef]
- Wang, J.; Lan, C.; Liu, C.; Ouyang, Y.; Qin, T. Generalizing to unseen domains: A survey on domain generalization. IEEE Trans. Knowl. Data Eng. 2021, 35, 8052–8072. [Google Scholar]
- Qian, H.; Pan, S.J.; Miao, C. Latent independent excitation for generalizable sensor-based cross-person activity recognition. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021. [Google Scholar]
- Lu, W.; Wang, J.; Chen, Y. Local and global alignments for generalizable sensor-based human activity recognition. In Proceedings of the ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022. [Google Scholar]
- Liu, S.; Chen, Z.; Wu, M.; Liu, C.; Chen, L. WiSR: Wireless domain generalization based on style randomization. IEEE Trans. Mob. Comput. 2024, 23, 4520–4532. [Google Scholar] [CrossRef]
- Liu, S.; Chen, Z.; Wu, M.; Wang, H.; Xing, B.; Chen, L. Generalizing wireless cross-multiple-factor gesture recognition to unseen domains. IEEE Trans. Mob. Comput. 2024, 23, 5083–5096. [Google Scholar] [CrossRef]
- Huang, G.; Li, Y.; Pleiss, G.; Liu, Z.; Hopcroft, J.E.; Weinberger, K.Q. Snapshot Ensembles: Train 1, get M for free. arXiv 2017, arXiv:1704.00109. [Google Scholar]
- Kaya, M.; Bilge, H.S. Deep metric learning: A survey. Symmetry 2019, 11, 1066. [Google Scholar] [CrossRef]
- Wen, Y.; Zhang, K.; Li, Z.; Qiao, Y. A discriminative feature learning approach for deep face recognition. In Proceedings of the Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016. [Google Scholar]
- Xu, S.; He, Z.; Shi, W.; Wang, Y.; Ohtsuki, T.; Gui, G. Cross-person activity recognition method using snapshot ensemble learning. In Proceedings of the IEEE 96th Vehicular Technology Conference (VTC2022-Fall), London, UK, 26–29 September 2022. [Google Scholar]
- Kwapisz, J.R.; Weiss, G.M.; Moore, S.A. Activity recognition using cell phone accelerometers. ACM SigKDD Explor. Newsl. 2011, 12, 74–82. [Google Scholar] [CrossRef]
- Anguita, D.; Ghio, A.; Oneto, L.; Parra, X.; Reyes-Ortiz, J.L. A public domain dataset for human activity recognition using smartphones. In Proceedings of the European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, Bruges, Belgium, 24–26 April 2013. [Google Scholar]
- Kumar, D.; Jeuris, S.; Bardram, J.E.; Dragoni, N. Mobile and wearable sensing frameworks for mHealth Studies and applications: A systematic review. ACM Trans. Comput. Healthc. 2020, 2, 1–28. [Google Scholar] [CrossRef]
- Dietterich, T.G. Ensemble methods in machine learning. In International Workshop on Multiple Classifier Systems; Springer: Berlin/Heidelberg, Germany, 2000. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
- Wellek, S. Testing Statistical Hypotheses of Equivalence and Noninferiority, 2nd ed.; Chapman and Hall/CRC: Boca Raton, FL, USA, 2010. [Google Scholar]
(a) CSI Datasets for HAR | |||
---|---|---|---|
Study | CSI Collection Tool | No. of Data | Publicity |
(subj. × act. × round) | |||
S. Arshad, et al. [23] | 5300 CSI Tool | 720 (12 × 3 × 20) | No |
S. Yousefi, et al. [24] | 5300 CSI Tool | 720 (6 × 6 × 20) | Yes |
X. Ding, et al. [25] | 5300 CSI Tool | 200–400 (- × 4 × -) | No |
Y. Zhang, et al. [26] | 5300 CSI Tool | 500 (1 × 10 × 50) | No |
G. Forbes, et al. [27] | Raspberry Pi | 1100 (1 × 11 × 100) | No |
F. Moshiri, et al. [28] | Raspberry Pi | 420 (3 × 7 × 20) | Yes |
Our dataset | ESP32 CSI Tool | 700 (5 × 7 × 20) | Yes |
(b) Basic Information of Five Subjects in Our Dataset | |||
Subject | Gender | Height (cm) | Weight (kg) |
A | Male | 172 | 52 |
B | Male | 173 | 60 |
C | Male | 165 | 50 |
D | Male | 175 | 68 |
E | Male | 168 | 59 |
Parameter | Specification |
---|---|
Project | ACTIVE_AP/STA |
Compiler version | IDF v4.3.2 |
No. of antennae | 1 |
Sampling rate | 63 Hz |
No. of subcarriers | 32 |
No. of Wi-Fi channels | 6 |
Packet rate | 100 packet/s |
Voltage | 1100 mV |
Distance of LoS path | 2.5 m |
Height of placement | 1.0 m |
(a) CSI Dataset Preparation | |
---|---|
Parameter | Specification |
Observation time per round | 16.0 s |
Time window length | 3.2 s |
Sliding window length | 0.8 s |
Sample length | 200 |
No. of samples per round | 17 |
Total no. of samples | 11,900 |
(b) Hyper-Parameter for Model Training | |
Parameter | Specification |
Python framework | Tensorflow 2.4.1 |
Loss function | Softmax + Center |
Balance factor | 1.0 |
Optimizer | Adam |
Batch size | 128 |
Initial learning rate | 0.01 |
No. of cycles M | 10 |
No. of epochs per cycle | 50 |
CNN [42] | LSTM [24] | BLSTM | ABLSTM [45] |
---|---|---|---|
91.01% | 92.77% | 94.37% | 97.23% |
Approach | Waving | Clapping | Walking | Lying Down | Sitting Down | Falling | Picking Up | Avg. | Pattern |
---|---|---|---|---|---|---|---|---|---|
ABLSTM [45] | 86.76% | 92.06% | 74.12% | 78.24% | 62.65% | 89.71% | 85.88% | 81.34% | ABCD-E |
ABLSTM-T | 87.35% | 91.47% | 69.12% | 79.41% | 61.18% | 90.29% | 79.41% | 79.75% | |
ABLSTM-C | 84.12% | 88.53% | 75.59% | 81.47% | 57.65% | 92.06% | 89.12% | 81.22% | |
ABLSTM [45] | 94.41% | 86.18% | 85.84% | 88.53% | 66.47% | 6.47% | 75.53% | 71.63% | ABCE-D |
ABLSTM-T | 97.65% | 84.41% | 93.81% | 90.00% | 69.41% | 0.88% | 77.06% | 73.31% | |
ABLSTM-C | 97.35% | 92.94% | 85.84% | 90.00% | 75.79% | 0.88% | 78.53% | 74.48% | |
ABLSTM [45] | 97.35% | 82.06% | 73.16% | 99.71% | 94.71% | 22.12% | 66.18% | 76.49% | ABDE-C |
ABLSTM-T | 98.53% | 77.35% | 46.02% | 100.00% | 96.47% | 17.40% | 72.06% | 72.55% | |
ABLSTM-C | 90.59% | 81.18% | 96.11% | 99.71% | 94.12% | 26.55% | 73.82% | 77.46% | |
ABLSTM [45] | 95.29% | 69.71% | 81.47% | 90.27% | 93.24% | 50.00% | 85.88% | 80.83% | ACDE-B |
ABLSTM-T | 98.53% | 71.47% | 93.53% | 80.83% | 96.47% | 53.82% | 80.88% | 82.22% | |
ABLSTM-C | 97.06% | 80.00% | 92.35% | 89.97% | 95.59% | 49.71% | 90.59% | 85.04% | |
ABLSTM [45] | 91.18% | 92.35% | 92.65% | 98.53% | 98.53% | 70.59% | 42.30% | 83.57% | BCDE-A |
ABLSTM-T | 87.16% | 98.24% | 95.59% | 99.41% | 98.82% | 74.41% | 44.71% | 85.46% | |
ABLSTM-C | 97.65% | 96.47% | 92.94% | 98.53% | 97.06% | 82.06% | 32.65% | 85.34% |
Approach | Waving | Clapping | Walking | Lying Down | Sitting Down | Falling | Picking Up | Avg. | Avg. imp. | Pattern |
---|---|---|---|---|---|---|---|---|---|---|
ABLSTM [45] | 86.76% | 92.06% | 74.12% | 78.24% | 62.65% | 89.71% | 85.88% | 81.34% | 3.03% | ABCD-E |
SE-ABLSTM [60] | 82.65% | 96.18% | 74.41% | 85.29% | 66.18% | 76.47% | 89.12% | 81.47% | 2.90% | |
ABLSTM-C | 84.12% | 88.53% | 75.59% | 81.47% | 57.65% | 92.06% | 89.12% | 81.22% | 3.15% | |
SE-ABLSTM-C (prop.) | 81.18% | 96.47% | 74.71% | 87.06% | 68.82% | 94.41% | 87.94% | 84.37% | – | |
ABLSTM [45] | 94.41% | 86.18% | 85.84% | 88.53% | 66.47% | 6.47% | 75.53% | 71.63% | 7.23% | ABCE-D |
SE-ABLSTM [60] | 97.94% | 94.12% | 91.45% | 92.94% | 68.53% | 2.35% | 83.53% | 75.83% | 3.03% | |
ABLSTM-C | 97.35% | 92.94% | 85.84% | 90.00% | 75.79% | 0.88% | 78.53% | 74.48% | 4.38% | |
SE-ABLSTM-C (prop.) | 98.82% | 96.47% | 91.15% | 94.12% | 80.29% | 3.82% | 87.35% | 78.86% | – | |
ABLSTM [45] | 97.35% | 82.06% | 73.16% | 99.71% | 94.71% | 22.12% | 66.18% | 76.49% | 4.25% | ABDE-C |
SE-ABLSTM [60] | 97.94% | 85.29% | 72.57% | 100.00% | 97.94% | 28.32% | 70.29% | 78.93% | 1.81% | |
ABLSTM-C | 90.59% | 81.18% | 76.11% | 99.71% | 94.12% | 26.55% | 73.82% | 77.46% | 3.28% | |
SE-ABLSTM-C (prop.) | 95.59% | 88.53% | 79.35% | 100.00% | 95.88% | 32.45% | 73.24% | 80.74% | – | |
ABLSTM [45] | 95.29% | 69.71% | 81.47% | 90.27% | 93.24% | 50.00% | 85.88% | 80.83% | 8.20% | ACDE-B |
SE-ABLSTM [60] | 98.53% | 85.29% | 93.53% | 99.12% | 100.00% | 54.41% | 91.18% | 88.87% | 0.16% | |
ABLSTM-C | 97.06% | 80.00% | 92.35% | 89.97% | 95.59% | 49.71% | 90.59% | 85.04% | 3.99% | |
SE-ABLSTM-C (prop.) | 99.41% | 85.59% | 97.65% | 97.64% | 100.00% | 50.88% | 92.06% | 89.03% | – | |
ABLSTM [45] | 91.18% | 92.35% | 92.65% | 98.53% | 98.53% | 70.59% | 42.30% | 83.57% | 3.11% | BCDE-A |
SE-ABLSTM [60] | 96.18% | 95.88% | 95.00% | 100.00% | 100.00% | 75.00% | 37.65% | 85.67% | 1.01% | |
ABLSTM-C | 97.65% | 96.47% | 92.94% | 98.53% | 97.06% | 82.06% | 32.65% | 85.34% | 1.34% | |
SE-ABLSTM-C (prop.) | 96.47% | 99.12% | 94.41% | 100.00% | 98.24% | 81.47% | 37.06% | 86.68% | – |
Approach | Waving | Clapping | Walking | Lying Down | Sitting Down | Falling | Picking Up | Avg. | Pattern |
---|---|---|---|---|---|---|---|---|---|
CNN [42] | 50.59% | 92.65% | 67.65% | 81.76% | 65.29% | 84.12% | 65.88% | 72.56% | ABCD-E |
LSTM [24] | 96.76% | 92.65% | 76.18% | 82.35% | 65.00% | 85.29% | 89.41% | 83.95% | |
ABLSTM [45] | 86.76% | 92.06% | 74.12% | 78.24% | 62.65% | 89.71% | 85.88% | 81.34% | |
LAGMAT [54] | 35.88% | 91.76% | 75.29% | 83.24% | 57.06% | 73.53% | 77.65% | 70.63% | |
SE-ABLSTM-C (prop.) | 81.18% | 96.47% | 74.71% | 87.06% | 68.82% | 94.41% | 87.94% | 84.37% | |
CNN [42] | 92.94% | 94.12% | 90.86% | 88.24% | 79.71% | 2.94% | 75.00% | 74.83% | ABCE-D |
LSTM [24] | 95.88% | 61.47% | 83.78% | 83.24% | 67.94% | 2.65% | 85.29% | 68.61% | |
ABLSTM [45] | 94.41% | 86.18% | 85.84% | 88.53% | 66.47% | 6.47% | 75.53% | 71.63% | |
LAGMAT [54] | 77.65% | 84.71% | 72.57% | 95.59% | 67.06% | 62.65% | 78.82% | 77.01% | |
SE-ABLSTM-C (prop.) | 98.82% | 96.47% | 91.15% | 94.12% | 80.29% | 3.82% | 87.35% | 78.86% | |
CNN [42] | 96.47% | 19.71% | 77.88% | 100.00% | 96.76% | 38.05% | 80.59% | 72.78% | ABDE-C |
LSTM [24] | 97.94% | 82.35% | 43.36% | 100.00% | 95.29% | 9.14% | 70.00% | 71.15% | |
ABLSTM [45] | 97.35% | 82.06% | 73.16% | 99.71% | 94.71% | 22.12% | 66.18% | 76.49% | |
LAGMAT [54] | 97.06% | 65.59% | 83.78% | 95.29% | 71.18% | 53.69% | 72.06% | 76.95% | |
SE-ABLSTM-C (prop.) | 95.59% | 88.53% | 79.35% | 100.00% | 95.88% | 32.45% | 73.24% | 80.74% | |
CNN [42] | 97.94% | 22.65% | 95.88% | 88.79% | 97.35% | 66.76% | 94.71% | 80.58% | ACDE-B |
LSTM [24] | 99.41% | 71.76% | 95.88% | 47.20% | 97.35% | 57.65% | 77.94% | 78.17% | |
ABLSTM [45] | 95.29% | 69.71% | 81.47% | 90.27% | 93.24% | 50.00% | 85.88% | 80.83% | |
LAGMAT [54] | 91.18% | 81.18% | 83.53% | 83.19% | 79.12% | 63.82% | 90.59% | 81.80% | |
SE-ABLSTM-C (prop.) | 99.41% | 85.89% | 97.65% | 97.64% | 100.00% | 50.88% | 92.06% | 89.03% | |
CNN [42] | 86.76% | 99.12% | 94.41% | 99.71% | 98.24% | 71.76% | 37.94% | 83.99% | BCDE-A |
LSTM [24] | 93.82% | 97.94% | 88.82% | 98.53% | 98.24% | 71.18% | 33.82% | 83.29% | |
ABLSTM [45] | 91.18% | 92.35% | 92.65% | 98.53% | 98.53% | 70.59% | 42.30% | 83.57% | |
LAGMAT [54] | 78.53% | 94.41% | 86.47% | 94.12% | 90.00% | 74.41% | 81.76% | 85.67% | |
SE-ABLSTM-C (prop.) | 96.47% | 99.12% | 94.41% | 100.00% | 98.24% | 81.47% | 37.06% | 86.68% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Ye, C.; Xu, S.; He, Z.; Yin, Y.; Ohtsuki, T.; Gui, G. Non-Contact Cross-Person Activity Recognition by Deep Metric Ensemble Learning. Bioengineering 2024, 11, 1124. https://doi.org/10.3390/bioengineering11111124
Ye C, Xu S, He Z, Yin Y, Ohtsuki T, Gui G. Non-Contact Cross-Person Activity Recognition by Deep Metric Ensemble Learning. Bioengineering. 2024; 11(11):1124. https://doi.org/10.3390/bioengineering11111124
Chicago/Turabian StyleYe, Chen, Siyuan Xu, Zhengran He, Yue Yin, Tomoaki Ohtsuki, and Guan Gui. 2024. "Non-Contact Cross-Person Activity Recognition by Deep Metric Ensemble Learning" Bioengineering 11, no. 11: 1124. https://doi.org/10.3390/bioengineering11111124
APA StyleYe, C., Xu, S., He, Z., Yin, Y., Ohtsuki, T., & Gui, G. (2024). Non-Contact Cross-Person Activity Recognition by Deep Metric Ensemble Learning. Bioengineering, 11(11), 1124. https://doi.org/10.3390/bioengineering11111124