Human Activity Recognition Based on an Efficient Neural Architecture Search Framework Using Evolutionary Multi-Objective Surrogate-Assisted Algorithms
Abstract
:1. Introduction
- (1)
- We propose an efficient NAS for HAR application. The proposed method significantly reduces the training cost using the surrogate model. Besides, we propose a CNN–LSTM-based model for search space, which uses CNN to extract the features of the data automatically and uses an LSTM neural network to classify the action into a specific category.
- (2)
- We add floating-point operations per second (FLOPs) and the number of parameters into the model as the second and third objectives. The objective function ensures the model has low computation and communication overhead, which can achieve excellent performance in resource-limited scenarios.
- (3)
- The portability of the proposed method is proved by migrating the models trained on the OPPORTUNITY dataset to the UniMiB-SHAR dataset. The experimental results show that the model obtained through the searched network and the surrogate model can be applied to data with different distributions, which is of great significance for practical application.
2. Literature Review and Background
2.1. Literature Review on NAS Theory and Applications
2.2. Literature Review on HAR Methods
2.3. Background and Significance
3. Methology
3.1. Search Space
3.2. Search Strategy
Algorithm 1 General process of NSGA-II. |
Input: Populations P, number of generations g, offspring S, basic operations O, select operations , crossover operations , mutation operations Output: Final population
|
3.3. Surrogate Model
3.4. Proposed Approach
4. Experiments
4.1. Datasets Description
4.2. Data Processing
4.3. Evaluation Indexes
Model Training Details
5. Results and Discussion
5.1. Performance of the Searched Architecture
5.2. Performance of the Surrogate Predictors
5.3. Search Efficiency
5.4. Transferring from OPPORTUNITY to UniMiB-SHAR
6. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Beddiar, D.R.; Nini, B.; Sabokrou, M.; Hadid, A. Vision-based human activity recognition: A survey. Multimed. Tools Appl. 2020, 79, 30509–30555. [Google Scholar] [CrossRef]
- Pareek, P.; Thakkar, A. A survey on video-based human action recognition: Recent updates, datasets, challenges, and applications. Artif. Intell. Rev. 2021, 54, 2259–2322. [Google Scholar] [CrossRef]
- Braunagel, C.; Kasneci, E.; Stolzmann, W.; Rosenstiel, W. Driver-activity recognition in the context of conditionally autonomous driving. In Proceedings of the 2015 IEEE 18th International Conference on Intelligent Transportation Systems, Gran Canaria, Spain, 15–18 September 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1652–1657. [Google Scholar]
- Civitarese, G. Human Activity Recognition in Smart-Home Environments for Health-Care Applications. In Proceedings of the IEEE International Conference on Pervasive Computing and Communications Workshops 2019, Kyoto, Japan, 11–15 March 2019; p. 1. [Google Scholar]
- Sarngadharan, D.; Rajeesh, C.; Nandu, K. Human Agency, Social Structure and Forming of Health Consciousness and Perception. Eur. J. Mol. Clin. Med. 2021, 7, 5910–5916. [Google Scholar]
- Uddin, M.Z.; Hassan, M.M.; Alsanad, A.; Savaglio, C. A body sensor data fusion and deep recurrent neural network-based behavior recognition approach for robust healthcare. Inf. Fusion 2020, 55, 105–115. [Google Scholar] [CrossRef]
- Yang, C.; Xu, Y.; Shi, J.; Dai, B.; Zhou, B. Temporal pyramid network for action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 591–600. [Google Scholar]
- Li, Y.; Ji, B.; Shi, X.; Zhang, J.; Kang, B.; Wang, L. Tea: Temporal excitation and aggregation for action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 909–918. [Google Scholar]
- Feichtenhofer, C. X3d: Expanding architectures for efficient video recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 203–213. [Google Scholar]
- Kalfaoglu, M.E.; Kalkan, S.; Alatan, A.A. Late temporal modeling in 3d cnn architectures with bert for action recognition. In Proceedings of the European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2020; pp. 731–747. [Google Scholar]
- Mihanpour, A.; Rashti, M.J.; Alavi, S.E. Human action recognition in video using db-lstm and resnet. In Proceedings of the 2020 6th International Conference on Web Research (ICWR), Tehran, Iran, 22–23 April 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 133–138. [Google Scholar]
- Chen, K.; Zhang, D.; Yao, L.; Guo, B.; Yu, Z.; Liu, Y. Deep Learning for Sensor-based Human Activity Recognition: Overview, Challenges, and Opportunities. ACM Comput. Surv. (CSUR) 2021, 54, 1–40. [Google Scholar] [CrossRef]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
- He, X.; Zhao, K.; Chu, X. AutoML: A Survey of the State-of-the-Art. Knowl.-Based Syst. 2021, 212, 106622. [Google Scholar] [CrossRef]
- Liu, C.; Zoph, B.; Neumann, M.; Shlens, J.; Hua, W.; Li, L.J.; Fei-Fei, L.; Yuille, A.; Huang, J.; Murphy, K. Progressive neural architecture search. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 19–34. [Google Scholar]
- Li, Y.; Dong, M.; Wang, Y.; Xu, C. Neural architecture search in a proxy validation loss landscape. In Proceedings of the International Conference on Machine Learning, PMLR, Vienna, Austria, 12–18 July 2020; pp. 5853–5862. [Google Scholar]
- He, C.; Ye, H.; Shen, L.; Zhang, T. Milenas: Efficient neural architecture search via mixed-level reformulation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11993–12002. [Google Scholar]
- Li, Y.; Jin, X.; Mei, J.; Lian, X.; Yang, L.; Xie, C.; Yu, Q.; Zhou, Y.; Bai, S.; Yuille, A.L. Neural architecture search for lightweight non-local networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10297–10306. [Google Scholar]
- Zhang, T.; Lei, C.; Zhang, Z.; Meng, X.B.; Chen, C.P. AS-NAS: Adaptive Scalable Neural Architecture Search with Reinforced Evolutionary Algorithm for Deep Learning. IEEE Trans. Evol. Comput. 2021, 25, 830–841. [Google Scholar] [CrossRef]
- Lu, Z.; Deb, K.; Goodman, E.; Banzhaf, W.; Boddeti, V.N. Nsganetv2: Evolutionary multi-objective surrogate-assisted neural architecture search. In Proceedings of the European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2020; pp. 35–51. [Google Scholar]
- Cergibozan, A.; Tasan, A.S. Genetic algorithm based approaches to solve the order batching problem and a case study in a distribution center. J. Intell. Manuf. 2020, 33, 1–13. [Google Scholar] [CrossRef]
- Real, E.; Aggarwal, A.; Huang, Y.; Le, Q.V. Regularized evolution for image classifier architecture search. AAAI Conf. Artif. Intell. 2019, 33, 4780–4789. [Google Scholar] [CrossRef] [Green Version]
- Su, B.; Xie, N.; Yang, Y. Hybrid genetic algorithm based on bin packing strategy for the unrelated parallel workgroup scheduling problem. J. Intell. Manuf. 2021, 32, 957–969. [Google Scholar] [CrossRef]
- Liu, H.; Simonyan, K.; Yang, Y. DARTS: Differentiable Architecture Search. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
- Wang, L.; Xie, S.; Li, T.; Fonseca, R.; Tian, Y. Neural Architecture Search by Learning Action Space for Monte Carlo Tree Search 2019. Available online: https://openreview.net/pdf?id=SklR6aEtwH (accessed on 11 November 2022).
- White, C.; Neiswanger, W.; Savani, Y. BANANAS: Bayesian Optimization with Neural Architectures for Neural Architecture Search. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 2–9 February 2021; Volume 35, pp. 10293–10301. [Google Scholar]
- Guo, R.; Lin, C.; Li, C.; Tian, K.; Sun, M.; Sheng, L.; Yan, J. Powering one-shot topological nas with stabilized share-parameter proxy. In Proceedings of the European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2020; pp. 625–641. [Google Scholar]
- Liang, T.; Wang, Y.; Tang, Z.; Hu, G.; Ling, H. OPANAS: One-Shot Path Aggregation Network Architecture Search for Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10195–10203. [Google Scholar]
- Zhong, Z.; Yan, J.; Wu, W.; Shao, J.; Liu, C.L. Practical block-wise neural network architecture generation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2423–2432. [Google Scholar]
- Wang, D.; Li, M.; Gong, C.; Chandra, V. Attentivenas: Improving neural architecture search via attentive sampling. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 6418–6427. [Google Scholar]
- Yang, Z.; Wang, Y.; Chen, X.; Guo, J.; Zhang, W.; Xu, C.; Xu, C.; Tao, D.; Xu, C. Hournas: Extremely fast neural architecture search through an hourglass lens. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10896–10906. [Google Scholar]
- Zhang, X.; Hou, P.; Zhang, X.; Sun, J. Neural architecture search with random labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10907–10916. [Google Scholar]
- Bobick, A.; Davis, J. The recognition of human movement using temporal templates. IEEE Trans. Pattern Anal. Mach. Intell. 2001, 23, 257–267. [Google Scholar] [CrossRef]
- Laptev, I. On space-time interest points. Int. J. Comput. Vis. 2005, 64, 107–123. [Google Scholar] [CrossRef]
- Laptev, I.; Marszalek, M.; Schmid, C.; Rozenfeld, B. Learning realistic human actions from movies. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 1–8. [Google Scholar]
- Simonyan, K.; Zisserman, A. Two-stream convolutional networks for action recognition. In Proceedings of the Neural Information Processing Systems (NIPS), Montreal, QC, USA, 7–10 December 2015. [Google Scholar]
- Tang, J.; Zhang, J.; Yin, J. Temporal consistency two-stream CNN for human motion prediction. Neurocomputing 2022, 468, 245–256. [Google Scholar] [CrossRef]
- Ji, S.; Xu, W.; Yang, M.; Yu, K. 3D convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 221–231. [Google Scholar] [CrossRef] [Green Version]
- Al-Amin, M.; Qin, R.; Moniruzzaman, M.; Yin, Z.; Ming, C.L. An individualized system of skeletal data-based CNN classifiers for action recognition in manufacturing assembly. J. Intell. Manuf. 2021, 2, 1–7. [Google Scholar] [CrossRef]
- Guo, J.; Shi, M.; Zhu, X.; Huang, W.; He, Y.; Zhang, W.; Tang, Z. Improving human action recognition by jointly exploiting video and WiFi clues. Neurocomputing 2021, 458, 14–23. [Google Scholar] [CrossRef]
- Martindale, C.F.; Christlein, V.; Klumpp, P.; Eskofier, B.M. Wearables-based multi-task gait and activity segmentation using recurrent neural networks. Neurocomputing 2021, 432, 250–261. [Google Scholar] [CrossRef]
- Gautam, A.; Panwar, M.; Biswas, D.; Acharyya, A. MyoNet: A transfer-learning-based LRCN for lower limb movement recognition and knee joint angle prediction for remote monitoring of rehabilitation progress from sEMG. IEEE J. Transl. Eng. Health Med. 2020, 8, 1–10. [Google Scholar] [CrossRef]
- Li, X.; Luo, J.; Younes, R. ActivityGAN: Generative adversarial networks for data augmentation in sensor-based human activity recognition. In Proceedings of the Adjunct Proceedings of the 2020 ACM International Joint Conference on Pervasive and Ubiquitous Computing and Proceedings of the 2020 ACM International Symposium on Wearable Computers, Virtual Event, 12–17 September 2020; pp. 249–254. [Google Scholar]
- Zhang, J.; Wu, F.; Wei, B.; Zhang, Q.; Huang, H.; Shah, S.W.; Cheng, J. Data augmentation and dense-LSTM for human activity recognition using WiFi signal. IEEE Internet Things J. 2020, 8, 4628–4641. [Google Scholar] [CrossRef]
- Meng, F.; Liu, H.; Liang, Y.; Tu, J.; Liu, M. Sample fusion network: An end-to-end data augmentation network for skeleton-based human action recognition. IEEE Trans. Image Process. 2019, 28, 5281–5295. [Google Scholar] [CrossRef] [PubMed]
- Steven Eyobu, O.; Han, D.S. Feature Representation and Data Augmentation for Human Activity Classification Based on Wearable IMU Sensor Data Using a Deep LSTM Neural Network. Sensors 2018, 18, 2892. [Google Scholar] [CrossRef] [PubMed]
- Li, X.; Wang, Y.; Zhang, B.; Ma, J. PSDRNN: An efficient and effective HAR scheme based on feature extraction and deep learning. IEEE Trans. Ind. Inform. 2020, 16, 6703–6713. [Google Scholar] [CrossRef]
- Xiao, Z.; Xu, X.; Xing, H.; Song, F.; Wang, X.; Zhao, B. A federated learning system with enhanced feature extraction for human activity recognition. Knowl.-Based Syst. 2021, 229, 107338. [Google Scholar] [CrossRef]
- Ahmed Bhuiyan, R.; Ahmed, N.; Amiruzzaman, M.; Islam, M.R. A robust feature extraction model for human activity characterization using 3-axis accelerometer and gyroscope data. Sensors 2020, 20, 6990. [Google Scholar] [CrossRef] [PubMed]
- Garcia, N.C.; Bargal, S.A.; Ablavsky, V.; Morerio, P.; Murino, V.; Sclaroff, S. Distillation Multiple Choice Learning for Multimodal Action Recognition. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2021; pp. 2755–2764. [Google Scholar]
- Ji, X.; Zhao, Q.; Cheng, J.; Ma, C. Exploiting spatio-temporal representation for 3D human action recognition from depth map sequences. Knowl.-Based Syst. 2021, 4, 107040. [Google Scholar] [CrossRef]
- Herruzo, P.; Gruca, A.; Lliso, L.; Calbet, X.; Rípodas, P.; Hochreiter, S.; Kopp, M.; Kreil, D.P. High-resolution multi-channel weather forecasting–First insights on transfer learning from the Weather4cast Competitions 2021. In Proceedings of the 2021 IEEE International Conference on Big Data (Big Data), Orlando, FL, USA, 15–18 December 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 5750–5757. [Google Scholar]
- Wu, C.Y.; Zaheer, M.; Hu, H.; Manmatha, R.; Smola, A.J.; Krähenbühl, P. Compressed video action recognition. In Proceedings of the Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6026–6035. [Google Scholar]
- Zhang, Z.; Lv, Z.; Gan, C.; Zhu, Q. Human action recognition using convolutional LSTM and fully-connected LSTM with different attentions. Neurocomputing 2020, 410, 304–316. [Google Scholar] [CrossRef]
- Yu, T.; Li, X.; Cai, Y.; Sun, M.; Li, P. S2-mlp: Spatial-shift mlp architecture for vision. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 297–306. [Google Scholar]
- Lewis, R.J. An introduction to classification and regression tree (CART) analysis. In Proceedings of the Annual Meeting of the Society for Academic Emergency Medicine, San Francisco, CA, USA, 22–25 May 2000; Volume 14. [Google Scholar]
- Orr, M.J. Introduction to Radial Basis Function Networks. 1996. Available online: https://faculty.cc.gatech.edu/~isbell/tutorials/rbf-intro.pdf (accessed on 11 November 2022).
- Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef] [Green Version]
- Srinivas, N.; Deb, K. Muiltiobjective optimization using nondominated sorting in genetic algorithms. Evol. Comput. 1994, 2, 221–248. [Google Scholar] [CrossRef]
- Sagha, H.; Digumarti, S.T.; Millán, J.d.R.; Chavarriaga, R.; Calatroni, A.; Roggen, D.; Tröster, G. Benchmarking classification techniques using the Opportunity human activity dataset. In Proceedings of the 2011 IEEE International Conference on Systems, Man, and Cybernetics, Anchorage, AK, USA, 9–12 October 2011; IEEE: Piscataway, NJ, USA, 2011; pp. 36–40. [Google Scholar]
- Micucci, D.; Mobilio, M.; Napoletano, P. Unimib shar: A dataset for human activity recognition using acceleration data from smartphones. Appl. Sci. 2017, 7, 1101. [Google Scholar] [CrossRef] [Green Version]
- Jhuang, H.; Gall, J.; Zuffi, S.; Schmid, C.; Black, M.J. Towards understanding action recognition. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–3 December 2013; pp. 3192–3199. [Google Scholar]
- Huang, C. Event-based action recognition using timestamp image encoding network. arXiv 2020, arXiv:2009.13049. [Google Scholar]
- Li, F.; Shirahama, K.; Nisar, M.A.; Köping, L.; Grzegorzek, M. Comparison of feature learning methods for human activity recognition using wearable sensors. Sensors 2018, 18, 679. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Zeng, M.; Nguyen, L.T.; Yu, B.; Mengshoel, O.J.; Zhu, J.; Wu, P.; Zhang, J. Convolutional neural networks for human activity recognition using mobile sensors. In Proceedings of the 6th International Conference on Mobile Computing, Applications and Services, Austin, TX, USA, 6–7 November 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 197–205. [Google Scholar]
- Ordóñez, F.J.; Roggen, D. Deep convolutional and lstm recurrent neural networks for multimodal wearable activity recognition. Sensors 2016, 16, 115. [Google Scholar] [CrossRef] [PubMed]
- Yang, J.; Nguyen, M.N.; San, P.P.; Li, X.L.; Krishnaswamy, S. Deep convolutional neural networks on multichannel time series for human activity recognition. In Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July 2015. [Google Scholar]
- Hammerla, N.Y.; Halloran, S.; Ploetz, T. Deep, Convolutional, and Recurrent Models for Human Activity Recognition using Wearables. J. Sci. Comput. 2016, 61, 454–476. [Google Scholar]
- Yang, Z.; Raymond, O.I.; Zhang, C.; Wan, Y.; Long, J. DFTerNet: Towards 2-bit dynamic fusion networks for accurate human activity recognition. IEEE Access 2018, 6, 56750–56764. [Google Scholar] [CrossRef]
- Brandenburg, F.J.; Gleißner, A.; Hofmeier, A. Comparing and aggregating partial orders with kendall tau distances. Discret. Math. Algorithms Appl. 2013, 5, 1360003. [Google Scholar] [CrossRef]
Ref. | Method | Description | Type |
---|---|---|---|
[26] 2018 | DARTS | Continuous relaxation in architecture representation | |
[30] 2021 | OPANAS | An efficient one-shot search method | |
[21] 2021 | AS-NAS | Solve the non-convexity problem in NAS. | |
[24] 2019 | AmoebaNet-A | Age property to favor the younger genotypes | |
[31] 2018 | BlockQNN | Use Q-Learning to build networks | |
[17] 2018 | PNAS | Use SMBO strategy to search | |
[27] 2019 | MCTS | Learn action space by MCTS | |
[28] 2019 | BANANAS | Use bayesian optimization | |
[32] 2020 | AttentiveNAS | Focuses on sampling the networks | |
[33] 2020 | HourNAS | Make the vital blocks the priority | |
[17] 2016 | NASnet | Use RL-based model to search | |
[22] 2020 | Nsganetv2 | Use surrogate model for efficiency | |
[29] 2020 | ST-NAS | Stabilized share-parameter proxy | |
[34] 2021 | RLNAS | Ease-of-convergence hypothesis |
Ref. | Method | Description | Type |
---|---|---|---|
[44] 2020 | LRCN | Transfer-learning-based approach | |
[45] 2020 | ActivityGAN | GAN-based data generation architecture | |
[46] 2020 | CSI | Eight channel state information transformation | |
[47] 2019 | SFGM | Sample fusion-based generation model | |
[48] 2018 | FS-LSTM | An ensemble of data augmentations in feature space | |
[48] 2018 | FS-LSTM | A spectrogram-based feature extraction approach | |
[49] 2020 | PSDRNN | An explicit feature extraction | |
[50] 2021 | PEN-based | Perceptive extraction net (PEN) feature extractor | |
[51] 2020 | EPS-LDA | An efficient and reduce dimension feature extractor | |
[56] 2020 | STDAN | Enrich the initial level of video representation | |
[52] 2021 | DMCL | A distillation multiple choice learning framework | |
[53] 2021 | DOGV | capture the cues between spatial appearance and temporal motion | |
[54] 2021 | Swin-B | An inductive bias of locality Transformers | |
[55] 2020 | PAN | A persistence of Appearance (PA)-based model |
Operations | Kernel Size |
---|---|
Skip connection | - |
Average pooling | (3,1),(1,3),(3,3) |
Max pooling | (3,1),(1,3),(3,3) |
Convolution | (1,1),(3,3),(5,5),(7,7),(3,1) (5,1),(7,1),(11,1) |
Dilated convolution | (3,3), (5,5) |
Positive | Negative | |
---|---|---|
Positive | True positive (TP) | False positive (FP) |
Negative | False negative (FN) | True negative (TN) |
Type | Model Name | OPPORTUNITY | UniMiB-SHAR | ||||
---|---|---|---|---|---|---|---|
Basic model | MLP | 91.11 | 68.17 | 90.86 | 71.62 | 59.97 | 70.81 |
LSTM | 91.29 | 69.71 | 91.16 | 71.47 | 59.32 | 70.82 | |
CNN | 90.58 | 65.26 | 90.19 | 74.97 | 64.65 | 74.29 | |
AE | 87.80 | 55.62 | 87.60 | 65.67 | 55.04 | 64.84 | |
Previous model | CNN-based [55] | 76.83 | - | - | - | - | - |
CNN + LSTM [67] | 78.90 | 70.40 | 91.70 | - | - | - | |
CNN-based [68] | 89.60 | - | 85.10 | - | - | - | |
CNN-based [69] | 90.58 | 65.26 | 90.19 | 74.66 | 64.65 | 74.29 | |
CNN + RNN [66] | - | - | 92.07 | - | - | - | |
CNN + NAS [70] | - | - | - | 72.8 | - | - | |
CNN + LSTM [71] | - | - | - | - | - | 73.19 | |
LSTM-based [66] | - | - | - | - | - | 69.24 | |
The proposed | |||||||
NAS model | HAR-NAS | 91.41 | 68.87 | 92.09 | 75.52 | 64.47 | 76.10 |
Model Name | Parameters Setting |
---|---|
MLP | Neurons numbers: 2000, 2000, 2000 |
LSTM | LSTM cells: 64, 64 Output dimensions: 600, 600 |
CNN | Convolution kernel sizes: (11, 1) (10, 1) (6, 1) Convolution siding strides: (1, 1) (1, 1) (1, 1) Pooling kernel sizes: (2, 1) (3, 1) (1, 1) |
AE | Neurons in encoder and decoder layer (5000) |
Dataset | Model | F1-Score | Trained Model | Speed up |
---|---|---|---|---|
Opportunity | Original NAS | 92.17 | 200 | 1× |
Opportunity | The proposed method | 92.09 | 120 | 1.67× |
UniMiB SHAR | Original NAS | 75.64 | 100 | 1× |
UniMiB SHAR | The proposed method | 76.12 | 24 | 4.17× |
Model | Weight F1-Score | FLOPs | Number of Parameters | Number of Training Model |
---|---|---|---|---|
NAS | 75.64 | 52.52 | 0.233 | 100 |
Nas with Surrogate Model | 76.12 | 20.66 | 0.140 | 24 |
Nas with old Surrogate Model (online) | 75.37 | 19.40 | 0.122 | 88 |
Nas with old Surrogate Model (offline) | 73.51 | 23.77 | 0.183 | 16 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wang, X.; He, M.; Yang, L.; Wang, H.; Zhong, Y. Human Activity Recognition Based on an Efficient Neural Architecture Search Framework Using Evolutionary Multi-Objective Surrogate-Assisted Algorithms. Electronics 2023, 12, 50. https://doi.org/10.3390/electronics12010050
Wang X, He M, Yang L, Wang H, Zhong Y. Human Activity Recognition Based on an Efficient Neural Architecture Search Framework Using Evolutionary Multi-Objective Surrogate-Assisted Algorithms. Electronics. 2023; 12(1):50. https://doi.org/10.3390/electronics12010050
Chicago/Turabian StyleWang, Xiaojuan, Mingshu He, Liu Yang, Hui Wang, and Yun Zhong. 2023. "Human Activity Recognition Based on an Efficient Neural Architecture Search Framework Using Evolutionary Multi-Objective Surrogate-Assisted Algorithms" Electronics 12, no. 1: 50. https://doi.org/10.3390/electronics12010050
APA StyleWang, X., He, M., Yang, L., Wang, H., & Zhong, Y. (2023). Human Activity Recognition Based on an Efficient Neural Architecture Search Framework Using Evolutionary Multi-Objective Surrogate-Assisted Algorithms. Electronics, 12(1), 50. https://doi.org/10.3390/electronics12010050