TricP: A Novel Approach for Human Activity Recognition Using Tricky Predator Optimization Based on Inception and LSTM
Abstract
1. Introduction
- To introduce an efficient deep learning-based HAR technique for realistic, multi-view surveillance video sequences. Pretrained models are adopted for deep feature extraction to reduce training cost and improve generalization.
- To propose a novel optimization algorithm, Tricky Predator (TricP), inspired by the poaching behavior of predators and the social dynamics of Latrans, to enhance exploration and exploitation during search.
- To perform HAR using a Tricky Predator-based Incept-LSTM classifier, where TricP acts as an outer-loop optimizer to fine-tune a restricted subset of parameters, avoiding prohibitive full-network search while improving recognition performance.
2. Related Works
3. Materials and Methods
3.1. Proposed Methodology for Human Activity Recognition Using an Optimized Incept-LSTM Classifier
3.2. Data Set
3.3. Extraction of Distinctive Video Frames from the Dataset
3.4. Feature Extraction
3.4.1. Histogram of Optical Flow (HOOF)
3.4.2. ResNet-101
3.4.3. Feature Concatenation
3.5. Human Action Recognition System
Inception LSTM
4. TricP: Proposed Tricky Predator—An Optimization Algorithm
4.1. Proposed Tricky Predator Optimization Algorithm
- Predator poaching behavior: Predators live in packs and hunt wild and domestic animals using stealthy crawling behavior and intelligent movement patterns. A dominant predator couple leads and protects the territory. Offspring may inherit the parents’ territory or migrate to establish a new territory upon maturity.
- Latrans social behavior: Latrans (coyotes) are social predators that follow dominance rules and cultural tendencies [28]. These social rules support herd maintenance and promote a balance between exploration and exploitation.
- A predator may leave its territory to form a new herd or be removed (e.g., killed by a poacher).
- Reproduction is modeled by replacing weak predators with new offspring at each iteration, which helps retain diversity and refine the best solution found so far.
4.2. Mathematical Modeling
4.2.1. Solution Encoding and Search-Space Dimension
4.2.2. Population Representation and Objective Function
4.2.3. Fitness Evaluation
4.3. Exploration Phase
4.4. Exploitation Phase
4.5. Latrans Social Update and Combined Rule
4.6. Breeding and Leaving (Population Maintenance)
4.7. Termination and Pseudo-Code
| Algorithm 1. Pseudo-code for the proposed Tricky Predator Optimization algorithm. |
Begin
a. For each predator , evaluate fitness using (26). b. Sort predators by fitness; select the best predator . c. For each predator, update its position according to the exploration and exploitation rules (30)–(36). d. Sort predators by updated fitness. e. Remove the worst predators (simulate hunting). f. Generate and include new predators using (40) (and rule in (39)). g. Update iteration counter: .
End |
5. Experimentation and Results
5.1. Experimental Setup
5.2. Implementation Details
5.3. Experimental Parameters
| Method | Training Time/Epoch | Total Epochs | Relative Training Time |
|---|---|---|---|
| Stochastic CNN | ~2.4 min | ~40 epochs | Baseline (1×) |
| DeepCNN-BiLSTM | ~3.1 min | ~35 epochs | 1.1× |
| isplInception | ~3.3 min | ~31 epochs | 1.2× |
| Tricky Predator-based Incept-LSTM | ~5.0 min | ~30 epochs | 2.0× |
6. Analysis
6.1. Interpretation Based on Training Data Percentage
6.2. Interpretation Based on K-Fold
6.3. Interpretation Based on ROC
6.4. Comparative Discussion
- Specificity: 8.03%, 6.47%, 3.54%, and 1.45%, respectively.
- Sensitivity: 21.19%, 6.67%, 3.93%, and 1.92%, respectively.
- Accuracy: 8.40%, 5.15%, 3.22%, and 1.61%, respectively.
7. Conclusions and Future Scope
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Mekruksavanich, S.; Jitpattanakul, A. Biometric user identification based on human activity recognition using wearable sensors: An experiment using deep learning models. Electronics 2021, 10, 308. [Google Scholar] [CrossRef]
- Ronald, M.; Poulose, A.; Han, D.S. iSPLInception: An inception-ResNet deep learning architecture for human activity recognition. IEEE Access 2021, 9, 68985–69001. [Google Scholar] [CrossRef]
- Garcia, K.D.; de Sá, C.R.; Poel, M.; Carvalho, T.; Mendes-Moreira, J.; Cardoso, J.M.; de Carvalho, A.C.; Kok, J.N. An ensemble of autonomous auto-encoders for human activity recognition. Neurocomputing 2021, 439, 271–280. [Google Scholar] [CrossRef]
- Anagnostis, A.; Benos, L.; Tsaopoulos, D.; Tagarakis, A.; Tsolakis, N.; Bochtis, D. Human activity recognition through recurrent neural networks for human–robot interaction in agriculture. Appl. Sci. 2021, 11, 2188. [Google Scholar] [CrossRef]
- Tasnim, N.; Islam, M.K.; Baek, J.H. Deep learning based human activity recognition using spatio-temporal image formation of skeleton joints. Appl. Sci. 2021, 11, 2675. [Google Scholar] [CrossRef]
- Alemayoh, T.T.; Lee, J.H.; Okamoto, S. New sensor data structuring for deeper feature extraction in human activity recognition. Sensors 2021, 21, 2814. [Google Scholar] [CrossRef]
- Zhang, L.; Lim, C.P.; Yu, Y. Intelligent human action recognition using an ensemble model of evolving deep networks with swarm-based optimization. Knowl. Based Syst. 2021, 220, 106918. [Google Scholar] [CrossRef]
- Russell, B.; McDaid, A.; Toscano, W.; Hume, P. Moving the Lab into the Mountains: A Pilot Study of Human Activity Recognition in Unstructured Environments. Sensors 2021, 21, 654. [Google Scholar] [CrossRef]
- Xu, Y.; Qiu, T.T. Human activity recognition and embedded application based on convolutional neural network. J. Artif. Intell. Technol. 2021, 1, 51–60. [Google Scholar] [CrossRef]
- Ma, C.Y.; Chen, M.H.; Kira, Z.; AlRegib, G. TS-LSTM and temporal-inception: Exploiting spatiotemporal dynamics for activity recognition. Signal Process. Image Commun. 2019, 71, 76–87. [Google Scholar] [CrossRef]
- Jandhyam, L.A.; Rengaswamy, R.; Satyala, N. An optimized Deep LSTM model for human action recognition. Rev. D’intelligence Artif. 2024, 38, 11–23. [Google Scholar] [CrossRef]
- Weng, Z.; Li, W.; Jin, Z. Human activity prediction using saliency-aware motion enhancement and weighted LSTM network. EURASIP J. Image Video Process. 2021, 2021, 3. [Google Scholar] [CrossRef]
- Aghaei, A.; Nazari, A.; Moghaddam, M.E. Sparse deep LSTMs with convolutional attention for human action recognition. SN Comput. Sci. 2021, 2, 151. [Google Scholar] [CrossRef]
- Saoudi, E.M.; Jaafari, J.; Andaloussi, S.J. Advancing human action recognition: A hybrid approach using attention-based LSTM and 3D CNN. Sci. Afr. 2023, 21, e01796. [Google Scholar] [CrossRef]
- Pandey, N.N.; Muppalaneni, N.B. Temporal and spatial feature based approaches in drowsiness detection using deep learning technique. J. Real-Time Image Process. 2021, 18, 2287–2299. [Google Scholar] [CrossRef]
- Guo, H.; Chen, J. Dynamic facial expression recognition based on ResNet and LSTM. IOP Conf. Ser. Mater. Sci. Eng. 2020, 790, 012145. [Google Scholar]
- Ahmad, W.; Kazmi, B.M.; Ali, H. Human activity recognition using multi-head CNN followed by LSTM. In Proceedings of the 2019 15th International Conference on Emerging Technologies (ICET); IEEE: Washington, DC, USA, 2019; pp. 1–6. [Google Scholar]
- Alom, M.Z.; Hasan, M.; Yakopcic, C.; Taha, T.M.; Asari, V.K. Improved inception-residual convolutional neural network for object recognition. Neural Comput. Appl. 2020, 32, 279–293. [Google Scholar] [CrossRef]
- Xu, C.; Chai, D.; He, J.; Zhang, X.; Duan, S. InnoHAR: A deep neural network for complex human activity recognition. IEEE Access 2019, 7, 9893–9902. [Google Scholar] [CrossRef]
- Xia, K.; Huang, J.; Wang, H. LSTM-CNN architecture for human activity recognition. IEEE Access 2020, 8, 56855–56866. [Google Scholar] [CrossRef]
- Mustafa, T.; Dhavale, S.; Kuber, M.M. Performance Analysis of Inception-v2 and Yolov3-Based Human Activity Recognition in Videos. SN Comput. Sci. 2020, 1, 138. [Google Scholar] [CrossRef][Green Version]
- Dua, N.; Singh, S.N.; Semwal, V.B. Multi-input CNN-GRU based human activity recognition using wearable sensors. Computing 2021, 103, 1461–1478. [Google Scholar] [CrossRef]
- Nafea, O.; Abdul, W.; Muhammad, G.; Alsulaiman, M. Sensor-based human activity recognition with spatio-temporal deep learning. Sensors 2021, 21, 2141. [Google Scholar] [CrossRef]
- UCF Crime Dataset. Available online: https://www.kaggle.com/datasets/odins0n/ucf-crime-dataset (accessed on 15 July 2024).
- Prabha, B.; Shanker, N.R.; Priya, M.; Ganesh, E. Human Anomalous Activity Detection: Shape and Motion Approach in Crowded Scenes. J. Phys. Conf. Ser. 2021, 1921, 012074. [Google Scholar] [CrossRef]
- Hnoohom, N.; Maitrichit, N.; Chotivatunyu, P.; Sornlertlamvanich, V.; Mekruksavanich, S.; Jitpattanakul, A. Blister Package Classification Using ResNet-101 for Identification of Medication. In Proceedings of the 2021 25th International Computer Science and Engineering Conference (ICSEC); IEEE: Washington, DC, USA, 2021; pp. 406–410. [Google Scholar]
- Hosseini, M.; Maida, A.S.; Hosseini, M.; Raju, G. Inception-inspired lstm for next-frame video prediction. arXiv 2019, arXiv:1909.05622. [Google Scholar]
- Pierezan, J.; Coelho, L.D.S. Coyote optimization algorithm: A new metaheuristic for global optimization problems. In Proceedings of the 2018 IEEE Congress on Evolutionary Computation (CEC); IEEE: Washington, DC, USA, 2018; pp. 1–8. [Google Scholar]
- Połap, D.; Woźniak, M. Red fox optimization algorithm. Expert Syst. Appl. 2021, 166, 114107. [Google Scholar] [CrossRef]
- Binu, D.; Kariyappa, B.S. Rider-deep-LSTM network for hybrid distance score-based fault prediction in analog circuits. IEEE Trans. Ind. Electron. 2020, 68, 10097–10106. [Google Scholar] [CrossRef]











| Method | Feature Pipeline | Optimization Mode | Training Time/Epoch (min) | Total Epochs | Total Training Time (min) | Relative Cost |
|---|---|---|---|---|---|---|
| Adam (Baseline) | End-to-End (Raw Frames) | Gradient-Based | ~2.4 | 40 | ~96 | 1.00× |
| Adam (Baseline) | Cached Features (HOOF + ResNet-101) | Gradient-Based | ~1.6 | 40 | ~64 | 0.67× |
| TricP (Proposed) | Cached Features (HOOF + ResNet-101) | Hybrid (Population + Gradient) | ~5.0 | 30 | ~150 | 1.56× (vs. E2E)/2.34× (vs. Cached Adam) |
| Methods/Metrics | EKVN | Stochastic-CNN | iSPLInception | DeepCNN-BiLSTM | Tricky Predator-Based Incept-LSTM |
|---|---|---|---|---|---|
| Specificity based on Training | |||||
| 40% | 82.94 | 86.44 | 88.48 | 89.87 | 91.80 |
| 50% | 84.98 | 87.98 | 89.54 | 90.92 | 93.08 |
| 60% | 86.42 | 89.03 | 91.51 | 92.90 | 94.29 |
| 70% | 87.38 | 90.11 | 92.70 | 94.17 | 95.56 |
| 80% | 89.06 | 90.57 | 93.41 | 95.44 | 96.84 |
| Sensitivity Based on Training | |||||
| 40% | 57.40 | 78.80 | 80.08 | 85.68 | 87.31 |
| 50% | 61.84 | 79.47 | 81.49 | 87.33 | 89.12 |
| 60% | 70.09 | 85.12 | 86.81 | 88.04 | 89.53 |
| 70% | 71.85 | 85.57 | 87.69 | 89.22 | 90.45 |
| 80% | 72.63 | 86.01 | 88.54 | 90.39 | 92.16 |
| Accuracy Based on Training | |||||
| 40% | 78.25 | 83.48 | 84.96 | 87.57 | 89.13 |
| 50% | 81.07 | 83.78 | 85.63 | 89.51 | 90.89 |
| 60% | 82.81 | 86.82 | 88.32 | 90.36 | 91.69 |
| 70% | 84.09 | 87.51 | 89.30 | 90.96 | 92.46 |
| 80% | 85.76 | 88.80 | 90.61 | 92.11 | 93.62 |
| Methods/Metrics | EKVN | Stochastic-CNN | iSPL Inception | DeepCNN-BiLSTM | Incept-LSTM (Adam) | Tricky Predator-Based Incept-LSTM |
|---|---|---|---|---|---|---|
| Specificity based on K-Fold | ||||||
| 5 | 81.73 | 85.39 | 87.19 | 89.59 | 90.10 | 91.39 |
| 6 | 82.65 | 85.98 | 87.52 | 89.74 | 90.36 | 91.45 |
| 7 | 85.15 | 86.57 | 88.07 | 89.89 | 90.32 | 91.75 |
| 8 | 85.65 | 87.28 | 88.73 | 91.06 | 91.45 | 92.74 |
| 9 | 85.73 | 87.84 | 89.26 | 91.20 | 92.11 | 93.33 |
| 10 | 86.39 | 88.43 | 89.94 | 91.82 | 92.38 | 93.70 |
| Sensitivity based on K-Fold | ||||||
| 5 | 66.15 | 82.68 | 84.54 | 87.30 | 87.85 | 88.64 |
| 6 | 68.28 | 83.28 | 85.18 | 87.99 | 88.40 | 89.25 |
| 7 | 69.01 | 84.59 | 86.32 | 88.50 | 90.12 | 89.85 |
| 8 | 69.18 | 84.85 | 86.79 | 88.96 | 90.78 | 90.37 |
| 9 | 70.80 | 85.32 | 86.97 | 89.71 | 92.02 | 91.19 |
| 10 | 72.85 | 85.80 | 87.64 | 90.23 | 90.78 | 91.64 |
| Accuracy based on K-Fold | ||||||
| 5 | 74.65 | 84.39 | 86.27 | 88.68 | 89.12 | 90.26 |
| 6 | 76.22 | 85.34 | 87.04 | 89.32 | 89.74 | 90.84 |
| 7 | 81.76 | 86.23 | 87.93 | 89.84 | 90.45 | 91.42 |
| 8 | 82.49 | 86.49 | 88.32 | 90.42 | 91.10 | 91.94 |
| 9 | 82.57 | 87.13 | 88.96 | 91.00 | 91.82 | 92.59 |
| 10 | 82.69 | 87.77 | 89.54 | 91.51 | 92.26 | 93.17 |
| Methods/Metrics | EKVN | Stochastic-CNN | iSPLInception | DeepCNN-BiLSTM | Tricky Predator-Based Incept-LSTM |
|---|---|---|---|---|---|
| Interpretation based on Training | |||||
| Specificity | 89.06 | 90.57 | 93.41 | 95.44 | 96.84 |
| Sensitivity | 72.63 | 86.01 | 88.54 | 90.39 | 92.16 |
| Accuracy | 85.76 | 88.80 | 90.61 | 92.11 | 93.62 |
| Interpretation based on K-Fold | |||||
| Specificity | 86.39 | 88.43 | 89.94 | 91.82 | 93.70 |
| Sensitivity | 72.85 | 85.80 | 87.64 | 90.23 | 91.64 |
| Accuracy | 82.69 | 87.77 | 89.54 | 91.51 | 93.17 |
| Method | Run 1 (%) | Run 2 (%) | Run 3 (%) | Run 4 (%) | Run 5 (%) | Mean ± Std (%) | p-Value |
|---|---|---|---|---|---|---|---|
| Adam (Baseline) | 92.18 | 92.34 | 92.09 | 92.41 | 92.28 | 92.26 ± 0.13 | — |
| TricP (Proposed) | 93.05 | 93.22 | 93.11 | 93.26 | 93.21 | 93.17 ± 0.09 | 0.0036 |
| Method | Run | Specificity (%) | Sensitivity (%) | Accuracy (%) |
|---|---|---|---|---|
| Adam | 1 | 90.74 | 95.91 | 92.18 |
| Adam | 2 | 90.88 | 96.03 | 92.34 |
| Adam | 3 | 90.69 | 95.82 | 92.09 |
| Adam | 4 | 90.93 | 96.11 | 92.41 |
| Adam | 5 | 90.81 | 95.97 | 92.28 |
| TricP | 1 | 91.82 | 96.41 | 93.05 |
| TricP | 2 | 92.04 | 96.66 | 93.22 |
| TricP | 3 | 91.91 | 96.53 | 93.11 |
| TricP | 4 | 92.18 | 96.78 | 93.26 |
| TricP | 5 | 92.09 | 96.71 | 93.21 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Girdhar, P.; Al-Saidi, M.; Johri, P.; Virmani, D.; Taha, H.; Hassen, O.A. TricP: A Novel Approach for Human Activity Recognition Using Tricky Predator Optimization Based on Inception and LSTM. Telecom 2026, 7, 32. https://doi.org/10.3390/telecom7020032
Girdhar P, Al-Saidi M, Johri P, Virmani D, Taha H, Hassen OA. TricP: A Novel Approach for Human Activity Recognition Using Tricky Predator Optimization Based on Inception and LSTM. Telecom. 2026; 7(2):32. https://doi.org/10.3390/telecom7020032
Chicago/Turabian StyleGirdhar, Palak, Muslem Al-Saidi, Prashant Johri, Deepali Virmani, Hussein Taha, and Oday Ali Hassen. 2026. "TricP: A Novel Approach for Human Activity Recognition Using Tricky Predator Optimization Based on Inception and LSTM" Telecom 7, no. 2: 32. https://doi.org/10.3390/telecom7020032
APA StyleGirdhar, P., Al-Saidi, M., Johri, P., Virmani, D., Taha, H., & Hassen, O. A. (2026). TricP: A Novel Approach for Human Activity Recognition Using Tricky Predator Optimization Based on Inception and LSTM. Telecom, 7(2), 32. https://doi.org/10.3390/telecom7020032

