Optimizing IoMT Security: Performance Trade-Offs Between Neural Network Architectural Design, Dimensionality Reduction, and Class Imbalance Handling
Abstract
1. Introduction
- Contributions. This work advances the development of secure IoMT frameworks through improved intrusion detection techniques based on sophisticated neural network approaches. The key contributions of this research include:
- 1.
- A systematic evaluation of multiple advanced ANN architectures for intrusion detection in IoMT environments, including standard feedforward networks, dual-branch models with addition and concatenation operations, and networks incorporating shortcut connections.
- 2.
- Comprehensive assessment of autoencoder preprocessing for dimensionality reduction in intrusion detection, revealing critical trade-offs between feature compression and detection performance.
- 3.
- Comparative analysis of three class imbalance mitigation strategies (SMOTE, weighted loss functions, and hybrid sampling) across different neural architectures, identifying optimal combinations for effective attack detection.
- Outline. The remainder of this paper is structured as follows. Section 2 discusses related work in intrusion detection. Section 3 presents our comprehensive research methodology. Section 4 details the WUSTL-EHMS-2020 dataset selection, preprocessing steps, and feature engineering techniques employed. Section 5 introduces our proposed neural network architectures, highlighting their structural elements and theoretical advantages. Section 6 provides a detailed analysis of experimental findings, comparing performance across architectures and preprocessing approaches, with statistical validation of key results. Finally, Section 7 synthesizes our contributions, discusses limitations, and outlines promising directions for future research in IoMT intrusion detection.
2. Related Work
2.1. Traditional Machine Learning Approaches
2.2. IoMT-Specific Approaches
2.3. Research Gaps and Opportunities
- Most studies focus on either traditional machine learning or basic neural network structures, with limited exploration of advanced neural network architectures specifically designed for healthcare intrusion detection.
- While class imbalance is acknowledged as a challenge, comprehensive comparisons of different balancing techniques and their impact on various neural network architectures are scarce.
- The interaction between dimensionality reduction techniques and different neural network designs remains underexplored, particularly in healthcare-specific contexts and with advanced techniques like Autoencoders.
- The impact of channel number optimization on neural network performance for intrusion detection has received insufficient attention.
- RQ1: How do different neural network architectures compare in their effectiveness for IoMT intrusion detection relative to traditional machine learning models?
- RQ2: What is the impact of various class balancing techniques on the detection accuracy and robustness of neural network–based intrusion detection systems in healthcare environments?
- RQ3: How do Autoencoders for dimensionality reduction affect detection performance and computational efficiency across neural network architectures?
3. Methodology
- 1.
- Dataset Selection and Preparation: We utilize the WUSTL-EHMS-2020 dataset specifically designed for IoMT environments, performing initial cleaning and feature standardization.
- 2.
- Feature Processing: We implement two parallel processing paths:
- Direct Feature Processing: Original features are standardized and used directly for model training.
- Autoencoder Preprocessing: Features undergo dimensionality reduction through an autoencoder network before being fed to classification models.
- 3.
- Class Imbalance Handling: We implement and compare three distinct strategies:
- Synthetic Minority Over-sampling Technique (SMOTE)
- Weighted loss function approach
- Hybrid over-under sampling method
- 4.
- Neural Network Architecture Design: We implement five distinct ANN architectures:
- Standard ANN (baseline)
- Enhanced Channel ANN (ANN_v1)
- Dual-Branch Addition ANN (ANN_v2)
- Dual-Branch Concatenation ANN (ANN_v3)
- Shortcut Connection ANN (ANN_v4)
- 5.
- Model Training and Validation: Each architecture is trained with consistent hyperparameters across multiple class balancing configurations, using evaluation at regular intervals to assess performance.
- 6.
- Performance Evaluation: Models are evaluated using multiple metrics (Area Under the Curve (AUC), Accuracy, Precision, Recall, and F1-score) to provide a comprehensive assessment of their detection capabilities.
4. Dataset Selection and Preprocessing
- Spoofing attacks: Intercept communications between gateway and server, potentially exposing confidential patient information.
- Data injection attackss: Alter packets in transit, compromising data integrity.
4.1. Data Preprocessing
4.2. Dimensionality Reduction
- Encoder: Compresses the 34-dimensional input features to a 16-dimensional bottleneck representation
- Decoder: Reconstructs the original 34-dimensional features from the bottleneck representation
4.3. Class Imbalance Strategies
4.3.1. Synthetic Minority Over-Sampling Technique (SMOTE)
4.3.2. Hybrid Over-Under Sampling
4.3.3. Weighted Cross-Entropy Loss Function
5. Neural Network Architectures
5.1. Standard and Enhanced Channel Neural Network Architectures
5.2. Dual-Branch Architectures (ANN_v2 and ANN_v3)
- Data Splitting: The 34-dimensional input is divided into network metrics (first 26 features) and biometric parameters (remaining 8 features)
- Specialized Processing: Each feature type is processed through dedicated network branches
- Different Fusion Mechanisms: The two models differ in how they combine branch outputs
5.3. Shortcut Connection ANN (ANN_v4)
6. Testing and Discussion
6.1. Evaluation Metrics
- Area Under the ROC Curve (AUC): Measures the model’s discrimination capability across all possible classification thresholds. AUC represents the probability that the classifier will rank a randomly chosen positive instance higher than a randomly chosen negative instance. Mathematically:where TPR is the true positive rate and FPR is the false positive rate at threshold t. AUC values range from 0.5 (random classification) to 1.0 (perfect classification). This metric is particularly valuable for imbalanced datasets as it is insensitive to class distribution.
- Accuracy (ACC): The proportion of correctly classified instances among all instances:While intuitive, accuracy can be misleading in imbalanced datasets, as high accuracy can be achieved by simply classifying all instances as the majority class.
- Precision (PR): The proportion of true positive predictions among all positive predictions:High precision indicates a low false positive rate, which is particularly important in intrusion detection systems where false alarms can lead to alert fatigue and reduced trust in the system.
- Recall (RC): Also known as sensitivity or true positive rate, recall measures the proportion of actual positives that are correctly identified:High recall indicates that the model successfully captures most attack instances, which is critical in security applications where missing an attack (false negative) can have severe consequences.
- F1-score (F1): The harmonic mean of precision and recall, providing a balance between these two potentially competing metrics:F1-score ranges from 0 to 1, with higher values indicating better performance. This metric is particularly useful when seeking a balance between precision and recall.
6.2. Training Dynamics
6.3. Testing Results
6.3.1. Impact of Neural Network Architecture
- Standard vs. Enhanced ANNs: The enhancement through increased channel numbers (ANN_v1) consistently improves performance, confirming that greater parametrization enables better feature learning for this task.
- Dual-Branch Architectures: The dual-branch models (ANN_v2 and ANN_v3) consistently achieve the highest performance across all balancing methods. The addition-based combination (ANN_v2) generally outperforms the concatenation approach (ANN_v3), suggesting that the summation of features from parallel branches provides more effective feature integration for intrusion detection.
- Shortcut Connections: The ANN_v4 model with shortcut connections shows comparable performance to ANN_v1, indicating that for this particular task and dataset size, shortcut connections do not provide substantial additional benefits over simply increasing channel numbers.
6.3.2. Effectiveness of Class Balancing Methods
6.3.3. Impact of Dimensionality Reduction
6.3.4. Robustness Against Overfitting
6.4. Computational Complexity
6.5. Comparative Analysis with Previous Work
- Performance Improvement: The suggested dual-branch ANN architecture with addition operations (ANN_v2) combined with a weighted loss function achieves an AUC of 0.8786 and an F1-score of 0.8716, representing relative improvements of 12.8% and 20.7% respectively over the best ELM model from [13].
- Architectural Sophistication: Moving beyond the single hidden layer constraint of ELM, our current work explores multi-layer architectures with various connectivity patterns, demonstrating that architectural design choices significantly impact detection performance.
- Dimensionality Reduction Analysis: While [13] work focused on direct classification of input features, this study provides critical insights into the trade-offs associated with autoencoder preprocessing, revealing that the information loss during dimensionality reduction compromises recall—a crucial metric for security applications.
7. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Islam, S.R.; Kwak, D.; Kabir, M.H.; Hossain, M.; Kwak, K.S. The Internet of Things for health care: A comprehensive survey. IEEE Access 2015, 3, 678–708. [Google Scholar] [CrossRef]
- Grand View Research. Internet of Medical Things Market Size, Share & Trends Analysis Report by Component (Hardware, Software, Services), by Deployment (On-Premise, Cloud), by Application (Telemedicine, Clinical Operations, Medication Management), by End Use (Hospitals, Clinics, Research Institutes), by Region, and Segment Forecasts, 2025–2030; Technical Report; Report ID: GVR-4-68040-134-8; Grand View Research, Inc.: San Francisco, CA, USA, 2025. [Google Scholar]
- Mordor Intelligence. Internet of Medical Things Market Size & Share Analysis—Growth Trends & Forecasts (2025–2030); Technical Report; Mordor Intelligence: Hyderabad, India, 2025. [Google Scholar]
- Armis Security. Mastering Patient-Centric Cybersecurity for Healthcare: A Playbook for Healthcare Delivery Organizations; Armis Security: San Francisco, CA, USA, 2025. [Google Scholar]
- Federal Bureau of Investigation. Unpatched and Outdated Medical Devices Provide Cyber Attack Opportunities; Private Industry Notification PIN 20220912-001; Federal Bureau of Investigation, Cyber Division: Washington, DC, USA, 2022.
- Zarpelão, B.B.; Miani, R.S.; Kawakani, C.T.; de Alvarenga, S.C. A survey of intrusion detection in Internet of Things. J. Netw. Comput. Appl. 2017, 84, 25–37. [Google Scholar] [CrossRef]
- Butun, I.; Morgera, S.D.; Sankar, R. A survey of intrusion detection systems in wireless sensor networks. IEEE Commun. Surv. Tutor. 2013, 16, 266–282. [Google Scholar] [CrossRef]
- Mujahid, M.; Mirdad, A.R.; Alamri, F.S.; Ara, A.; Khan, A. Software defined network intrusion system to detect malicious attacks in computer Internet of Things security using deep extractor supervised random forest technique. PeerJ Comput. Sci. 2025, 11, e3103. [Google Scholar] [CrossRef] [PubMed]
- Farhan, S.; Mubashir, J.; Haq, Y.U.; Mahmood, T.; Rehman, A. Enhancing network security: An intrusion detection system using residual network-based convolutional neural network. Clust. Comput. 2025, 28, 251. [Google Scholar] [CrossRef]
- Alrayes, F.S.; Zakariah, M.; Amin, S.U.; Khan, Z.I.; Alqurni, J.S. CNN Channel Attention Intrusion Detection System Using NSL-KDD Dataset. Comput. Mater. Contin. 2024, 79, 4319. [Google Scholar] [CrossRef]
- Mitchell, R.; Chen, I.R. A survey of intrusion detection techniques for cyber-physical systems. ACM Comput. Surv. (CSUR) 2014, 46, 55. [Google Scholar] [CrossRef]
- Farnaaz, N.; Jabbar, M. Random forest modeling for network intrusion detection system. Procedia Comput. Sci. 2016, 89, 213–217. [Google Scholar] [CrossRef]
- Cherif, A. Intrusion Detection for Internet of Medical Things (IoMT) using Extreme Learning Machine. In Proceedings of the 2025 2nd International Conference on Advanced Innovations in Smart Cities (ICAISC), Jeddah, Saudi Arabia, 9–11 February 2025; pp. 1–7. [Google Scholar] [CrossRef]
- Yamashita, T.; Hirasawa, K.; Hu, J.; Murata, J. Multi-branch structure of layered neural networks. In Proceedings of the 9th International Conference on Neural Information Processing (ICONIP’02), Singapore, 18–22 November 2002; Volume 1, pp. 243–247. [Google Scholar] [CrossRef]
- Geirhos, R.; Jacobsen, J.H.; Michaelis, C.; Zemel, R.; Brendel, W.; Bethge, M.; Wichmann, F.A. Shortcut learning in deep neural networks. Nat. Mach. Intell. 2020, 2, 665–673. [Google Scholar] [CrossRef]
- Diro, A.A.; Chilamkurti, N. Distributed attack detection scheme using deep learning approach for Internet of Things. Future Gener. Comput. Syst. 2018, 82, 761–768. [Google Scholar] [CrossRef]
- Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [PubMed]
- Hady, A.A. WUSTL-EHMS-2020. 2020. Available online: https://www.cse.wustl.edu/~jain/ehms/index.html (accessed on 20 October 2024).
- Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
- Information and Computer Science, University of California, Irvine. KDD Cup 1999 Data. 2007. Available online: http://kdd.ics.uci.edu/databases/kddcup99/kddcup99.html (accessed on 19 October 2024).
- Zhang, J.; Zulkernine, M.; Haque, A. Random-Forests-Based Network Intrusion Detection Systems. IEEE Trans. Syst. Man Cybern. Part C (Appl. Rev.) 2008, 38, 649–659. [Google Scholar] [CrossRef]
- Hady, A.A.; Ghubaish, A.; Salman, T.; Unal, D.; Jain, R. Intrusion Detection System for Healthcare Systems Using Medical and Network Data: A Comparison Study. IEEE Access 2020, 8, 106576–106584. [Google Scholar] [CrossRef]
- Li, Y.; Xia, J.; Zhang, S.; Yan, J.; Ai, X.; Dai, K. An efficient intrusion detection system based on support vector machines and gradually feature removal method. Expert Syst. Appl. 2012, 39, 424–430. [Google Scholar] [CrossRef]
- Tesfahun, A.; Bhaskari, D.L. Intrusion Detection Using Random Forests Classifier with SMOTE and Feature Reduction. In Proceedings of the 2013 International Conference on Cloud & Ubiquitous Computing & Emerging Technologies, Pune, India, 15–16 November 2013; pp. 127–132. [Google Scholar] [CrossRef]
- Shah, B.; Trivedi, B.H. Reducing Features of KDD CUP 1999 Dataset for Anomaly Detection Using Back Propagation Neural Network. In Proceedings of the 2015 Fifth International Conference on Advanced Computing & Communication Technologies, Haryana, India, 21–22 February 2015; pp. 247–251. [Google Scholar] [CrossRef]
- Zaib, M.H. NSL KDD Dataset. 2024. Available online: https://www.kaggle.com/datasets/hassan06/nslkdd (accessed on 19 July 2024).
- Kale, R.; Lu, Z.; Fok, K.W.; Thing, V.L.L. A Hybrid Deep Learning Anomaly Detection Framework for Intrusion Detection. In Proceedings of the 2022 IEEE 8th Intl Conference on Big Data Security on Cloud (BigDataSecurity), IEEE Intl Conference on High Performance and Smart Computing, (HPSC) and IEEE Intl Conference on Intelligent Data and Security (IDS), Jinan, China, 6–8 May 2022; pp. 137–142. [Google Scholar] [CrossRef]
- Albulayhi, K.; Abu Al-Haija, Q.; Alsuhibany, S.A.; Jillepalli, A.A.; Ashrafuzzaman, M.; Sheldon, F.T. IoT Intrusion Detection Using Machine Learning with a Novel High Performing Feature Selection Method. Appl. Sci. 2022, 12, 5015. [Google Scholar] [CrossRef]
- Iwendi, C.; Anajemba, J.H.; Biamba, C.; Ngabo, D. Security of Things Intrusion Detection System for Smart Healthcare. Electronics 2021, 10, 1375. [Google Scholar] [CrossRef]
- Alani, M.M.; Mashatan, A.; Miri, A. Explainable Ensemble-Based Detection of Cyber Attacks on Internet of Medical Things. In Proceedings of the 2023 IEEE Intl Conf on Dependable, Autonomic and Secure Computing, Intl Conf on Pervasive Intelligence and Computing, Intl Conf on Cloud and Big Data Computing, Intl Conf on Cyber Science and Technology Congress (DASC/PiCom/CBDCom/CyberSciTech), Abu Dhabi, United Arab Emirates, 14–17 November 2023; pp. 0609–0614. [Google Scholar] [CrossRef]
- Nayak, J.; Meher, S.K.; Souri, A.; Naik, B.; Vimal, S. Extreme learning machine and bayesian optimization-driven intelligent framework for IoMT cyber-attack detection. J. Supercomput. 2022, 78, 14866–14891. [Google Scholar] [CrossRef]
- Nour, M. ToN_IoT Datasets. 2024. Available online: https://research.unsw.edu.au/projects/toniot-datasets (accessed on 19 July 2024).
- Ramya, M.; Sudhakaran, P.; Sivagnanam, Y.; Santhana Krishnan, C. Advanced intrusion detection technique (AIDT) for secure communication among devices in internet of medical things (IoMT). EURASIP J. Wirel. Commun. Netw. 2025, 2025, 34. [Google Scholar] [CrossRef]
- Akar, G.; Sahmoud, S.; Onat, M.; Cavusoglu, Ü.; Malondo, E. L2D2: A Novel LSTM Model for Multi-Class Intrusion Detection Systems in the Era of IoMT. IEEE Access 2025, 13, 7002–7013. [Google Scholar] [CrossRef]
- Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
- Dadkhah, S.; Neto, E.C.P.; Ferreira, R.; Molokwu, R.C.; Sadeghi, S.; Ghorbani, A.A. CICIoMT2024: A Benchmark Dataset for Multi-Protocol Security Assessment in IoMT. Internet Things 2024, 28, 101351. [Google Scholar] [CrossRef]
- Alazab, M.; Awajan, A.; Obeidat, A.; Faruqui, N.; Rehman, H.U. IntruSafe: A FCNN-LSTM hybrid IoMT intrusion detection system for both string and 2D-spatial data using sandwich architecture. Neural Comput. Appl. 2025, 37, 23395–23422. [Google Scholar] [CrossRef]
- Wardhani, R.W.; Putranto, D.S.C.; Jo, U.; Kim, H. Toward enhanced attack detection and explanation in intrusion detection system-based IoT environment data. IEEE Access 2023, 11, 131661–131676. [Google Scholar]
- Nataraj, L.; Karthikeyan, S.; Jacob, G.; Manjunath, B.S. Malware images: Visualization and automatic classification. In Proceedings of the 8th International Symposium on Visualization for Cyber Security, Pittsburgh, PA, USA, 20 July 2011; pp. 1–7. [Google Scholar]
- Binbusayyis, A.; Alaskar, H.; Vaiyapuri, T.; Dinesh, M. An investigation and comparison of machine learning approaches for intrusion detection in IoMT network. J. Supercomput. 2022, 78, 17403–17422. [Google Scholar] [CrossRef] [PubMed]
- Alrayes, F.S.; Zakariah, M.; Amin, S.U.; Khan, Z.I.; Helal, M. Hybrid Cross-Temporal Contrastive Model with Spiking Energy-Efficient Network Intrusion Detection in IOMT. Int. J. Comput. Intell. Syst. 2025, 18, 270. [Google Scholar] [CrossRef]
- Liu, X.Y.; Wu, J.; Zhou, Z.H. Exploratory undersampling for class-imbalance learning. IEEE Trans. Syst. Man Cybern. Part B (Cybern.) 2020, 39, 539–550. [Google Scholar]







| Authors | Dataset | Methodology | Results |
|---|---|---|---|
| Zhang et al. [21] | KDD 1999 | Random Forest (RF) for anomaly detection | 95% accuracy, 1% false-positive rate |
| Li et al. [23] | KDD 1999 | Clustering, Ant Colony Algorithm, SVM | 98.62% accuracy, MCC of 0.861 |
| Shah et al. [25] | KDD 1999 | Information Gain (IG) for feature reduction | Improved model performance with reduced dataset |
| Tesfahun et al. [24] | KDD 1999 | Random Forest with IG | Enhanced generalization capacity |
| Kale et al. [27] | NSL-KDD, CIC-IDS2018, TON IoT | Three-stage deep learning framework (K-means, GANomaly, CNN) | 91.6% accuracy on NSL-KDD |
| Albulayhi et al. [28] | NSL-KDD | Feature selection using set theory | 99.98% classification accuracy |
| Iwendi et al. [29] | NSL-KDD | RF with Genetic Algorithm for feature optimization | 98.81% detection rate, 0.8% false alarm rate |
| Nayak et al. [31] | ToN_IoT | Bayesian Optimization and ELM | High precision and recall, but no class imbalance solution |
| Hady et al. [22] | Custom dataset WUSTL-EHMS-2020 (16,000 records) | Integration of medical and network data using EHMS testbed | Improved performance by 7% to 25%; SVM accuracy 92.46%, ANN AUC 92.98% |
| Mohammed M. et al. [30] | WUSTL-EHMS-2020 | Ensemble learning and explainable AI with random over sampling | 99.96% accuracy and 0.998 F1 score |
| Cherif A. [13] | WUSTL-EHMS-2020 | Multiple neural network architectures with three class balancing approaches | 93.05% accuracy, 0.9518 precision and 0.8037 F1-score with weighted loss |
| Ramya et al. [33] | Dataset proposed in [40] | Particle swarm Optimization combined with a probabilistic neural network | Accuracy: 96.4%, F1-score: up to 95.67% |
| Akar et al. [34] | CIC-IoMT2024 [36] | LSTM-based model | Accuracy: 98%, F1-score: 98% |
| AlAzab et al. [37] | CIC-IoMT2023 and Malimg datasets | FCNN combined with LSTM | Accuracy: 97.66%, F1-score: 97.85% |
| Ref. | Advanced Architectures | IoMT Dataset | Class Imbalance Handling | Feature Reduction |
|---|---|---|---|---|
| [21] | - | - | - | - |
| [23] | - | - | - | ✓ |
| [25] | - | - | - | ✓ |
| [24] | - | - | - | - |
| [27] | ✓ | ✓ | - | - |
| [28] | - | ✓ | - | ✓ |
| [29] | - | - | - | ✓ |
| [31] | - | - | - | ✓ |
| [22] | - | ✓ | ✓ | - |
| [30] | ✓ | ✓ | ✓ | |
| [13] | - | ✓ | ✓ | - |
| [33] | - | ✓ | - | ✓ |
| [34] | - | ✓ | ✓ | - |
| [37] | ✓ | - | - | - |
| [41] | ✓ | - | ✓ | - |
| Measurement | Value |
|---|---|
| Size | 4.4 MB |
| Normal samples | 14,272 (87.5%) |
| Attack samples | 2046 (12.5%) |
| Total number of samples | 16,318 |
| Model | Layer Dimensions | Activation Function | Dropout and Fusion Details |
|---|---|---|---|
| ANN | Input: 34 Hidden: [40, 40, 20, 10, 10, 10, 10] Output: 2 | ReLU | None |
| ANN_v1 | Input: 34 Hidden: [256, 256, 128, 64, 64, 64, 64] Output: 2 | ReLU | None |
| ANN_v2 | Network Branch: [256, 256] Biometric Branch: [256, 256] Shared Layers: [128, 64, 64] Output: 2 | ReLU | Dropout (0.4) after shared layers; Fusion: element-wise addition scaled by 0.5 |
| ANN_v3 | Network Branch: [256, 256] Biometric Branch: [256, 256] Shared Layers: [256, 128, 64] Output: 2 | ReLU | Dropout (0.4) after shared layers; Fusion: concatenation of branch outputs |
| ANN_v4 | Input: 34 Hidden: [256, 256, 128, 64, 64, 64, 64] Output: 2 | ReLU | Shortcut connections between layers 1–2, 4–5, 5–6, 6–7; No dropout |
| Hyperparameter | Value |
|---|---|
| Learning rate | |
| Batch size | 64 |
| Weight decay | |
| Optimizer | AdamW |
| Loss function | Cross Entropy (standard or weighted) |
| Evaluation interval | 50 epochs |
| Maximum epochs | 200–500 (architecture dependent) |
| Model | Class Balancing Method | AUC | ACC | PR | RC | F1 |
|---|---|---|---|---|---|---|
| ANN | SMOTE | 0.8491 | 0.8753 | 0.7427 | 0.8491 | 0.7783 |
| ANN_v1 | SMOTE | 0.8544 | 0.8983 | 0.7750 | 0.8544 | 0.8062 |
| ANN_v2 | SMOTE | 0.8766 | 0.8955 | 0.7721 | 0.8766 | 0.8096 |
| ANN_v3 | SMOTE | 0.8740 | 0.9032 | 0.7851 | 0.8740 | 0.8182 |
| ANN_v4 | SMOTE | 0.8554 | 0.9035 | 0.7839 | 0.8554 | 0.8129 |
| AE+ANN_v4 | SMOTE | 0.7485 | 0.9308 | 0.9383 | 0.7485 | 0.8085 |
| ANN | Hybrid | 0.8518 | 0.9179 | 0.8132 | 0.8518 | 0.8308 |
| ANN_v1 | Hybrid | 0.8577 | 0.9213 | 0.8201 | 0.8577 | 0.8373 |
| ANN_v2 | Hybrid | 0.8750 | 0.9323 | 0.8437 | 0.8750 | 0.8582 |
| ANN_v3 | Hybrid | 0.8671 | 0.9203 | 0.8163 | 0.8671 | 0.8387 |
| ANN_v4 | Hybrid | 0.8739 | 0.9151 | 0.8047 | 0.8739 | 0.8335 |
| AE+ANN_v4 | Hybrid | 0.7925 | 0.9059 | 0.7942 | 0.7925 | 0.7934 |
| ANN | Weighted cross-entropy Loss | 0.8588 | 0.9145 | 0.8049 | 0.8588 | 0.8283 |
| ANN_v1 | Weighted cross-entropy Loss | 0.8559 | 0.9197 | 0.8168 | 0.8559 | 0.8345 |
| ANN_v2 | Weighted cross-entropy Loss | 0.8786 | 0.9403 | 0.8650 | 0.8786 | 0.8716 |
| ANN_v3 | Weighted cross-entropy Loss | 0.8467 | 0.8161 | 0.6938 | 0.8467 | 0.7216 |
| ANN_v4 | Weighted cross-entropy Loss | 0.8534 | 0.8382 | 0.7096 | 0.8534 | 0.7431 |
| AE+ANN_v4 | Weighted cross-entropy Loss | 0.7463 | 0.9200 | 0.8705 | 0.7463 | 0.7909 |
| Model | Total Training Time (s) | Training Time per Epoch (s) | Inference Time (s) |
|---|---|---|---|
| ANN | 234.992 | 0.47 | 0.0557 |
| ANN_v1 | 471.676 | 0.94 | 0.0680 |
| ANN_v2 | 730.803 | 1.56 | 0.0921 |
| ANN_v3 | 950.360 | 1.90 | 0.0631 |
| ANN_v4 | 164.865 | 0.82 | 0.0331 |
| AE+ANN_v4 | 1110.085 | 2.77 | 0.0329 |
| Study | Approach | Acc | F1 | AUC | PR | RC |
|---|---|---|---|---|---|---|
| Cherif A. [13] | ELM (256) + SMOTE | 0.8444 | 0.7223 | 0.7789 | 0.6949 | 0.7789 |
| Cherif A. [13] | ELM (256) + Weighted cross-entropy Loss | 0.9305 | 0.8037 | 0.7404 | 0.9518 | 0.7404 |
| Hady et al. [22] | SVM with SMOTE | 0.9246 | Not reported | 0.8237 | Not reported | Not reported |
| Hady et al. [22] | ANN with SMOTE | 0.9040 | Not reported | 0.9342 | Not reported | Not reported |
| Mohammed et al. [30] | Ensemble with random over-sampling | 0.9980 | 0.9980 | Not reported | 0.9980 | 0.9980 |
| Current study | ANN_v2 + Weighted cross-entropy Loss | 0.9403 | 0.8716 | 0.8786 | 0.8650 | 0.8786 |
| Current study | AE+ANN_v4 + SMOTE | 0.9308 | 0.8085 | 0.7485 | 0.9383 | 0.7485 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Ammar, H.; Cherif, A. Optimizing IoMT Security: Performance Trade-Offs Between Neural Network Architectural Design, Dimensionality Reduction, and Class Imbalance Handling. IoT 2025, 6, 74. https://doi.org/10.3390/iot6040074
Ammar H, Cherif A. Optimizing IoMT Security: Performance Trade-Offs Between Neural Network Architectural Design, Dimensionality Reduction, and Class Imbalance Handling. IoT. 2025; 6(4):74. https://doi.org/10.3390/iot6040074
Chicago/Turabian StyleAmmar, Heyfa, and Asma Cherif. 2025. "Optimizing IoMT Security: Performance Trade-Offs Between Neural Network Architectural Design, Dimensionality Reduction, and Class Imbalance Handling" IoT 6, no. 4: 74. https://doi.org/10.3390/iot6040074
APA StyleAmmar, H., & Cherif, A. (2025). Optimizing IoMT Security: Performance Trade-Offs Between Neural Network Architectural Design, Dimensionality Reduction, and Class Imbalance Handling. IoT, 6(4), 74. https://doi.org/10.3390/iot6040074

