Next Article in Journal
Recursive Batch Smoother with Multiple Linearization for One Class of Nonlinear Estimation Problems: Application for Multisensor Navigation Data Fusion
Next Article in Special Issue
GECO: A Real-Time Computer Vision-Assisted Gesture Controller for Advanced IoT Home System
Previous Article in Journal
Recent Real-Time Aerial Object Detection Approaches, Performance, Optimization, and Efficient Design Trends for Onboard Performance: A Survey
Previous Article in Special Issue
AIoT-Based Eyelash Extension Durability Evaluation Using LabVIEW Data Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid AI Intrusion Detection: Balancing Accuracy and Efficiency

by
Vandit R Joshi
1,
Kwame Assa-Agyei
1,
Tawfik Al-Hadhrami
1 and
Sultan Noman Qasem
2,*
1
Department of Computer Science, Nottingham Trent University, Nottingham NG1 4FQ, UK
2
Computer Science Department, College of Computer and Information Sciences, Imam Mohammad Ibn Saud Islamic University (IMSIU), Riyadh 11432, Saudi Arabia
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(24), 7564; https://doi.org/10.3390/s25247564
Submission received: 9 October 2025 / Revised: 26 November 2025 / Accepted: 3 December 2025 / Published: 12 December 2025
(This article belongs to the Special Issue AI-Empowered Internet of Things)

Abstract

The Internet of Things (IoT) has transformed industries, healthcare, and smart environments, but introduces severe security threats due to resource constraints, weak protocols, and heterogeneous infrastructures. Traditional Intrusion Detection Systems (IDS) fail to address critical challenges including scalability across billions of devices, interoperability among diverse protocols, real-time responsiveness under strict latency, data privacy in distributed edge networks, and high false positives in imbalanced traffic. This study provides a systematic comparative evaluation of three representative AI models, CNN-BiLSTM, Random Forest, and XGBoost for IoT intrusion detection on the NSL-KDD and UNSW-NB15 datasets. The analysis quantifies the achievable detection performance and inference latency of each approach, revealing a clear accuracy–latency trade-off that can guide practical model selection: CNN-BiLSTM offers the highest detection capability (F1 up to 0.986) at the cost of higher computational overhead, whereas XGBoost and Random Forest deliver competitive accuracy with significantly lower inference latency (sub-millisecond on conventional hardware). These empirical insights support informed deployment decisions in heterogeneous IoT environments where accuracy-critical gateways and latency-critical sensors coexist.

1. Introduction

The rapid integration of Internet of Things (IoT) devices across multiple domains—from smart homes and healthcare monitoring to industrial automation and large-scale sensor networks has fundamentally reshaped modern digital infrastructures.
Despite these advances, IoT ecosystems face severe security vulnerabilities. Most devices operate on constrained hardware, lack built-in security, and are deployed with minimal oversight, making them prime targets for DoS attacks, data leaks, man-in-the-middle exploits, and unauthorized control [1].
Traditional Intrusion Detection Systems (IDS), reliant on centralized processing and high computational overhead, are ill-suited for such environments. They fail to overcome five core challenges: (1) scalability to support millions of heterogeneous devices, (2) interoperability across incompatible protocols (e.g., MQTT, CoAP, Zigbee), (3) real-time detection under microsecond latency constraints, (4) preservation of data privacy in decentralized edge settings, and (5) reliable performance amid highly imbalanced attack distributions.
Artificial Intelligence (AI)-driven IDS leveraging machine learning and deep learning offer adaptive, scalable alternatives capable of learning evolving attack patterns and responding in real time. However, existing solutions remain hindered by high resource demands, inconsistent accuracy, and limited generalizability across diverse IoT contexts (Akif et al., 2025) [2]. This study addresses these gaps through a performance-centric evaluation of lightweight AI models tailored for IoT constraints. The key contributions are as follows: (1) a reproducible and well-performing CNN-BiLSTM baseline that combines 1D-CNN and Bi-LSTM layers; (2) a detailed measurement of inference latency for all three models on the same hardware platform; and (3) an empirical accuracy–latency trade-off analysis to guide model selection in IoT IDS deployments.
This study proposes the development and evaluation of lightweight AI-powered IDS models tailored specifically for IoT environments. The aim is to enhance the detection and classification of malicious activity without compromising device efficiency or network responsiveness. By exploring hybrid modeling techniques and evaluating model effectiveness through metrics like accuracy, precision, recall, F1-score, and latency per sample, the aim of this study is to develop a scalable and efficient framework capable of detecting intrusions within IoT environments. The selection of CNN-BiLSTM, Random Forest, and XGBoost is strategically motivated by their complementary strengths in addressing IoT-specific constraints. CNN-BiLSTM integrates convolutional layers for local spatial feature extraction with bidirectional LSTM for long-range temporal dependency modeling, making it ideal for detecting sophisticated, sequential attack patterns (e.g., multi-stage exploits) prevalent in modern IoT traffic [3,4]. Random Forest, a tree-based ensemble, offers inherent robustness to noisy and imbalanced data common in real-world IoT logs while providing interpretable feature importance rankings without requiring extensive hyper parameter tuning. XGBoost, an optimized gradient boosting framework, delivers state-of-the-art performance with minimal inference latency (sub-millisecond), rendering it suitable for edge deployment on battery-powered sensors. This tri-model evaluation enables a comprehensive performance-centric analysis, quantifying trade-offs between detection capability and computational cost across accuracy-prioritized and latency-critical IoT scenarios.
The structure of this paper is organized as follows: Section 2 discusses previous studies; Section 3 explains the research methodology, covering the experiment setup, NSL-KDD and UNSW-NB15 dataset preparation, the CNN-BiLSTM framework, and Exploratory Data Analysis (EDA); Section 4 outlines the obtained results and their discussion, with evaluations of CNN-BiLSTM, Random Forest, and XGBoost models; Section 5 provides the merged conclusions and future work, summarizing key findings, highlighting model trade-offs for IoT Intrusion Detection Systems (IDS), and outlining potential directions for further investigation.

2. Related Work

The growing adoption of Internet of Things (IoT) technologies across critical sectors has increased the demand for resilient, scalable, and interpretable Intrusion Detection Systems (IDS). Recent literature reflects a convergence of Artificial Intelligence (AI), Explainable AI (XAI), deep learning, Federated Learning-based solutions, and optimization techniques to tackle the specific challenges found in IoT systems, including evasion attacks and data security enhancements.

2.1. AI and XAI-Driven IDS Models

Explainable AI has become recognized as pivotal theme in IDS research, aiming enhance transparency and user trust. Alabbadi and Bajaber (2025) [5] introduced an XAI-based IDS framework for IoT data streams, demonstrating improved interpretability but lacking scalability analysis. Similarly, Patil et al., 2022 [6] and Arreche et al., 2024 [7] integrated XAI tools such as SHAP and LIME into ensemble models, enabling interpretable alerts for analysts. However, these studies often overlook the trade-offs between explainability and performance, particularly in resource-constrained deployments. Baral et al. (2024) [8] extended this paradigm by incorporating large language models (LLMs), yet complexity of implementation continues to hinder practical deployment.

2.2. Deep Learning and Hybrid Architectures

Hybrid deep learning models, especially those combining Convolutional Neural Networks (CNNs) and Long Short-Term Memory (LSTM) networks, have shown promise in enhancing detection accuracy. Altunay and Albayrak (2023) [3] and Gueriani et al., 2024 [4] reported high classification performance using CNN-LSTM architectures on benchmark datasets such as UNSW-NB15 and NSL-KDD. Hossain (2025) [9] further validated the scalability of deep learning models across multiple architectures, including 1D CNNs and RNNs. Liang et al., 2025 [10] introduced a CO-GNN (Combinatorial Optimization Graph Neural Network) framework for heterogeneous secure transmissions in IRS-assisted NOMA communications, optimizing beamforming, power allocation, and phase shifts to maximize secrecy rates against eavesdroppers in dynamic IoT networks; this highlights GNNs’ potential for proactive, scalable security in heterogeneous environments, aligning with the adaptive hybrid models explored in the study. Despite these advances, most studies lack deployment-focused evaluations, leaving questions about latency, memory footprint, and adaptability in heterogeneous IoT environments.

2.3. Lightweight and Modular IDS Designs

Efforts to design lightweight IDS models suitable for low-resource devices have gained traction. Jouhari and Guizani (2024) [11] proposed a CNN-BiLSTM model optimized for constrained environments, achieving notable accuracy but struggling with complex attack patterns. Sowmya and Anita (2023) [12] and Faisal and Mubeen (2024) [13] emphasized the importance of adaptive and modular IDS frameworks, though their work remained largely conceptual without empirical validation. Complementing these, Abu Laila (2025) [14] developed a responsive machine-learning framework for preventing evasion attacks in IoT-based IDS, incorporating adversarial training, feature-space transformations, and ensemble detection. This lightweight approach achieved 94.7% accuracy on clean traffic and 89.3% against evasions, with only 15% additional computational overhead, making it ideal for resource-constrained IoT deployments.

2.4. Federated Learning and Privacy-Preserving IDS

Researchers have investigated Federated Learning (FL) as a collaborative and privacy-aware approach for intrusion detection. Campos et al., 2022 [15], Javeed et al., 2024 [16], and Ragab et al., 2025 [17] demonstrated the potential of FL-enabled IDS models to maintain data privacy while achieving high detection rates. However, communication overhead, latency, and model convergence issues persist. Liu et al., 2021 [18] and Rampone et al., 2025 [19] attempted to mitigate these challenges by integrating blockchain and near-real-time architectures, yet the complexity of these systems limits their scalability.

2.5. Feature Selection and Optimization Techniques

Feature selection remains critical for improving IDS efficiency. Alkahla et al., 2024 [20] and Fang et al., 2023 [21] reviewed filter-based, wrapper-based, and genetic algorithm-driven techniques, showing that optimized feature sets enhance classification accuracy. Bibi et al., 2024 [22] explored pruning and quantization in NLP models, offering insights into model compression strategies, though their applicability to IDS remains underexplored. Alshuaibi et al., 2025 [23] proposed a hybrid genetic algorithm (GA) and hidden Markov model (HMM)-based hashing technique within a block hashing framework to bolster data security against cyber threats like dictionary and brute-force attacks. Drawing from evolutionary biology, the GA incorporates mutation, crossover, and selection operators, paired with Hill cipher encryption using a singular key matrix to minimize hash collisions and reversals. Experimental results highlight its superior resistance to attacks and robustness over existing methods, evaluated via break attempts, hash key size, and algorithmic strength offering a cryptographic layer that complements IDS by ensuring data integrity in IoT preprocessing pipelines. Similarly, Alshinwan et al., 2025 [24] introduced an unsupervised text feature selection approach using an improved Prairie Dog optimization (PDO) algorithm for text clustering, which selects informative feature subsets to mitigate sparsity and uninformative traits in high-dimensional documents. Assessed via K-means clustering against literature benchmarks, it reduces computational time while yielding precise clusters; though focused on text data, this has strong potential for IDS applications, such as optimizing log analysis or anomaly detection in network traffic datasets by enhancing feature relevance and efficiency in unsupervised IoT security tasks. Table 1 provides a comprehensive comparative assessment of recent IoT intrusion detection studies, clearly delineating their objectives, principal contributions, and limitations.

3. Methodology

A quantitative experimental design was adopted using the NSL-KDD and UNSW-NB15 dataset. Data preprocessing included normalization, label encoding, and feature selection. Three models, CNN, Bi-LSTM, and hybrid CNN-BiLSTM, were developed and trained using TensorFlow. Performance metrics included accuracy, precision, recall, F1-score, and latency. Benchmarking was conducted against Random Forest and XGBoost classifiers. The CNN-BiLSTM architecture leveraged spatial feature extraction via CNN and temporal sequence learning via Bi-LSTM.

3.1. Experimental Setup

Experiments were carried out using a high-capacity computing workstation running Windows 11, equipped with an AMD Ryzen 7 5800X 8-core processor (3.80 GHz), 32 GB DDR4 RAM, a 128-bit Nvidia RTX 4060Ti GPU with 16 GB VRAM (2500 MHz core clock), and 1 TB SSD storage. This configuration provided a robust environment for training and evaluating deep learning models efficiently.

3.2. Data Preparation

The study uses the NSL-KDD [25] and UNSW-NB15 [26] datasets. The datasets contain labeled records of both normal and malicious network traffic, covering a broad spectrum of attack types. The data undergoes a structured preparation process to ensure quality and relevance for model training.
Initial steps include preprocessing the dataset by eliminating null entries, duplicate entries, and irrelevant features. Numerical attributes are scaled into a normalized range (commonly 0–1) to ensure uniform influence during training. Categorical variables such as protocol type, service, and connection flags are transformed into numerical representations through methods such as one-hot or label encoding.
Feature selection is guided by statistical correlation, domain expertise, and practical relevance to network behaviors. The final dataset is divided into training and testing portions following an 80:20 proportion, enabling robust model evaluation. This processing ensures the data is clean, balanced, and optimized for deep-learning applications in intrusion detection.

3.3. Proposed Model

The proposed model as shown in Figure 1 is a hybrid CNN-BiLSTM model designed to tackle the complex nature of IoT intrusion detection, where attacks often unfold dynamically across multiple dimensions.
Model Explanation:
(1)
Input Layer: Input Shape (T, 1) here T is the sequence length and 1 represents single Feature channel. One sequence is fed per sample, but this can be extended to multiple channels when incorporating additional feature types.
(2)
1D convolution Layer:
(i)
Config: 96 Filter, each kernel size of 5.
(ii)
Activation ‘Relu’, ReLU   ( F ( x ) = m a x ( 0 , x ) ) [27].
(iii)
Convolution operation [21].
y i ( k ) = R e L U j = 0 m 1 w j ( k ) x i + j + b ( k )
Here w ( k ) and b ( k ) are kernel weights and bias for k-th filter, m is kernel size and ReLU activation. y i ( k ) is the feature map.
(iv)
Convolution layer (CNN) captures local temporal patterns, i.e., traffic signatures in network data.
(3)
MaxPooling1D layer: pool size 2. Reduces sequence length by half, retaining only the most outstanding features. It is used for dimensionality reduction and Noise suppression [28].
(4)
Bidirectional LSTM layer:
(i)
48 units per direction, total 96 states.
(ii)
Dropout 0.5 input connection and recurrent dropout 0.2 for recurrent connection.
(iii)
LSTM equation [29]
The LSTM updates are computed as follows (all gates and states have dimension 48):
Notation:
x t R d : input at time t h t 1 , h t R 48 : previous and current hidden state
c t 1 , c t R 48 : previous and current cell state
i t , f t , o t R 48 : input, forget, output gates
c ~ t R 48 : candidate cell state
W i , W f , W o , W c R 48 x ( d + 48 ) : weight matrices (applied to concatenated h t 1 , x t )
b i , b f , b o , b c R 48 : bias vectors
σ : sigmoid function
: element-wise multiplication
[ht−1, xt]: concatenation of previous hidden state and current input
i t = σ ( W i h t 1 , x t + b i )   / / Input   gate
f t = σ ( W f h t 1 , x t + b f )   / / forget   gate
o t = σ ( W o h t 1 , x t + b o )   / / output   gate
c ~ t = tanh ( W c h t 1 , x t + b c )   / / candidate   gate
c t = f t     c t 1 + i t c ~ t   / / cell   state   update
h t = o t tanh ( c t )   / / Hidden   state   output
(iv)
Bidirectional extension [30]:
Forward LSTM processes the sequence as follows:
h t = L S T M ( x t , h t 1 )
Backward LSTM processes in reverse order:
h t = L S T M ( x t , h t + 1 )
The final representation is the concatenation:
h t = [ h t ; h t ]
This ensure that temporal pattern such as attack bursts followed by silence or l gitimate activity interrupted by anomalies are both captured.
(5)
Dense layer (64 units, ReLU activation): This layer projects the sequential representations generated by the BiLSTM into a higher-dimensional, non-linear feature space. The ReLU activation introduces sparsity and non-linearity, enabling the network to capture complex interactions across features, such as the joint influence of TCP flags, packet timing, and connection duration, thus yielding richer higher-order representations.
(6)
Dropout layer: 0.5 dropout rate regularizes the network by randomly disabling neurons during training, this will prevent over fitting [31].
(7)
Output layer: number of units = number of classes (5 classes in this case), activation SoftMax. SoftMax Equation:
P y = c   |   x = exp z c j = 1 C exp z j
where z c is the logit for class c and C is quantity of target classes.
SoftMax ensures outputs are interpretable as class probabilities, facilitating multiclass classification.
The early stop function added to stop the model training when there is no significant improvement on the Val_loss. This will prevent the model from completing the training which can make it overly fit for the dataset.

3.4. Hyper Parameter Tuning

Hyper parameter optimization for the CNN-BiLSTM intrusion detection model was conducted using a grid search combined with 5-fold cross-validation on the training subsets of the NSL-KDD and UNSW-NB15 datasets. This process systematically evaluated combinations of key hyper parameters to maximize the validation F1-score, balancing accuracy, generalization, and over-fitting prevention. Early stopping with a patience of 10 epochs was applied, monitoring validation loss to halt training when no improvement was observed. The tuned hyper parameters included one CNN layer to minimize latency without sacrificing accuracy, 96 filters in the CNN layer for optimal temporal feature extraction, 48 LSTM units per direction to balance sequence modeling and efficiency, an Adam learning rate of 0.001 for stable convergence, a batch size of 32 to enhance gradient stability on imbalanced data, and dropout rates of 0.2 for recurrent layers and 0.5 for input and dense layers, reducing validation loss by approximately 15%. A total of 216 hyper parameter combinations were evaluated, with the optimal configuration achieving validation F1-scores of 0.85 on NSL-KDD (multiclass) and 0.97 on UNSW-NB15, resulting in a robust model suitable for IoT environments with computational constraints.

3.5. Data Visualization and EDA

Figure 2 depicts the occurrence distribution of various attack types in the NSL-KDD training dataset. The most prevalent label is “normal,” accounting for roughly 67,000 instances. “Neptune” follows as the second-most frequent attack, with around 41,000 occurrences. Compared to the dominant categories, most other attack types show up far less frequently. For instance, “warezclient” appears around 2500 times, while others like “teardrop,” “nmap,” “satan,” “smurf,” and “pod” each fall below 2000. Many additional attacks have fewer than 1000 examples, which clearly points to a strong class imbalance across the dataset.
Figure 2 shows how different attack categories are distributed the UNSW-NB15 dataset. Approximately 56,000 instances are labeled as ‘normal’, followed by around 40,000 occurrences of the ‘generic’ attack type. ‘Exploits’ account for roughly 34,000 instances, while ‘Fuzzer’ and ‘DoS’ attacks register about 18,000 and 13,000, respectively. The ‘Reconnaissance’ category includes close to 10,000 samples. In contrast, all other attack types—such as ‘analysis’, ‘backdoor’, ‘shellcode’, and ‘worms’—each appear fewer than 2000 times, highlighting a significant class imbalance across the dataset.

3.6. Feature Importance Analysis

A feature importance analysis was conducted to improve interpretability and understand the decision-making process of the Random Forest and XGBoost models. Using Gini impurity decrease for Random Forest and gain-based importance for XGBoost, the analysis identified the network features most influential in intrusion detection. For the NSL-KDD dataset, both models highlighted traffic volume and connection-based attributes such as src_bytes, dst_bytes, and same_srv_rate as key indicators of probing and DoS attacks, while XGBoost additionally emphasized error-related features like serror_rate and dst_host_srv_serror_rate. For the UNSW-NB15 dataset, both models prioritized flow duration and rate-based features, including sttl, ct_state_ttl, sbytes, and rate, which capture the temporal and throughput characteristics of IoT traffic. XGBoost further weighted stateful and load-based attributes (sload, dload), enhancing its ability to detect complex attack patterns. The models consistently focused on lightweight traffic intensity and flow dynamic features, demonstrating their suitability for efficient, real-time intrusion detection in IoT environments.
Figure 3 presents the distribution of the protocol_type feature within the NSL-KDD and UNSW-NB15 datasets. TCP emerges as dominant protocol, appearing in approximately 18,500 instances in NSL-KDD and around 45,000 instances in UNSW-NB15. UDP ranks second, though significantly less frequent, with roughly 2500 occurrences in NSL-KDD and 30,000 in UNSW-NB15. ICMP is the least represented in NSL-KDD, with just over 1000 examples. Additional protocols such as UNAS, ARP, OSPF, and SCTP are present in the UNSW-NB15 dataset but absent from NSL-KDD. The chart highlights a clear skew in protocol representation, with TCP overwhelmingly dominating the test set. For multiclass classification of attack scenarios, the label-to-class mapping is defined as follows: Class 0 represents normal traffic; Class 1 corresponds to Denial of Service (DoS) attacks; Class 2 captures probing activities; Class 3 denotes Root-to-Local (R2L) intrusions; and Class 4 identifies User-to-Root (U2R) attacks.

3.7. Confusion Matrix for Multiclass Classification of Attack Types

Figure 4, Figure 5 and Figure 6 illustrate the model’s classification performance by contrasting predicted outcomes with actual labels. Correct classifications are shown along the diagonal of each confusion matrix, whereas the off-diagonal entries indicate incorrect classification instances, revealing patterns of inter-class confusion. The label-to-class mapping is as follows: Class 0 denotes normal traffic; Class 1 identifies Denial of Service (DoS) attacks; Class 2 captures probing activities; Class 3 corresponds to Root-to-Local (R2L) intrusions; and Class 4 represents User-to-Root (U2R) attacks. CNN-BiLSTM and XGBoost handled the majority classes (Class 0, 1, and 2) effectively, with particular showing very high recall for Class 1. However, the real challenge appeared in the minority classes. CNN-BiLSTM often misclassified Classes 3 and 4 into the larger categories, though it managed to identify a few Class 4 cases and XGBoost showed a similar pattern, where Random Forest failed entirely. Random Forest’s confusion matrix was the most concerning, as it was heavily skewed towards Class 1 and showed almost no recognition of Classes 2, 3, and 4.

3.8. Confusion Matrix for Binary Classification:

Both datasets encompass a wide spectrum of attack categories across multiple protocols and services. However, the UNSW-NB15 dataset is more refined in its labeling structure, representing attack traffic as ‘1’ and benign traffic as ‘0’. In contrast, the NSL-KDD dataset lacks this binary classification. To enable consistent training and testing, we manually assigned ‘0’ to normal instances and ‘1’ to all attack types, effectively transforming the labels into a binary format. This process aligns with the principles of one-hot encoding for binary classification tasks.
Figure 7 displays the confusion matrix corresponding to the CNN-BiLSTM model applied across both the NSL-KDD and UNSW-NB15 datasets. In this context, Class 0 represents normal traffic, while Class 1 denotes attack instances. For the NSL-KDD dataset, the model successfully identified 9275 normal instances and 10,889 attack instances, resulting in an overall accuracy near 89% and a precision rate of around 90%. Performance on the UNSW-NB15 dataset was stronger. The model correctly classified all 37,000 normal traffic instances and misclassified only 1122 out of 46,000 attack samples. This corresponds to an accuracy and precision of approximately 98%, demonstrating the resilience and generalization strength of the model across heterogeneous attack scenarios.
Figure 8 shows the confusion matrix for the Random Forest model used in binary classification of attacks. For the NSL-KDD test dataset, the model performed reasonably well, correctly identifying 9342 normal instances and 8456 attack instances. This translates to an accuracy and precision of about 78%. However, its performance on the UNSW-NB15 dataset was noticeably weaker. The confusion matrix indicates that numerous normal samples were incorrectly labeled as attacks, potentially increasing the false alarm rate in practical intrusion detection applications.
Figure 9 presents the confusion matrix for the XGBoost model under binary classification of attack instances. Findings indicate that the model achieves satisfactory performance on Class 0 (normal traffic) in the NSL-KDD dataset, with relatively few misclassifications outperforming both Random Forest and CNN-BiLSTM in this category. However, its performance on Class 1 (attack traffic) is less effective, with the lowest correct prediction rate among the three models. A similar trend is observed with the UNSW-NB15 dataset: while the model excels in identifying normal traffic, it struggles to accurately classify attack instances, indicating limitations in its generalization across diverse intrusion patterns.

4. Discussion

The IDS models were evaluated using the following performance metrics:

4.1. CNN-BiLSTM Model

4.1.1. Model Accuracy

Figure 10 illustrates a rapid rise in training accuracy (Train_Acc) with multiclass attack NSL-KDD, which remains consistently high across all epochs signifying effective learning from the training data. In contrast, validation accuracy (Val_Acc) starts around 0.82 but fluctuates between 0.76 and 0.81 throughout the training process. This persistent gap between training and validation performance highlights a generalization issue: while the model performs reliably on known data, its ability to handle unseen inputs is less stable. The observed variability in validation accuracy implies that certain attack categories may not be consistently detected across epochs.
Figure 11 shows how the model performed when using binary classification on both the NSL-KDD and UNSW-NB15 datasets. For NSL-KDD, validation accuracy stayed fairly consistent, ranging from 0.83 to 0.89 across epochs. On the other hand, the UNSW-NB15 dataset showed a sharp improvement, with validation accuracy rising from 0.77 to 0.98. The number of epochs shown is limited due to the use of an early stopping function, which automatically halts training when there is no meaningful improvement in validation loss. This helps prevent over fitting and ensures the model does not continue training unnecessarily.

4.1.2. Model Loss

As shown in Figure 12, the training loss begins at approximately 0.15 and quickly drops to near zero, indicating minimal error and strong convergence during multiclass classification on the NSL-KDD training set. In contrast, the validation loss starts around 1.0 and steadily increases, peaking near 5.0 with noticeable fluctuations. This widening gap between training and validation losses suggests noticeable overfitting, where the model becomes tailored to training data yet fails to generalize well to new samples.
Figure 13 shows how the model performed during binary classification training using benchmark datasets; the NSL-KDD and UNSW-NB15 datasets. The CNN framework showed steady improvement in validation loss. For NSL-KDD, the Val_loss stayed between 0.5 and 0.9, which suggests the model was learning effectively without signs of over fitting. On the UNSW-NB15 dataset, Val_loss started at 1.1, briefly spiked to 1.3, and then dropped sharply to around 0.2–0.1. This drop indicates that the model successfully captured training patterns and generalized effectively to previously unseen data.

4.2. Random Forest Model

The Random Forest model was evaluated on both the NSL-KDD and UNSW-NB15 datasets for multiclass and binary attack detection tasks. As shown in Table 2, the model achieved an accuracy of 63.74% for multiclass classification on the NSL-KDD dataset, with corresponding recall and F1-score values of 0.6374 and 0.5474, respectively. The relatively low precision (0.4994) indicates a tendency to misclassify some attack categories, reflecting moderate discriminative capability across multiple classes.
For binary classification, performance improved significantly. On the NSL-KDD dataset, the Random Forest achieved an accuracy of 78.95%, precision of 83.88%, and an F1-score of 0.7880, suggesting a balanced and reliable detection of attack versus normal traffic. Similarly, on the UNSW-NB15 dataset, accuracy reached 75.53% with precision of 83.06% and F1-score of 0.7318. Although slightly lower than on NSL-KDD, this still demonstrates reasonable generalization. The minor performance drop may stem from dataset heterogeneity and variations in attack distribution.

4.3. XGBoost Model

As summarized in Table 2, the XGBoost model also exhibited strong classification performance across both datasets. For multiclass attack detection on the NSL-KDD dataset, it achieved an accuracy of 77.69%, precision of 79.19%, recall of 77.69%, and F1-score of 0.7287. These results indicate that XGBoost effectively balances precision and recall in identifying diverse attack types.
In binary classification tasks, XGBoost demonstrated robust generalization. On the NSL-KDD dataset, it reached 77.30% accuracy, 83.51% precision, and an F1-score of 0.7701. Notably, when applied to the UNSW-NB15 dataset, the model’s performance improved substantially, with accuracy and recall of 96.89%, precision of 97.09%, and F1-score of 0.9690. These outcomes show that XGBoost performs competitively with deep learning architectures, particularly on more balanced datasets like UNSW-NB15.

4.4. Latency per Sample

The latency per sample was calculated by dividing the total inference time by the number of test instances, providing a standardized measure of computational efficiency. As reflected in Table 2, traditional machine-learning models such as Random Forest and XGBoost demonstrated extremely low latency (approximately 0.0002–0.0008 ms/sample), making them suitable for real-time intrusion detection. In contrast, the CNN-BiLSTM models exhibited slightly higher latency (ranging from 0.059 to 0.126 ms/sample), which is expected given their deeper architecture and sequential processing nature. Despite this, all models maintained acceptable inference speeds suitable for deployment in IoT-based IDS environments.

4.5. Discussion of Results

Table 2 presents a detailed comparison of three models CNN-BiLSTM, Random Forest, and XGBoost evaluated across multiclass and binary attack classification tasks employing both the NSL-KDD and UNSW-NB15 benchmark datasets. The analysis considers core performance metrics (F1-score, recall, precision, accuracy) alongside latency per sample, offering a holistic view of each model’s predictive capability and computational efficiency. The CNN-BiLSTM model consistently delivered top-tier classification performance across all tasks. On the UNSW-NB15 dataset, it achieved near-perfect scores (F1: 0.9864, Accuracy: 0.9864), demonstrating its strength in capturing complex attack patterns. Binary classification on NSL-KDD also showed strong results (F1: 0.8949), while multiclass performance remained competitive (F1: 0.7680). However, this accuracy came with a trade-off in latency. Inference times ranged from 0.0590 ms to 0.1261 ms per sample, the highest among all models. The deeper architecture and sequential processing of CNN-BiLSTM contribute to this overhead, rendering it less practical for real-time use in environments where latency is critical.
Random Forest showed mixed performance. It performed reasonably well in binary classification on NSL-KDD (F1: 0.7880), with strong precision (0.8388), but struggled with multiclass classification (F1: 0.5474), suggesting limitations in handling more granular attack categories.
Its standout feature is speed. With latency as low as 0.0002 ms per sample, Random Forest offers rapid inference, making it attractive for systems where computational resources are constrained or response time is critical. However, its lower predictive performance, especially on multiclass tasks, may limit its utility in high-stakes detection scenarios.
XGBoost emerged as the most balanced model in the study. It delivered strong classification metrics across both datasets and tasks, particularly on UNSW-NB15 (F1: 0.9690, Precision: 0.9709), rivaling CNN-BiLSTM in accuracy while outperforming it in latency.
Its multiclass performance on NSL-KDD (F1: 0.7287) was also robust, and binary classification results were consistent (F1: 0.7701). With latency as low as 0.0001 ms per sample, XGBoost offers exceptional computational efficiency without compromising predictive quality, making it a viable solution for real-time Intrusion Detection Systems.
However, although NSL-KDD and UNSW-NB15 are well-established benchmarks with diverse attack taxonomies enabling reproducible comparisons, neither was specifically designed for IoT scenarios, NSL-KDD reflects legacy enterprise traffic and UNSW-NB15 focuses on general LAN/WAN environments, potentially limiting model generalization to real IoT networks characterized by device heterogeneity, low-power protocols (e.g., MQTT, CoAP), and extreme class imbalance; consequently, the observed minority-class weaknesses may underestimate challenges in actual IoT deployments, where rare protocol-specific attacks (e.g., Mirai botnets) could yield higher false negatives, underscoring the need for future validation on IoT-native datasets despite the current findings positioning CNN-BiLSTM for accuracy-critical contexts and XGBoost for latency-sensitive real-time IDS. Table 3 maps the limitations of previous studies to the key findings and contributions of the present work.

5. Conclusions and Future Work

This work systematically compared a hybrid CNN-BiLSTM model with two lightweight machine-learning baselines (Random Forest and XGBoost) for intrusion detection in IoT environments. The CNN-BiLSTM consistently achieved the highest detection performance, with F1-scores of 0.7680, 0.8949, and 0.9864 supported by strong precision and recall. However, its deep architecture introduced higher inference latency (0.0590–0.1261 ms). XGBoost offered the best balance between accuracy and computational efficiency, maintaining competitive F1-scores (up to 0.9690) and the lowest latency (0.0001–0.0002 ms), while Random Forest provided fast inference but more variable performance, especially for multiclass tasks. Confusion matrix analysis showed that all models struggled with minority attack classes, emphasizing the impact of dataset imbalance. Overall, CNN-BiLSTM is well suited to accuracy-critical IoT environments, whereas XGBoost is more appropriate for latency-sensitive deployments. These findings clarify the trade-offs between predictive capability and efficiency when designing scalable IDS solutions for heterogeneous IoT ecosystems.
Future work should focus on improving model interpretability through explainable AI, enabling privacy-preserving training via Federated Learning, and validating IDS models on realistic IoT-specific datasets such as IoT-23, CIC-IoT2023, and TON_IoT. Addressing class imbalance remains a priority; techniques such as SMOTE, class-weighted loss functions, and other imbalance-aware training strategies could enhance detection of rare and zero-day attacks. Additional optimizations including pruning, quantization, and edge-oriented deployment strategies—may further improve scalability and efficiency in real-world IoT scenarios.

Author Contributions

Conceptualization, V.R.J., K.A.-A., T.A.-H. and S.N.Q.; Methodology, V.R.J., K.A.-A., T.A.-H. and S.N.Q.; Software, V.R.J., K.A.-A., T.A.-H. and S.N.Q.; Validation, V.R.J., K.A.-A., T.A.-H. and S.N.Q.; Formal analysis, V.R.J., K.A.-A., T.A.-H. and S.N.Q.; Investigation V.R.J., K.A.-A., T.A.-H. and S.N.Q.; Resources, V.R.J., K.A.-A., T.A.-H. and S.N.Q.; Data curation, V.R.J., K.A.-A., T.A.-H. and S.N.Q.; Writing—original draft, V.R.J., K.A.-A. and T.A.-H.; Writing—review & editing, K.A.-A., T.A.-H. and S.N.Q.; Visualization, V.R.J., K.A.-A., T.A.-H. and S.N.Q.; Supervision, K.A.-A. and T.A.-H.; Project administration, K.A.-A., T.A.-H. and S.N.Q.; Funding acquisition, T.A.-H. and S.N.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported and funded by the Deanship of Scientific Research at Imam. Mohammad Ibn Saud Islamic University (IMSIU) (grant number IMSIU-DDRSP2504).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The research data supporting this publication are accessible through the project’s GitHub repository https://github.com/kwameassa/Vandit_IDS_Project (accessed on 31 August 2025).

Acknowledgments

The authors appreciate the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University (IMSIU).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript
AIArtificial Intelligence
Bi-LSTMBidirectional Long Short-Term Memory
CNNConvolutional Neural Network
DLDeep Learning
DoSDenial of Service
EDAExploratory Data Analysis
FLFederated Learning
IDSIntrusion Detection System
IoTInternet of Things
MLMachine Learning
NSL-KDDNetwork Security Laboratory Knowledge Discovery and Data Mining dataset
SHAPSHapley Additive exPlanations
XAIExplainable Artificial Intelligence

References

  1. Bar, S.; Prasad, P.W.C.; Sayeed, S. Enhancing Internet of Things Intrusion Detection Using Artificial Intelligence. Comput. Mater. Contin. 2024, 81, 1–23. [Google Scholar] [CrossRef]
  2. Akif, A.; Butun, I.; Williams, A.; Mahgoub, I. Hybrid machine learning models for intrusion detection in IoT: Leveraging a real-world IoT dataset. arXiv 2025, arXiv:2502.12382. [Google Scholar]
  3. Altunay, H.C.; Albayrak, Z. A hybrid CNN + LSTM-based intrusion detection system for industrial IoT networks. Eng. Sci. Technol. Int. J. 2023, 38, 101322. [Google Scholar] [CrossRef]
  4. Gueriani, A.; Kheddar, H.; Mazari, A.C. Enhancing IoT security with CNN and LSTM-based intrusion detection system. In Proceedings of the 2024 6th International Conference on Pattern Analysis and Intelligent Systems (PAIS), El Oued, Algeria, 24–25 April 2024; pp. 1–7. [Google Scholar]
  5. Alabbadi, A.; Bajaber, F. An Intrusion Detection System over the IoT Data Streams Using eXplainable Artificial Intelligence (XAI). Sensors 2025, 25, 847. [Google Scholar] [CrossRef]
  6. Patil, S.; Varadarajan, V.; Mazhar, S.M.; Sahibzada, A.; Ahmed, N.; Sinha, O.; Kumar, S.; Shaw, K.; Kotecha, K. Explainable Artificial Intelligence for Intrusion Detection System. Electronics 2022, 11, 3079. [Google Scholar] [CrossRef]
  7. Arreche, O.; Guntur, T.; Abdallah, M. XAI-IDS: Toward Proposing an Explainable Artificial Intelligence Framework for Enhancing Network Intrusion Detection Systems. Appl. Sci. 2024, 14, 4170. [Google Scholar] [CrossRef]
  8. Baral, S.; Saha, S.; Haque, A. An adaptive end-to-end IoT security framework using explainable AI and LLMs. In Proceedings of the 2024 IEEE 10th World Forum on Internet of Things (WF-IoT), Ottawa, ON, Canada, 10–13 November 2024; pp. 469–474. [Google Scholar] [CrossRef]
  9. Hossain, M.A. Deep learning-based intrusion detection for IoT networks: A scalable and efficient approach. EURASIP J. Inf. Secur. 2025, 2025, 28. [Google Scholar] [CrossRef]
  10. Liang, L.; Tian, Z.; Huang, H.; Li, X.; Yin, Z.; Zhang, D.; Zhang, N.; Zhai, W. Heterogeneous Secure Transmissions in IRS-Assisted NOMA Communications: CO-GNN Approach. IEEE Internet Things J. 2025, 12, 34113–34125. [Google Scholar] [CrossRef]
  11. Jouhari, M.; Guizani, M. Lightweight CNN-BiLSTM based intrusion detection systems for resource-constrained IoT devices. In Proceedings of the 2024 International Wireless Communications and Mobile Computing (IWCMC), Ayia Napa, Cyprus, 27–31 May 2024; pp. 1558–1563. [Google Scholar] [CrossRef]
  12. Sowmya, T.; Anita, E.M. A comprehensive review of AI based intrusion detection system. Meas. Sensors. 2023, 28, 100827. [Google Scholar] [CrossRef]
  13. Faisal, S.; Mubeen, S. Artificial Intelligence-Based Intrusion Detection Systems for IoT Devices. 2024. Available online: https://www.researchgate.net/publication/384321903_Artificial_Intelligence_based_Intrusion_Detection_Systems_for_IoT_Devices (accessed on 15 June 2025).
  14. Abu Laila, D. Responsive Machine Learning Framework and Lightweight Utensil of Prevention of Evasion Attacks in the IoT-Based IDS. STAP J. Secur. Risk Manag. 2025, 2025, 59–70. [Google Scholar] [CrossRef]
  15. Campos, E.M.; Saura, P.F.; González-Vidal, A.; Hernández-Ramos, J.L.; Bernabé, J.B.; Baldini, G.; Skarmeta, A. Evaluating Federated Learning for intrusion detection in Internet of Things: Review and challenges. Comput. Networks 2022, 203, 108661. [Google Scholar] [CrossRef]
  16. Javeed, D.; Saeed, M.S.; Adil, M.; Kumar, P.; Jolfaei, A. A federated learning-based zero trust intrusion detection system for Internet of Things. Ad Hoc Networks 2024, 162, 103540. [Google Scholar] [CrossRef]
  17. Ragab, M.; Ashary, E.B.; Alghamdi, B.M.; Aboalela, R.; Alsaadi, N.; Maghrabi, L.A.; Allehaibi, K.H. Advanced artificial intelligence with federated learning framework for privacy-preserving cyberthreat detection in IoT-assisted sustainable smart cities. Sci. Rep. 2025, 15, 4470. [Google Scholar] [CrossRef]
  18. Liu, H.; Zhang, S.; Zhang, P.; Zhou, X.; Shao, X.; Pu, G.; Zhang, Y. Blockchain and Federated Learning for Collaborative Intrusion Detection in Vehicular Edge Computing. IEEE Trans. Veh. Technol. 2021, 70, 6073–6084. [Google Scholar] [CrossRef]
  19. Rampone, G.; Ivaniv, T.; Rampone, S. A Hybrid Federated Learning Framework for Privacy-Preserving Near-Real-Time Intrusion Detection in IoT Environments. Electronics 2025, 14, 1430. [Google Scholar] [CrossRef]
  20. Alkahla, L.; Hussein, M.K.; Alqassab, A. Feature Selection Techniques in Intrusion Detection: A Comprehensive Review. Iraqi J. Comput. Informatics 2024, 50, 46–53. [Google Scholar] [CrossRef]
  21. Fang, Y.; Yao, Y.; Lin, X.; Wang, J.; Zhai, H. A feature selection based on genetic algorithm for intrusion detection of industrial control systems. Comput. Secur. 2023, 139, 103675. [Google Scholar] [CrossRef]
  22. Bibi, U.; Mazhar, M.; Sabir, D.; Butt, M.F.U.; Hassan, A.; Ghazanfar, M.A.; Khan, A.A.; Abdul, W. Advances in Pruning and Quantization for Natural Language Processing. IEEE Access. 2024, 12, 139113–139128. [Google Scholar] [CrossRef]
  23. Alshuaibi, A.; Arshad, M.W.; Maayah, M. A Hybrid Genetic Algorithm and Hidden Markov Model-Based Hashing Technique for Robust Data Security. J. Cyber Secur. Risk Audit. 2025, 2025, 42–56. [Google Scholar] [CrossRef]
  24. Alshinwan, M.; Memon, A.; Ghanem, M.; Almaayah, M. Unsupervised text feature selection approach based on improved Prairie dog algorithm for the text clustering. Jordanian J. Informatics Comput. 2025, 2025, 27–36. [Google Scholar] [CrossRef]
  25. Defcom. NSL-KDD Dataset. 2009. Available online: https://github.com/defcom17/NSL_KDD (accessed on 1 July 2025).
  26. UNSW Canberra Cyber Range Lab. UNSW-NB15 Dataset. 2015. Available online: https://research.unsw.edu.au/projects/unsw-nb15-dataset (accessed on 10 July 2025).
  27. Nair, V.; Hinton, G. Rectified linear units improve restricted Boltzmann machines. In Proceedings of the 27th International Conference on International Conference on Machine Learning(ICML-10), Haifa, Israel, 21–24 June 2010; pp. 807–814. [Google Scholar]
  28. Boureau, Y.-L.; Ponce, J.; LeCun, Y. A theoretical analysis of feature pooling in visual recognition. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), Haifa, Israel, 21–24 June 2010; pp. 111–118. [Google Scholar]
  29. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  30. Schuster, M.; Paliwal, K.K. Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 1997, 45, 2673–2681. [Google Scholar] [CrossRef]
  31. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
Figure 1. Flow Chart of CNN-BiLSTM.
Figure 1. Flow Chart of CNN-BiLSTM.
Sensors 25 07564 g001
Figure 2. Distribution of attack types in training dataset.
Figure 2. Distribution of attack types in training dataset.
Sensors 25 07564 g002aSensors 25 07564 g002b
Figure 3. Top 20 Protocol Types.
Figure 3. Top 20 Protocol Types.
Sensors 25 07564 g003
Figure 4. CNN-BiLSTM confusion matrix.
Figure 4. CNN-BiLSTM confusion matrix.
Sensors 25 07564 g004
Figure 5. Random Forest confusion matrix.
Figure 5. Random Forest confusion matrix.
Sensors 25 07564 g005
Figure 6. XGBoost model confusion matrix.
Figure 6. XGBoost model confusion matrix.
Sensors 25 07564 g006
Figure 7. Confusion matrix for CNN-BiLSTM.
Figure 7. Confusion matrix for CNN-BiLSTM.
Sensors 25 07564 g007
Figure 8. Random Forest confusion matrix.
Figure 8. Random Forest confusion matrix.
Sensors 25 07564 g008
Figure 9. XGBoost confusion matrix.
Figure 9. XGBoost confusion matrix.
Sensors 25 07564 g009
Figure 10. Model accuracy for multiclass attacks.
Figure 10. Model accuracy for multiclass attacks.
Sensors 25 07564 g010
Figure 11. Model accuracy on Binary class attacks.
Figure 11. Model accuracy on Binary class attacks.
Sensors 25 07564 g011
Figure 12. Model loss for multiclass attacks.
Figure 12. Model loss for multiclass attacks.
Sensors 25 07564 g012
Figure 13. Model loss for binary class attack.
Figure 13. Model loss for binary class attack.
Sensors 25 07564 g013aSensors 25 07564 g013b
Table 1. Comparative Analysis of Related Studies on IoT Intrusion Detection.
Table 1. Comparative Analysis of Related Studies on IoT Intrusion Detection.
StudyObjectiveKey Contribution/Gap AddressedLimitations
Alabbadi & Bajaber (2025) [5]Develop XAI-enhanced IDS for real-time IoT data streamsImproved interpretability using SHAP/LIME on streaming dataNo scalability testing; not evaluated on edge devices; missing methodological details prevent replication or deeper analysis
Patil et al., 2022 [6]; Arreche et al., 2024 [7]Integrate XAI into ensemble IDS modelsTransparent decision-making via feature attributionIgnores latency and resource overhead in constrained environments; missing methodological details prevent replication or deeper analysis
Altunay & Albayrak (2023) [3]; Gueriani et al., 2024 [4]High-accuracy hybrid CNN + LSTM for attack classificationStrong performance on UNSW-NB15/NSL-KDDNo latency analysis; not optimized for edge deployment; missing methodological details prevent replication or deeper analysis
Jouhari & Guizani (2024) [11]Lightweight CNN-BiLSTM for resource-constrained devicesReduced model size with acceptable accuracyStruggles with complex, multi-stage attacks and missing methodological details prevent replication or deeper analysis
Abu Laila (2025) [14]Evasion-resistant lightweight IDS with adversarial robustness94.7% accuracy, only 15% overhead versus baselineLimited to binary classification, no multiclass evaluation, and missing methodological details prevent replication or deeper analysis
Campos et al., 2022 [15]; Javeed et al., 2024 [16]Privacy-preserving IDS via Federated LearningHigh accuracy with no raw data sharingHigh communication overhead, convergence instability, and missing methodological details prevent replication or deeper analysis
Rampone et al., 2025 [19]Near-real-time FL with blockchain integrationEnhanced privacy and auditabilityIncreased system complexity; scalability concerns
Alkahla et al., 2024 [20]; Fang et al., 2023 [21]Feature selection for IDS efficiencyReduced dimensionality, improved accuracyNot integrated into end-to-end lightweight models, missing methodological details prevent deeper analysis
Alshuaibi et al., 2025 [23]GA + HMM-based secure hashing for data integrityStrong resistance to dictionary/brute-force attacksComplementary (not core IDS); adds preprocessing overhead; missing methodological details prevent deeper analysis
Alshinwan et al., 2025 [24]Unsupervised feature selection via PDOEfficient clustering in high-dimensional datamissing methodological details prevent deeper analysis
Liang et al., 2025 [10]Secure NOMA transmission using CO-GNNOptimized secrecy rate in IRS-assisted networksNot an IDS; focuses on physical-layer security
Table 2. Experimental Model Benchmark comparison.
Table 2. Experimental Model Benchmark comparison.
ModelDatasetF1-ScoreRecallPrecisionAccuracyLatency per Sample (ms)
CNN-BiLSTMNSL-KDD Multiclass attack0.76800.78620.81510.78620.0668
NSL-KDD Binary class attack0.89490.89440.90340.89440.0590
UNSW-NB150.98640.98640.98680.98640.1261
Random ForestNSL-KDD Multiclass attack0.54740.63740.49940.63740.0008
NSL-KDD Binary class attack0.78800.78950.83880.78950.0002
UNSW-NB150.73180.75530.83060.75530.0002
XGBoostNSL-KDD Multiclass attack0.72870.77690.79190.77690.0002
NSL-KDD Binary class attack0.77010.77300.83510.77300.0001
UNSW-NB150.96900.96890.97090.96890.0001
Table 3. Addressing Limitations of Prior IoT-IDS Research: Key Findings and Contributions of the Present Study.
Table 3. Addressing Limitations of Prior IoT-IDS Research: Key Findings and Contributions of the Present Study.
Limitations of Previous StudiesKey Findings of Our StudyContributions of Our Study
Lack of latency benchmarking and edge deployment analysisXGBoost achieves F1 = 0.9690 (UNSW-NB15) with 0.0001 ms/sample latencyLow-latency XGBoost model suitable for real-time edge inference on constrained IoT devices
High computational overhead of deep models; no accuracy–latency trade-off analysisCNN-BiLSTM delivers F1 = 0.9864 (UNSW-NB15 binary), 0.8949 (NSL-KDD binary) at 0.059–0.126 ms/sampleFully documented hybrid CNN-BiLSTM architecture with equations, tuning, and performance metrics for complex attack detection
No systematic comparison between deep and lightweight modelsAll models evaluated across multiclass and binary tasks on NSL-KDD and UNSW-NB15 with confusion matrices, EDA, and feature importanceTri-model benchmarking framework (CNN-BiLSTM, Random Forest, XGBoost) enabling performance-centric IDS design
Absence of context-aware deployment guidelines for heterogeneous IoTXGBoost offers best accuracy–speed balance; CNN-BiLSTM excels in accuracy-critical scenariosPerformance-centric deployment strategy: Use CNN-BiLSTM in accuracy-critical gateways, XGBoost in latency-sensitive sensors
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Joshi, V.R.; Assa-Agyei, K.; Al-Hadhrami, T.; Qasem, S.N. Hybrid AI Intrusion Detection: Balancing Accuracy and Efficiency. Sensors 2025, 25, 7564. https://doi.org/10.3390/s25247564

AMA Style

Joshi VR, Assa-Agyei K, Al-Hadhrami T, Qasem SN. Hybrid AI Intrusion Detection: Balancing Accuracy and Efficiency. Sensors. 2025; 25(24):7564. https://doi.org/10.3390/s25247564

Chicago/Turabian Style

Joshi, Vandit R, Kwame Assa-Agyei, Tawfik Al-Hadhrami, and Sultan Noman Qasem. 2025. "Hybrid AI Intrusion Detection: Balancing Accuracy and Efficiency" Sensors 25, no. 24: 7564. https://doi.org/10.3390/s25247564

APA Style

Joshi, V. R., Assa-Agyei, K., Al-Hadhrami, T., & Qasem, S. N. (2025). Hybrid AI Intrusion Detection: Balancing Accuracy and Efficiency. Sensors, 25(24), 7564. https://doi.org/10.3390/s25247564

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop