You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

13 September 2024

Deep Complex Gated Recurrent Networks-Based IoT Network Intrusion Detection Systems

,
,
,
,
and
1
Department of Computer Science, Faculty of Computers & Artificial Intelligence, University of Sadat City, Sadat City 32897, Egypt
2
Department of Information Technology, Faculty of Computers & Information Systems, Damanhour University, Damanhour 22511, Egypt
3
Computer Science Department, Faculty of Computers and Information, Kafrelsheikh University, Kafrelsheikh 33516, Egypt
4
Department of Computer Engineering, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia
This article belongs to the Section Internet of Things

Abstract

The explosive growth of the Internet of Things (IoT) has highlighted the urgent need for strong network security measures. The distinctive difficulties presented by Internet of Things (IoT) environments, such as the wide variety of devices, the intricacy of network traffic, and the requirement for real-time detection capabilities, are difficult for conventional intrusion detection systems (IDS) to adjust to. To address these issues, we propose DCGR_IoT, an innovative intrusion detection system (IDS) based on deep neural learning that is intended to protect bidirectional communication networks in the IoT environment. DCGR_IoT employs advanced techniques to enhance anomaly detection capabilities. Convolutional neural networks (CNN) are used for spatial feature extraction and superfluous data are filtered to improve computing efficiency. Furthermore, complex gated recurrent networks (CGRNs) are used for the temporal feature extraction module, which is utilized by DCGR_IoT. Furthermore, DCGR_IoT harnesses complex gated recurrent networks (CGRNs) to construct multidimensional feature subsets, enabling a more detailed spatial representation of network traffic and facilitating the extraction of critical features that are essential for intrusion detection. The effectiveness of the DCGR_IoT was proven through extensive evaluations of the UNSW-NB15, KDDCup99, and IoT-23 datasets, which resulted in a high detection accuracy of 99.2%. These results demonstrate the DCG potential of DCGR-IoT as an effective solution for defending IoT networks against sophisticated cyber-attacks.

1. Introduction

Globally, the Internet of Things (IoT) is expanding quickly due to rising consumer expectations, a growing range of products, and advanced technology. Our everyday lives now involve billions of gadgets that are connected thanks to this proliferation; by 2023, Gartner predicts that there will be about 25 billion connected devices [1]. IoT technology offers advantages such as smarter solutions and improved daily activities, but these are outweighed by serious privacy and security issues. Managing a variety of linked devices, complexities, conflicting trends, and diversities is necessary while developing Internet of Things solutions. Conventional security measures are insufficient for the extended sessions that are common in Internet of Things contexts since they were created for powerful computers and brief sessions. These elements make Internet of Things (IoT) devices appealing targets for cybercriminals, putting our lives in danger. Developing efficient security solutions based on “lightweight” and “adaptive” concepts is a pragmatic way to handle IoT complexity. Large-scale distributed networks have shown success using “Adaptive Lightweight” techniques in managing anomalies. However, because there are so many IoT devices on a network, it is not feasible to develop unique security solutions for every item.
Securing data shared between devices in an IoT network, on the other hand, could be a realistic technique. To generate adaptive solutions for the IoT system, artificial intelligence can be utilized to analyze a wide range of data quantities and types. Current security techniques are only appropriate for powerful machines and short-term sessions, and the same protection strategy cannot be employed for long-running sessions. As a result, IoT devices have become appealing targets for hackers, placing our lives in danger from unknown risks. Designing effective security solutions based on the principles of “lightweight” and “adaptation” may be a viable method for handling the complexity of the IoT. “Adaptive Lightweight” approaches have proven helpful in dealing with abnormalities in big, distributed networks. The sheer quantity of IoT devices on a network makes developing a security solution for each one difficult. However, protecting the data transmitted between IoT devices could be a viable approach.
Artificial neural networks with layers of artificial neurons are subjected to deep learning. As data move from the first layer, known as the input layer, via intermediate (hidden) layers to the last layer, known as the output layer, it is processed in such a network. An artificial neural network that has multiple intermediate layers is referred to as deep.
The activation function is the one in which neurons are represented. The neuron’s output is the outcome of applying the activation function to the scalar product of the input vector and the neuron’s weight vector, shifted by a specified distance [2]. The activation function’s value is determined by the weighted sum of the neuron’s inputs and the threshold value. Useful instances of nonlinear functions are the sigmoid, SoftMax function, hyperbolic tangent, and rectified linear unit (ReLU), which are a few nonlinear functions that have been utilized as activation functions.
A loss function, often known as error, is used to describe the discrepancy between the target variable’s actual value and the value the neural network predicts during training.
Neural network topologies vary according to the arrangements of neurons, layers, and connections between them. Thus, the neural network design that is employed can be utilized to categorize deep learning techniques.
Furthermore, deep learning techniques can be categorized into four primary types, like classic machine learning techniques: supervised, semi-supervised, unsupervised, and reinforced deep learning techniques. Hybrid approaches are made possible by the capacity to combine various neural network topologies into intricate ones within a single network.
The primary deep-learning techniques that surfaced throughout the examination of pertinent literature are briefly described here.
Artificial intelligence (AI) can deliver adaptable solutions for Internet of Things (IoT) systems by analyzing a variety of data types and quantities. By analyzing enormous volumes of IoT data, machine learning and data analytics are already being utilized to improve network performance and customer service. To spot potentially harmful behavior and stop anomalous activity, IoT networks can use behavioral analytics, anomaly detection, and pattern recognition. To characterize behaviors as “malware attack” or “normal”, this study presents a novel multi-layer architecture for an Internet of Things system that makes use of deep learning algorithms to forecast network data. This study makes use of two intrusion detection datasets—UNSW-NB15, KDDCup99, and IoT-23 datasets—which are commonly employed in machine learning studies on network data security. Because IoT datasets are high-dimensional, temporal, and multi-modal in nature, they pose considerable hurdles for analysis and detection. Compared to classic machine learning techniques, deep learning algorithms are more suited to complicated IoT datasets due to their huge processing capacity. Even while deep learning research on IoT security is still in its infancy, it has enormous promise for gleaning information from IoT data. The researchers think that IoT systems can be intelligently optimized using deep learning techniques. Recurrent neural networks, for instance, can process and store data at each stage before feeding it into the next because they can learn from prior time steps of incoming input. Regardless of the complexity of the neural network, hyperparameters can be changed to obtain lightweight functionality for Internet of Things solutions. This study aims to assess the efficacy of recurrent neural network techniques on intrusion detection datasets and investigate deep learning’s potential in cybersecurity and IoT. It is driven by the idea of applying deep learning to IoT network security. In the linked world of today, maintaining security necessitates evaluating enormous amounts of heterogeneous data—something that can only be completed with AI support.
Using real-time Internet of Things data, the researchers assessed several deep learning algorithms, such as recurrent neural networks, autoencoders, convolutional neural networks, and deep neural networks, to tackle this problem. The study issue finally led them to concentrate on recurrent neural networks because it required an algorithm that could learn from past data.
We evaluated various versions of gated recurrent unit (GRU), including bi-directional and multi-layer GRU, as well as CNN to devise efficient and lightweight solutions for the IoT network and obtain optimal results on the dataset. Most IDS-ML researchers utilize the UNSW-NB15, KDDCup99, and IoT-23 datasets for IoT network intrusion detection data. To make the data more suited for an IoT network, we layered it.
It has become impractical to rely on traditional intrusion detection systems (IDS) based on attack signatures to identify and block cyber-attacks against IoT infrastructures [3] due to the continual growth of cyber-attacks. While signature-based IDSs can accurately detect attacks that match pre-stored intrusion patterns, such as sequences of system calls or network traffic patterns, a problem arises when a new attack is discovered. Because signature-based IDSs require prior knowledge of an attack’s prospective signature, they are susceptible to new and unknown attacks. Traditional IDSs can only detect an attack if it already exists in their database, leaving them vulnerable to new and unknown threats. To address this limitation, anomaly-based intrusion detection systems (IDSs) have been proposed, which employ more intelligent methodologies than conventional IDSs. Anomaly-based IDSs establish a profile of “normal” behavior and then detect deviations, enabling them to detect potential new attacks (zero-day attacks). However, in a dynamic environment such as an IoT ecosystem, anomaly-based IDSs struggle to accurately detect all new intrusions, and the cost of false positives remains significant. Numerous zero-day assaults are undetected because of the limitations of IoT devices and conventional anomaly detection methods. Typically, critical defense mechanisms, such as a network intrusion detection system (NIDS), are used to analyze network traffic for abnormal behavior and facilitate such functions [4]. Assessing network traffic with anomaly-based IDSs can be used in conjunction with traditional cyber defense systems, such as firewall systems [5], to establish a secure environment and network. Using pre-trained models, anomaly-based IDSs can distinguish between benign and malicious traffic. This research paper is interdisciplinary, encompassing cyber security, artificial intelligence, and computer networks; consequently, a substantial quantity of effort has been devoted to comprehending the nuances of each subject. Our initial objective was to study the various attack categories present in an intrusion detection dataset. Then, we realized that deep learning techniques were required to classify the dataset for intrusion detection. In addition, we investigated IoT architecture and evaluated machine learning techniques that met IoT specifications. Signature-based methods, specification-based approaches, anomaly-based approaches, and hybrid strategies are the four primary categories of ID attacks [6]. Signature-based techniques entail comparing a set of network data to a feature database for similarities. If the scanned data match the database of signatures, they are deemed unlawful. This method is useful for precisely distinguishing the type of attack and permits system administrators to define rules and thresholds beforehand. The established criteria are then used to determine the current device and network status. If the threshold is exceeded or the rules are violated, the IDS will detect an abnormal condition and act accordingly [7].
The following are the main contributions of our paper.
  • An extensive analysis of pertinent studies in this field as well as an investigation of the importance of temporal and spatial characteristics in network data for ID intrusion detection.
  • The release of the DCGR_IoT model, which improves intrusion detection capabilities by integrating the temporal and spatial components of network data.
  • The creation of the DCGR_IoT model consists of two primary parts: a convolutional neural network (CNN)-based feature extraction module and a CGNN-based temporal feature extraction module. Notably, the DCGR_IoT improves the attention between local and global features as well as across various features by incorporating multi-object self-attention from the transformer architecture. Furthermore, to alleviate CNN’s restriction, a bidirectional memory network was integrated to preserve the temporal aspects of network traffic data.
  • Furthermore, to overcome CNN’s constraint in extracting temporal information, a bidirectional memory network was integrated to preserve the temporal properties of network traffic data.
Thorough testing of the suggested strategy using simulation tests was carried out with the Pytorch framework and Python programming language on the UNSW-NB15, KDDCup99, and IoT-23 datasets. Additionally, comparative experiments were performed, which showed that our model outperformed the other methods in terms of detection accuracy on both datasets. The effectiveness of building and refining large-scale intrusion detection systems (IDS) in an IoT environment is supported by these results.
The structure of this paper is as follows: Section 2 describes the fundamental architecture and related works for network intrusion detection. Our proposed technique, the deep complex gated recurrent network-based IoT network intrusion detection system (DCGR_IoT), is in Section 3. The fourth section introduces the verification and analysis of the efficacy and precision of the proposed model. Section 5 concludes our paper with a summary.

3. The Deep Neural Learning-Based Approach for IoT Anomaly Detection (DCGR_IoT)

There are three phases in the DCGR_IoT strategy described in this section:
The first phase is preprocessing the dataset to prepare the dataset and meet the input requirements. The original characteristics must be reshaped and encoded using a multilayer convolutional neural network before being fed into the model. To keep the information flowing through the network and stop gradient fading, residual links were introduced.
The second phase for temporal-spatial feature extraction is CNN-assisted spatial feature extraction: CNN is utilized to acquire spatial characteristics and temporal feature extraction with CGRNs: multiread self-attention is employed with multiple transformer encoders to extract greater feature information. To obtain temporal feature representations among the features, the temporal relationship between the pre- and post-features was mined using several CGRNs blocks as mentioned Section 4.
The combining features and categorization in the third phase, the acquired spatial and temporal characteristics, were combined through concatenation for the classification of intrusion attacks. The categorization module receives the fused features. Using the SoftMax function, several completely connected layers were used to perform the final classification.

3.1. The Pre-Processing Dataset

The dataset is encoded and structured to comply with the model’s requirements after the transform has been adjusted. Next, the dataset is split into three sections: with the sets for training and validation, and testing configuration, the testing set is used to assess the model’s performance, whilst the training and validation sets are used to train the model.
After encoding and normalization, the dataset is prepared for DCGR_IoT model training. However, depending on the CNN model, further preparation is needed:
  • A 1D vector must be created from the dataset to use it in a 1D CNN model.
  • The processed data must be transformed into a matrix format for a 2D CNN model.
  • With these changes, the data are guaranteed to be in the right format for any kind of CNN model.

3.2. Convolutional Neural Networks (CNN) for Feature Selection

The DCGR_IoT system mainly uses CNN for spatial feature extraction. This aids in separating crucial characteristics from irrelevant ones in network traffic, helping to remove unnecessary information. Not all layers in a convolutional neural network (CNN) are fully connected, and neurons have limited connections, being linked to only a tiny portion of the preceding layer. Convolutional layers that are stacked make it easier to abstract low-level characteristics into high-level features. A convolutional layer is made up of a sliding filter that is applied across the input data’s dimensions. The filter output is passed into a non-linear activation function, which produces activation maps. The network modifies the filters during operation so that activations happen when the filter passes over favorable traits, allowing it to learn and improve its performance [32]. The size of the filter in a convolutional layer can vary, although 3 × 3 is a popular size. A fully connected layer, also known as a dense or linear layer, is made up of neurons that are linked to all neurons in the layer above. These layers are used to build an ANN. The pooling layer subsamples the previous layer’s output, lowering dimensionality and the number of learned parameters. When paired with a convolutional layer, the pooling layer gives translation invariance to the network, as shown in Figure 1.
Figure 1. The architecture convolution neural network in DCGR_IoT.
A dropout layer mitigates overfitting by deleting data at random with a specified probability. It also minimizes the preceding layer’s bias towards weights, and its effects are amplified when the dataset is small. The activation function f ( x ) in rectified linear units (ReLU) accelerates the network’s training process. The sparsity of the activation maps is also influenced by ReLU, in Equation (1).
f ( x ) = m a x ( x ;   0 )
Batch normalization is a technique for normalizing variables in relation to the current batch that can be applied to either direct inputs or prior layer activations. It speeds up network training, improves regularization, and lowers network generalization errors. Skip connections establish new connections between neurons in different levels that may not be connected sequentially. Skip connections improve network training and minimize linear dependence between neurons by disrupting the symmetry of the neural network, avoiding singularities in the loss landscape. Trainable parameters in a neural network are those that change during training, whereas nontrainable parameters do not change. By dividing the dataset depending on these proportions, a stratified split of a dataset retains the normal distribution of classes. Cisco NetFlow is a network flow representation protocol [29]. It contains data characteristics and bidirectional flow information retrieved from network packet headers that are unrelated to the content.

3.3. Complex Gated Recurrent Network

A more intricate and spatial representation of network traffic is made possible by the construction of multidimensional feature subsets using complex gated recurrent networks (CGRNs). This helps much more in locating and removing unnecessary data.
A gated recurrent network (GRN) is a sort of recurrent neural network (RNN) that uses gating methods to selectively transfer and update information across the network. Unlike typical RNNs, which use a simple recurrent connection between hidden units, GRNs govern the flow of input using more complicated gating units.
Cho et al. [33] introduced the gated recurrent unit (GRU) as the most popular architecture for GRNs in 2014. To govern information flow, GRUs employ two gating units: the update gate and the reset gate. The update gate decides how much of the prior hidden state to keep and how much of the new input to add, whereas the reset gate decides how much of the previous hidden state to forget.
Another widely used GRN architecture is the Long Short-Term Memory (LSTM) network, which was first defined by Hochreiter and Schmidhuber in 1997. LSTMs, like GRUs, use gating mechanisms to selectively send and update data across networks, but their architecture is more complex, with separate input, output, and forget gates. GRNs are especially useful for modeling long-term dependencies because they allow the network to selectively recall and forget information over time, as seen in Figure 2. They have been used in a variety of fields, such as natural language processing, speech recognition, and image captioning.
Figure 2. The architecture of complex gated recurrent network.
Deep complex gated recurrent networks (Deep Complex GRUs) combined with convolutional neural networks (CNN) can be used to create a powerful model for processing complex-valued data and capturing spatial and temporal dependencies. In this setup, the CNN is used for feature extraction, while the Deep Complex GRU processes the extracted features and captures temporal dependencies. Here is a description of the combined model with equations:

3.3.1. CNN for Feature Extraction

Equation (2) represents a 2D convolution operation, where the input feature map is convolved with the convolution filter, and the result is passed through an activation function to produce the output feature map.
  • Convolutional layer:
    y i , j = α m = 0 M 1   n = 0 N 1   w m , n · x i + m , j + n + b
    In Equation(2), where x i , j is the input feature map, w m , n are the convolutional weights of m×n and b: Bias term. y i , j : Output feature at location ( i , j ) in the output feature map. α : Activation function (typically a non-linear function such as ReLU, tanh, or sigmoid). x i + m , j + n : Input feature values at locations ( i + m , j + n ) in the input feature map.
  • Pooling layer:
    The equation describes the max pooling operation in a neural network. The max pooling operation is used to down sample the input feature map by selecting the maximum value within a specified window size. The equation is as follows:
    y i , j = max m = 0 k 1 max n = 0 l 1 x i + m , j + n
    In Equation (3), where y i , j is the output feature map, and the max operation is performed over a specified window size (e.g., 2 × 2), y i , j is the output feature value at location ( i , j ) in the output feature map after pooling. x i + m , j + n : Input feature values within the pooling window, where m and n range over the dimensions of the pooling window of size k l , where k : Height of the pooling window and l : Width of the pooling window.
  • Flattening:
    The output feature maps from the CNN to create a 1D vector that can be fed into the Deep Complex GRU.

3.3.2. Deep Complex GRU for Processing Extracted Features and Capturing Temporal Dependencies

  • Update Gate:
    The update gate controls how much of the previously hidden state h t 1 is carried forward to the current time step in Equation (4). It uses the sigmoid activation function to squash the value between 0 and 1.
    z t = σ W z x t + U z h t 1 + b z
  • Reset Gate:
    The reset gate controls how much of the previous hidden state h t 1 is forgotten or “reset” at the current time step. It also uses the sigmoid activation function in Equation (5).
    r t = σ W r x t + U r h t 1 + b r
  • Candidate Hidden State:
    The candidate hidden state h ˜ t in Equation (6) is an intermediate representation computed at each time step in a gated recurrent neural network for GRU. It represents the new information that could potentially be added to the hidden state.
    h ˜ t = tanh W x t + r t U h t 1 + b
  • Hidden State:
    h t = 1 z t h t 1 + z t h ˜ t
    The hidden state h t is the final output of a gated recurrent neural network at time step t . It is computed as a weighted combination of the previous hidden state t 1 and the candidate hidden state h t .
    z t : The update gate value at time step t 1
    h t 1 : The previous hidden state
    h t : The candidate hidden state at time step t
The update gate controls how much of the previous hidden state h t 1 is carried forward and how much of the new candidate hidden state h t is incorporated. When z t is close to 0, the hidden state h t 1 will be mostly determined by the previously hidden state h t 1 . When z t is close to 1, the hidden state will be mostly determined by the candidate’s hidden state h t .
In these equations, the variables are defined as follows:
x t : Input at time step t (extracted features from the CNN)
h t : Hidden state at time step t
h t 1 : Hidden state at time step t 1
W ,   U ,   W z ,   U z ,   W r , U r : Weight matrices
b ,   b z , b r : Bias terms
* : Complex-valued matrix multiplication
σ : Sigmoid activation function
t a n h : Hyperbolic tangent activation function
By combining CNN for feature extraction and Deep Complex GRU for processing and capturing temporal dependencies, the model can effectively handle complex-valued data and provide a powerful solution for various applications, such as IoT network intrusion detection systems.
Complex gated recurrent networks (CGRNs) are an extension of the gated recurrent unit (GRU) model, designed to handle complex-valued data, which is particularly useful in scenarios where phase information is critical, such as in signal processing or communication systems. In the context of DCGR_IoT, CGRNs are used to construct multi-space feature subsets and to extract crucial features for intrusion detection.
Let h t represent the hidden state at time t , and x t represents the input at time t . The CGRN updates its hidden state using two gates: the reset gate r t and the update gate z t .
The reset gate determines how much past information to forget. It is computed as follows:
In the context of complex-valued data, the complex gated recurrent network (CGRN) updates its hidden state using two gates: the reset gate r t and the update gate z t . Here is how the computations are adapted for complex-valued data:
  • Reset Gate r t :
    r t = σ W x r x t + W h r h t 1 + b r
    where σ is a complex-valued activation function (e.g., complex sigmoid or complex softmax), W x r and W h r are complex-valued weight matrices for the input and hidden states, respectively, and b r is a complex-valued bias term.
  • Update Gate z t :
    z t = σ W x z x t + W h z h t 1 + b z
    with W x z   W h z and being complex-valued weight matrices, and b z being a complex-valued bias term.
  • The Hidden State Update where the new memory content n t is computed by combining the reset gate with the previous hidden state and the current input: New Memory Content n t :
    n t = t a n h W x n x t + r t W h n h t 1 + b n
    where t a n h is replaced with a complex-valued hyperbolic tangent function, W x n and W h n are complex-valued weight matrices, and b n is a complex-valued bias term. The element-wise multiplication is performed in the complex domain.
  • Final Hidden State h t :
    h t = 1 z t h t 1 + z t n t
Here, the element-wise multiplication and addition are performed in the complex domain. The final hidden state h t is also complex-valued.
By adapting these computations to the complex domain, CGRNs can effectively process complex-valued data, capturing both the magnitude and phase information. This makes them suitable for applications where complex-valued data are inherent, such as in signal processing, communication systems, and certain machine-learning tasks involving complex features.
In the context of complex-valued data, all operations are performed in the complex domain. The sigmoid function is replaced with a complex-valued activation function, and the weight matrices and bias terms are complex-valued as well. The element-wise multiplication and addition operations are also performed in the complex domain.
The CGRN’s ability to handle complex-valued data allows it to capture phase information and construct richer feature representations, which is crucial for identifying intricate intrusion patterns in IoT networks.
The combination of convolutional neural networks (CNNs) [34] and gated recurrent networks (GRNs) has been used in IoT anomaly detection applications to effectively capture both spatial and temporal features in sensor data.
The CNN is utilized in this approach to extract spatial characteristics from sensor data [35], while the GRN is used to model temporal dependencies in the data. CNN is commonly employed as the model’s front end, with numerous convolutional layers followed by pooling and normalization layers. The CNN output is then passed into the GRN, which processes the feature map sequence generated by the CNN. The GRN can be implemented using various architectures of GRUs which are used to model the temporal dependencies in the data. The output of the GRN is then fed into a fully connected layer, which performs classification to detect anomalies in the data.
One advantage of using CNNs and GRNs is that they can handle high-dimensional sensor data and capture complex patterns in the data. Additionally, the use of GRNs allows the model to effectively handle sequences with long-term dependencies, which is important for anomaly detection in IoT applications where anomalies may occur over a long period of time as shown in Figure 3.
Figure 3. The architecture proposes DCGR_IoT.
Overall, the combination of CNNs and GRNs has shown promising results for IoT anomaly detection applications and has the potential to improve the accuracy and efficiency of anomaly detection systems.
To evaluate the DCGR_IoT system on UNSW-NB15, KDDCup99, and IoT-23 datasets, the steps for the DCGR_IoT A approach are as follows.
  • Phase 1: Data preprocessing: Loading and preprocessing the UNSW-NB15, KDDCup99, and IoT-23 datasets. This may include normalizing data, handling missing values, and converting categorical variables to numerical representations. Ensure that the data are consistent with the Deep Complex GRU model.
  • Phase 2: Feature Extraction: Extract relevant features from pre-processed datasets using convolutional neural networks (CNN) as mentioned in the DCGR_IoT system. This step helps in selecting important features and filtering out unnecessary data to improve computational efficiency.
  • Phase 3: Model Training: Train the Deep Complex GRU (CGRN) model on the extracted features using a labeled dataset containing examples of normal and malicious network traffic. The model will learn to recognize patterns and temporal dependencies in the data that indicate intrusion attempts.
  • Phase 4: Model Evaluation: Evaluate the performance of the DCGR_IoT system using a separate validation dataset for both the UNSW-NB15, KDDCup99, and IoT-23 datasets. This step helps ensure that the model generalizes well to unseen data and can effectively detect intrusions.
  • Final stage:
Performance metrics: Calculate the detection accuracy, precision, recall, and F1 score of the DCGR_IoT system on the two datasets. Compare the results with other modern intrusion detection systems to prove the effectiveness of the proposed system.
Analysis: Analyze the results and identify any potential areas for improvement or further research. This may include exploring different architectures, hyperparameters, or feature extraction techniques to improve the performance of the DCGR_IoT system.
By following these steps, you can evaluate the DCGR_IoT system on the UNSW-NB15, KDDCup99, and IoT-23 datasets and demonstrate its potential as an effective solution for defending IoT networks against sophisticated cyberattacks. The high detection accuracy of 99.2% mentioned in the description indicates that the DCGR_IoT system can be a promising approach for IoT network intrusion detection.
  • Preprocessing: The sensor data are preprocessed to fill in missing values and normalize the data.
  • Feature selection: The preprocessed sensor data are fed into a CNN to extract spatial features from the sensor data. Feature selection in the proposed framework for intrusion detection involves identifying and selecting the most important features from the dataset to improve model performance and reduce computational costs. The method used for feature selection is a correlation-based filter approach, where features are chosen based on their correlation scores. Features with a correlation score higher than a specified threshold are considered highly correlated, and redundant features are removed to streamline the dataset. This process not only enhances the model’s accuracy but also decreases the training time by reducing the number of input variables. Once the dataset was normalized, we used feature selection algorithms to identify significant features. Among the different possibilities, we used the correlation-based filter method. This method improves storage economy while reducing computation time. We chose features according to their correlation scores by applying a correlation-based filter approach. We used the Pearson correlation approach, as shown in Equation (2) to evaluate the correlations between the attributes. Features with values close to −1 or 1 are highly correlated, whereas those near 0 are uncorrelated, according to the Pearson correlation coefficient, which ranges from −1 to 1. Features with high correlation values were considered redundant. We handle this by designating features as redundant if their correlation values are greater than a certain threshold, which we set as 0.95. Therefore, we eliminated one feature from each group of redundant features. During model building, we applied correlation for feature selection in UNSW-NB15, KDDCup99, and IoT-23 datasets’ features.
  • Sequence generation: The output of the CNN is then fed into a GRU to generate a sequence of feature maps, which capture the temporal dependencies in the data.
  • Classification: The output of the GRU is then fed into a fully connected layer, which performs classification to detect anomalies in the data.
  • Training: The DCGR_IoT are trained on a labeled dataset of sensor data, using backpropagation and gradient descent to optimize the model parameters.
  • Evaluation: The performance of the DCGR_IoT model is evaluated using various metrics, such as accuracy, precision, recall, and F1-score.
  • Deployment: The trained model is deployed in a real-world IoT environment, where it continuously monitors sensor data and detects anomalies in real time.

4. Implementations and Experiments

In this section, we showcase the DCGR_IoT model for conducting anomaly detection experiments on an IoT dataset. We assess the model’s performance in detecting anomalies within IoT network intrusion detection systems and compare it with other widely used anomaly detection algorithms to highlight its superiority. Furthermore, to validate the effectiveness of our proposed improvement method, we perform ablation experiments, comparing and analyzing the model’s anomaly detection performance before and after the enhancements. The outcomes of these experiments serve to further confirm the feasibility and effectiveness of our DCGR_IoT model and its improved method in practical applications. The experiments utilized Python’s TensorFlow library. After conducting multiple experimental validations, the model hyperparameters were set as follows: learning rate = 0.001, optimizer = SGD, and number of epochs = 100, Weight Decay: 0.0001, Dropout Rate: 0.01. Hyper Parameter tuning in DCGR_IoT model is CNN1 (Convolution layers 64 filters 3 × 3), CNN1 (Convolution layers 32 filters 3 × 3) GRU1 and GRU1 Hidden Units [64, 128], The Dropout [0.0, 0.1], The Activation “relu”as shown in Figure 4.
Figure 4. The flowchart proposes Model.
The following hardware configuration was used for the comparative studies and experimentation for the intrusion detection model presented in this paper: Operating System: Windows 64-bit, Processor: 2.80 GHz Intel Core i7 CPU, 16 GB of RAM for graphics processing unit: Python-compatible Nvidia GPU with 4 GB of RAM. All of the experimental and comparative tests covered in this paper were conducted using this setup.

4.1. Evaluation Metrics

The optimal metric for IoT anomaly detection is decided by the specific situation and the system’s goals. Different measures have different strengths and weaknesses, and the best metrics to employ are determined by the most important component of system performance. If the goal is to detect as many anomalous cases as possible, recall, which measures the proportion of true positive predictions out of all real anomalous instances in the dataset, may be the most relevant parameter. A high recall shows that the model can discover a significant fraction of the anomalous cases in the sample.
If the goal is to reduce false positives (occurrences that are anticipated to be abnormal but are normal), accuracy may be the most relevant metric. A high precision suggests that the model can anticipate anomalous cases effectively while minimizing false positives.
F1-score may be an appropriate metric to utilize if a balance between recall and precision is sought. The F1-score in Equation (15) is a harmonic mean of precision and recall in Equation (12), with equal weight given to both metrics.
AUC-ROC can provide a measure of the model’s overall performance in discriminating between anomalous and normal occurrences, and the confusion matrix can provide a thorough breakdown of the classification findings.
The best metric for detecting IoT anomalies is determined by the situation and the system’s aims. Distinct metrics have distinct strengths and limitations, and the optimal one to use is determined by the most critical aspect of system performance. The confusion matrix is a table that summarizes the results of an instance classification task, such as IoT anomaly detection. The number of true positives (TP), true negatives (TN), false positives (FP), and false negatives (FN) for each class (anomalous and normal) is shown in the confusion matrix.
The confusion matrix can be used to compute metrics like accuracy, precision, recall, and F1-score.
The accuracy is defined as (TP + TN)/(TP + TN + FP + FN)
Precision = (TP + FP)
TP = (TP + FN)
F1-score = 2 ∗ (precision ∗ recall)/(precision + recall)
The confusion matrix can be used in IoT anomaly detection to evaluate the performance of a classification model in recognizing anomalous instances from sensor data. True positives are instances that are correctly classified as anomalous, true negatives are instances that are correctly classified as normal, false positives are instances that are incorrectly classified as anomalous (normal instances classified as anomalous), and false negatives are instances that are incorrectly classified as normal (anomalies classified as normal). By analyzing the confusion matrix, it is possible to determine the classification model’s strengths and shortcomings and alter its parameters to improve its performance. For example, if the model has a high false positive rate, the classification threshold may need to be adjusted or more features added to the model to increase its capacity to distinguish between anomalous and typical cases. Overall, the confusion matrix is a good tool for evaluating classification model performance in IoT anomaly detection.

4.2. Performance Evaluation UNSW-NB15 Dataset

Dataset UNSW-NB15 includes records of both benign traffic and five different types of attacks, such as Fuzzers, Normal, Generic, DoS, and Exploits. It was developed by the Australian Centre for Cyber Security (ACCS). The selected features determined the optimal performance of the model (DoS, Exploits, Fuzzers, Normal and Generic). Furthermore, when the entire feature set was used to train the intrusion detection system (IDS) model on this dataset, it achieved an accuracy of 80.00%, as shown in Figure 5. The records were gathered from three real-world websites: BID (Symantec Corporation), CVE (Common Vulnerabilities and Exposures), and MSD (Microsoft Corporation) (Microsoft Security Bulletin). The results for the UNSW-NB15 dataset are presented in Figure 6. It was found that three layers of CNN achieved the highest accuracy. For the final feature selection testing, DCGR_IoT was utilized.
Figure 5. DCGR_IoT confusion matrix for UNSW-NB15 dataset.
Figure 6. DCGR_IoT training and testing performance on UNSW-NB15 dataset.

4.3. Performance Evaluation DCGR_IoT a Using the KDDCup99

KDDCup99 is a well-known dataset for intrusion detection research that was generated for the 1999 KDD Cup data mining competition. Researchers at the University of California, Irvine generated the dataset, which contains network traffic data taken from a simulated network environment.
The KDDCup99 dataset comprises both regular and malicious traffic, as well as 41 network traffic features such as protocol type, service, source and destination IP addresses, and connection length. There are four types of attacks in the dataset: DoS (Denial of Service), Probe, R2L (Remote-to-Local), and U2R (User-to-Root). The attacks were created with a variety of tools and approaches, and the dataset includes raw network traffic data as well as preprocessed CSV files for each attack scenario. The KDDCup99 dataset has been frequently utilized in research projects to assess the efficacy of machine learning algorithms in identifying network intrusions. However, the dataset has some drawbacks, including the fact that it is based on a simulated environment and contains outdated attack methodologies and traffic patterns. Because of these constraints, newer and more diverse datasets for intrusion detection studies have been generated throughout time. Nonetheless, the KDDCup99 dataset is still a popular resource for network security researchers and practitioners. The KDDCup99 dataset is open to the public and can be obtained from the dataset’s official website: http://kdd.ics.uci.edu/databases/kddcup99/kddcup99.html (accessed on 2 August 2024). The selected features determined the optimal performance of the model (DoS, U2R, R2L, Normal, and Probe). The technique was validated with 496,020 training records, 311,028 testing records, and a 99.2% total accuracy as shown in Figure 7. To assess it further, a confusion matrix was applied. Figure 8 shows the training and validation accuracies.
Figure 7. DCGR_IoT is a confusion matrix for the KDDCup99 dataset.
Figure 8. DCGR_IoT training and testing performance on KDDCup99 dataset.

4.4. Performance Evaluation DCGR_IoT Using the IoT-23 DATASET

The IoT-23 dataset includes both benign and malicious traffic, simulating an IoT network in a smart house. To simulate benign circumstances, the network traffic gathered from three IoT devices was included. The dataset also includes various attacks, including DDoS attacks (https://www.stratosphereips.org/datasets-iot23 (accessed on 2 August 2024)). The confusion matrix for the performance of the model on the IoT-23 dataset is shown in Figure 9. A table called a confusion matrix is frequently used to explain how well a classification model performs when applied to a set of test data whose real values are known. The anticipated labels are shown in the columns, and genuine labels are shown in the rows. Each cell in the matrix represents the number of times a true label is applied to a predicted label.
Figure 9. DCGR_IoT confusion matrix for IoT-23 dataset.
‘Benign’, ‘DDoS’, ‘Mirai’, ‘Okiru’, ‘Tori’, ‘PartofHorizontal-PortScan’, ‘Heartbeat’, ‘FileDownload’, and ‘Command and control’ are the labels assigned to the rows and columns in this confusion matrix. The genuine label for each row and the anticipated label for each column were matched.
The number of times the model accurately predicted ‘Benign’ traffic is indicated in the cell where the ‘Benign’ row and ‘Benign’ column overlap. Comparably, the number of times that ‘DDoS’ traffic was incorrectly projected as ‘Mirai’ is indicated in the cell where the ‘Mirai’ column crosses the ‘DDoS’ row.
The diagonal of the matrix shows the model’s correct predictions from top left to bottom right. Misclassifications were represented as off-diagonal cells. The number of instances that a cell represents increases with deeper hue. The accuracy of the predictions of the classification model is known. This is determined by dividing the total number of forecasts by the number of accurate estimates. The accuracy in this instance is 0.9985384650914202, which is extremely near to 1. This suggests that the model predicts the right labels for the cases in the IoT-23 dataset with a high degree of accuracy. This paper examines the integration of current security mechanisms to improve the overall security of the DCGR_IoT model’s overall security:
All network traffic between IoT devices and the DCGR_IoT system can be encrypted using TLS integration. This will prevent network traffic from being intercepted or tampered with while it is being transmitted. 1. To impose end-to-end encryption, the system can be set up to allow TLS connections. 2. Digital signatures and standards, such as X.509 certificates, can be used to authenticate IoT devices. This ensures that only approved devices can send network traffic and connect to the system. 3. To verify device IDs prior to allowing traffic, the system can incorporate device authentication. To identify and address security events, security monitoring systems such as Security Information and Event Management (SIEM) can be incorporated. This will facilitate quick incident response and give real-time visibility into system security. The DCGR_IoT system’s overall security will be greatly improved by integrating these current security mechanisms. There will be defenses in place to protect the system from threats and attacks. To guarantee smooth integration and compatibility, meticulous preparation and configuration are necessary. All things considered, the system and its data are well protected by this multilayered security technique. We only consider packets that belong to IPv4 and one of the three protocols: TCP, UDP, or ICMP. The DCGR_IoT model has achieved the highest accuracy of 0.9985384650914202, 99.85% on the IoT-23 dataset. With an accuracy of 0.9985384650914202, 99.85% of the dataset’s instances were accurately predicted by the model. Given the extremely high accuracy, it appears that the model is doing a good job with this task.
Figure 10 shows the simulation results. DCGR_IoT A training and testing performance on the IoT-23 dataset presented two line plots, one depicting the model accuracy and the other showing the model loss over 100 epochs. The x-axis in both plots represents the number of epochs, and the y-axis indicates the respective metrics. The model accuracy shows that the training accuracy line generally exhibits an upward trend, suggesting that the model is learning and improving its performance on the training data over time. There are some fluctuations in the training accuracy, which are common during the training process. Additionally, there was a rising trend in the testing accuracy line, suggesting that the model’s generalization performance improved. In general, the training loss line trends downward, showing that the model learns to reduce the error between its predictions and the actual values. The training loss shows variations similar to those in the accuracy plot. The testing loss line likewise exhibits a downward trend, indicating that the model is becoming more effective for data that have not yet been observed. The model learns effectively, as both training and testing accuracy increase, while training loss decreases.
Figure 10. DCGR_IoT A training and testing performance on IoT-23 dataset.

5. Conclusions

As the period of employing IoT network intrusion detection systems and other areas in real-life approaches, this technique can be linked with IoT situations as well. As we have highlighted in previous works [23,24,25,26,27,28], one of the directions of the modern technology and industrial revolution is IoT. We also examined the differences and parallels between IoT network intrusion detection approaches. One of the intrinsic new abilities of this intrusion detection technology is its security aspects. One of the primary drawbacks of current IoT solutions is the centralized system. The rapid expansion of the Internet of Things (IoT) has underscored the critical need for robust network security measures. Traditional intrusion detection systems (IDS) struggle to adapt to the unique challenges posed by IoT environments, including the diverse range of devices, complex network traffic patterns, and the imperative for real-time detection capabilities. To tackle these issues, we introduce DCGR_IoT, a cutting-edge intrusion detection system (IDS) based on deep neural learning designed to safeguard bidirectional communication networks in IoT settings. DCGR_IoT leverages advanced techniques to enhance anomaly detection capabilities. It employs convolutional neural networks (CNN) for spatial feature extraction and filters out redundant data to improve computational efficiency. Additionally, complex gated recurrent networks (CGRNs) are utilized in the temporal feature extraction module, which is a key component of DCGR_IoT. Moreover, DCGR_IoT utilizes complex gated recurrent networks (CGRNs) to create multidimensional feature subsets, enabling a more comprehensive spatial representation of network traffic and facilitating the extraction of crucial features vital for intrusion detection. The efficacy of DCGR_IoT was validated through extensive testing on the UNSW-NB15, KDDCup99, and IoT-23 datasets, achieving a high detection accuracy of 99.2%. These findings highlight the potential of DCGR-IoT as a potent solution for protecting IoT networks against advanced cyber threats.

Author Contributions

Conceptualization, E.E.-S.; Data curation, H.E. and G.F.E.; Formal analysis, W.M.E. and M.I.I.; Funding acquisition, M.A.; Methodology, H.E.; Resources, W.M.E., M.I.I. and G.F.E.; Software, E.E.-S.; Validation, E.E.-S., M.A. and M.I.I.; Visualization, M.A. and G.F.E.; Writing—original draft, E.E.-S., H.E. and G.F.E.; Writing—review and editing, E.E.-S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Researchers Supporting Project number (RSPD2024R636), King Saud University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The KDDCup99 dataset is open to the public and can be obtained from the dataset’s official website: http://kdd.ics.uci.edu/databases/kddcup99/kddcup99.html (accessed on 2 August 2024).

Conflicts of Interest

The authors declare that they have no conflicts of interest to report regarding the present study.

References

  1. Dahlqvist, F.; Patel, M.; Rajko, A.; Shulman, J. Growing Opportunities in the Internet of Things; McKinsey & Company: Atlanta, GA, USA, 2019. [Google Scholar]
  2. Apicella, A.; Donnarumma, F.; Isgrò, F.; Prevete, R. A survey on modern trainable activation functions. Neural Netw. 2021, 138, 14–32. [Google Scholar] [CrossRef] [PubMed]
  3. Le, K.H.; Nguyen, M.H.; Tran, T.D.; Tran, N.D. IMIDS: An intelligent intrusion detection system against cyber threats in IoT. Electronics 2022, 11, 524. [Google Scholar] [CrossRef]
  4. Saba, T.; Rehman, A.; Sadad, T.; Kolivand, H.; Bahaj, S.A. Anomaly-based intrusion detection system for IoT networks through deep learning model. Comput. Electr. Eng. 2022, 99, 107810. [Google Scholar] [CrossRef]
  5. Moustafa, N.; Hu, J.; Slay, J. A holistic review of network anomaly detection systems: A comprehensive survey. J. Netw. Comput. Appl. 2019, 128, 33–55. [Google Scholar] [CrossRef]
  6. Aldweesh, A.; Derhab, A.; Emam, A.Z. Deep learning approaches for anomaly-based intrusion detection systems: A survey, taxonomy, and open issues. Knowl.-Based Syst. 2020, 189, 105124. [Google Scholar] [CrossRef]
  7. Heidari, A.; Jabraeil Jamali, M.A. Internet of Things intrusion detection systems: A comprehensive review and future directions. Clust. Comput. 2023, 26, 3753–3780. [Google Scholar] [CrossRef]
  8. Alzubi, O.A.; Alzubi, J.A.; Alazab, M.; Alrabea, A.; Awajan, A.; Qiqieh, I. Optimized machine learning-based intrusion detection system for fog and edge computing environment. Electronics 2022, 11, 3007. [Google Scholar] [CrossRef]
  9. Le Jeune, L.; Goedeme, T.; Mentens, N. Machine learning for misuse-based network intrusion detection: Overview, unified evaluation and feature choice comparison framework. IEEE Access 2021, 9, 63995–64015. [Google Scholar] [CrossRef]
  10. Muhammad, K.; Obaidat, M.S.; Hussain, T.; Ser, J.D.; Kumar, N.; Tanveer, M.; Doctor, F. Fuzzy logic in surveillance big video data analysis: Comprehensive review, challenges, and research directions. ACM Comput. Surv. (CSUR) 2021, 54, 1–33. [Google Scholar] [CrossRef]
  11. Sen, J.; Mehtab, S. Machine learning applications in misuse and anomaly detection. In Security and Privacy from a Legal, Ethical, and Technical Perspective; IntechOpen: London, UK, 2020; p. 155. [Google Scholar]
  12. Sun, N.; Ding, M.; Jiang, J.; Xu, W.; Mo, X.; Tai, Y.; Zhang, J. Cyber threat intelligence mining for proactive cybersecurity defense: A survey and new perspectives. IEEE Commun. Surv. Tutor. 2023, 25, 1748–1774. [Google Scholar] [CrossRef]
  13. Silva JV, V.; Lopez, M.A.; Mattos, D.M. Attackers are not stealthy: Statistical analysis of the well-known and infamous KDD network security dataset. In Proceedings of the 2020 4th Conference on Cloud and Internet of Things (CIoT), Niteroi, Brazil, 7–9 October 2020; pp. 1–8. [Google Scholar]
  14. Li, T.; Wang, Z.; Liu, S.; Lin, W.Y. Deep unsupervised anomaly detection. In Proceedings of the 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2021; pp. 3636–3645. [Google Scholar]
  15. Ahmad, R.; Alsmadi, I. Machine learning approaches to IoT security: A systematic literature review. Internet Things 2021, 14, 100365. [Google Scholar] [CrossRef]
  16. Li, Y.; Zuo, Y.; Song, H.; Lv, Z. Deep learning in security of internet of things. IEEE Internet Things J. 2021, 9, 22133–22146. [Google Scholar] [CrossRef]
  17. Diro, A.; Mahmood, A.; Chilamkurti, N. Collaborative intrusion detection schemes in fog-to-things computing. In Fog/edge Computing for Security, Privacy, and Applications; Springer: Cham, Switzerland, 2021; pp. 93–119. [Google Scholar]
  18. Sharma, B.; Sharma, L.; Lal, C.; Roy, S. Anomaly based network intrusion detection for IoT attacks using deep learning technique. Comput. Electr. Eng. 2023, 107, 108626. [Google Scholar] [CrossRef]
  19. Yang, Z.; Liu, X.; Li, T.; Wu, D.; Wang, J.; Zhao, Y.; Han, H. A systematic literature review of methods and datasets for anomaly-based network intrusion detection. Comput. Secur. 2022, 116, 102675. [Google Scholar] [CrossRef]
  20. Muthamil Sudar, K.; Deepalakshmi, P. An intelligent flow-based and signature-based IDS for SDNs using ensemble feature selection and a multi-layer machine learning-based classifier. J. Intell. Fuzzy Syst. 2021, 40, 4237–4256. [Google Scholar] [CrossRef]
  21. Jiang, K.; Wang, W.; Wang, A.; Wu, H. Network intrusion detection combined hybrid sampling with deep hierarchical network. IEEE Access 2020, 8, 32464–32476. [Google Scholar] [CrossRef]
  22. Rahul-Vigneswaran, K.; Poornachandran, P.; Soman, K.P. A compendium on network and host based intrusion detection systems. In Lecture Notes in Electrical Engineering, ICDSMLA 2019: Proceedings of the 1st International Conference on Data Science, Machine Learning and Applications, 29–30 March 2019, Bengaluru, India; Springer: Singapore, 2020; pp. 23–30. [Google Scholar]
  23. Khan, M.A. HCRNNIDS: Hybrid convolutional recurrent neural network-based network intrusion detection system. Processes 2021, 9, 834. [Google Scholar] [CrossRef]
  24. Idrissi, I.; Boukabous, M.; Azizi, M.; Moussaoui, O.; El Fadili, H. Toward a deep learning-based intrusion detection system for IoT against botnet attacks. IAES Int. J. Artif. Intell. 2021, 10, 110. [Google Scholar] [CrossRef]
  25. Jayarajan, K.; Poongothai, T.; Patra, P. Building a Novel Intrusion Detection System using Long Short-Term Memory for Detecting Network Attacks. Turk. Online J. Qual. Inq. 2021, 12, 7184. [Google Scholar]
  26. Ramaiah, M.; Chandrasekaran, V.; Ravi, V.; Kumar, N. An intrusion detection system using optimized deep neural network architecture. Trans. Emerg. Telecommun. Technol. 2021, 32, e4221. [Google Scholar] [CrossRef]
  27. Le TT, H.; Kim, Y.; Kim, H. Network intrusion detection based on novel feature selection model and various recurrent neural networks. Appl. Sci. 2019, 9, 1392. [Google Scholar] [CrossRef]
  28. Abdalgawad, N.; Sajun, A.; Kaddoura, Y.; Zualkernan, I.A.; Aloul, F. Generative deep learning to detect cyberattacks for the IoT-23 dataset. IEEE Access 2021, 10, 6430–6441. [Google Scholar] [CrossRef]
  29. Ngo, D.M.; Lightbody, D.; Temko, A.; Pham-Quoc, C.; Tran, N.T.; Murphy, C.C.; Popovici, E. Network Attack Detection on IoT Devices Using 2D-CNN Models. In International Conference on Intelligence of Things; Springer Nature: Cham, Switzerland, 2023; pp. 237–247. [Google Scholar]
  30. Cao, B.; Li, C.; Song, Y.; Qin, Y.; Chen, C. Network intrusion detection model based on CNN and GRU. Appl. Sci. 2022, 12, 4184. [Google Scholar] [CrossRef]
  31. DeMedeiros, K.; Hendawi, A.; Alvarez, M. A survey of AI-based anomaly detection in IoT and sensor networks. Sensors 2023, 23, 1352. [Google Scholar] [CrossRef]
  32. Feldkamp, L.A.; Puskorius, G.V. A signal processing framework based on dynamic neural networks with application to problems in adaptation, filtering, and classification. Proc. IEEE 1998, 86, 2259–2277. [Google Scholar] [CrossRef]
  33. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  34. Krichen, M. Convolutional neural networks: A survey. Computers 2023, 12, 151. [Google Scholar] [CrossRef]
  35. Alahmari, F.; Naim, A.; Alqahtani, H. E-Learning Modeling Technique and Convolution Neural Networks in Online Education. In IoT-enabled Convolutional Neural Networks: Techniques and Applications; River Publishers: Aalborg, Denmark, 2023; pp. 261–295. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.