Next Article in Journal
Optimization of BBR Congestion Control Algorithm Based on Pacing Gain Model
Next Article in Special Issue
Binary Hunter–Prey Optimization with Machine Learning—Based Cybersecurity Solution on Internet of Things Environment
Previous Article in Journal
Low-Cost Water Quality Sensors for IoT: A Systematic Review
Previous Article in Special Issue
Information Security Threats and Working from Home Culture: Taxonomy, Risk Assessment and Solutions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Intrusion Detection Systems for IoT and Cloud Environments Using a Growth Optimizer Algorithm and Conventional Neural Networks

by
Abdulaziz Fatani
1,2,
Abdelghani Dahou
3,
Mohamed Abd Elaziz
4,5,6,7,
Mohammed A. A. Al-qaness
8,
Songfeng Lu
9,10,*,
Saad Ali Alfadhli
11 and
Shayem Saleh Alresheedi
12
1
School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan 430074, China
2
Computer Science Department, Umm Al-Qura University, Makkah 24381, Saudi Arabia
3
Faculty of Computer Sciences and Mathematics, Ahmed Draia University, Adrar 01000, Algeria
4
Department of Mathematics, Faculty of Science, Zagazig University, Zagazig 44519, Egypt
5
Artificial Intelligence Research Center (AIRC), Ajman University, Ajman 346, United Arab Emirates
6
Department of Artificial Intelligence Science and Engineering, Galala University, Suze 435611, Egypt
7
Department of Electrical and Computer Engineering, Lebanese American University, Byblos 13-5053, Lebanon
8
College of Physics and Electronic Information Engineering, Zhejiang Normal University, Jinhua 321004, China
9
Hubei Engineering Research Center on Big Data Security, School of Cyber Science and Engineering, Huazhong University of Science and Technology, Wuhan 430074, China
10
Shenzhen Huazhong University of Science and Technology Research Institute, Shenzhen 518057, China
11
Department of Computer Techniques Engineering, Imam Al-Kadhum College, Baghdad 10081, Iraq
12
War College, National Defense University, Riyadh 12211, Saudi Arabia
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(9), 4430; https://doi.org/10.3390/s23094430
Submission received: 12 February 2023 / Revised: 7 April 2023 / Accepted: 18 April 2023 / Published: 30 April 2023

Abstract

:
Intrusion detection systems (IDS) play a crucial role in securing networks and identifying malicious activity. This is a critical problem in cyber security. In recent years, metaheuristic optimization algorithms and deep learning techniques have been applied to IDS to improve their accuracy and efficiency. Generally, optimization algorithms can be used to boost the performance of IDS models. Deep learning methods, such as convolutional neural networks, have also been used to improve the ability of IDS to detect and classify intrusions. In this paper, we propose a new IDS model based on the combination of deep learning and optimization methods. First, a feature extraction method based on CNNs is developed. Then, a new feature selection method is used based on a modified version of Growth Optimizer (GO), called MGO. We use the Whale Optimization Algorithm (WOA) to boost the search process of the GO. Extensive evaluation and comparisons have been conducted to assess the quality of the suggested method using public datasets of cloud and Internet of Things (IoT) environments. The applied techniques have shown promising results in identifying previously unknown attacks with high accuracy rates. The MGO performed better than several previous methods in all experimental comparisons.

1. Introduction

The need to secure online data, information, and related systems has grown in importance with the development of information communication, particularly the Internet. Sensitive data is created, transported, stored, or updated worldwide daily in enormous quantities. Private emails, financial transactions, simple holiday photos, and military communications are all examples of sensitive information. Malicious parties have sought to steal, alter, or erase this information for a long time. Hackers and other hostile actors have developed, exploited, and enhanced various cyberattacks to accomplish these objectives [1]. A paradigm shift from straightforward defense mechanisms to complex defense systems was necessary for this new era of cyber security. While simple network security measures such as firewalls may have been enough in the past, the sophistication of cyberattacks has made them ineffective when used alone. Intrusion Detection Systems (IDS) are currently the cornerstone of cyber security to defend against these sophisticated attacks [1]. In the cloud and IoT, there are three primary divisions of cloud services: infrastructure-as-a-service (IaaS), platform-as-a-service (PaaS), and software-as-a-service (SaaS). To give users security, it is necessary to address the weaknesses and problems each of these services and methods possesses [2]. In recent years, different methods have been proposed for the IDS, such as the traditional machine learning techniques, for example, the support vector machine (SVM) [3,4], decision trees [5,6], k-means clustering [7,8], and others. The recent advances in deep neural networks, including conventional neural networks (CNNs) and recurrent neural networks (RNNs), were also adopted in this field [9]. Several IDS were developed based on ANNs, such as RNNs [10] and CNNs [11].
In recent years, a new direction was utilized for the IDS by employing the power of the metaheuristic optimization algorithms adopted in different and complex engineering and optimization problems, including IDS. For example, Alazab et al. [12] employed the moth–flame optimizer algorithm to build an IDS method. The MFO was a feature selection method that enhanced the classifier’s performance (Decision Tree). The evaluation showed that the classification accuracy of the DT was improved by applying the MFO. In [13], the authors applied a combined MH method using the firefly algorithm (FA) and ant lion optimization algorithm to build an efficient IDS system. Zhou et al. [14] employed the bat algorithm as a feature selection to build an IDS. It was evaluated with random forest classifier, C4.5, and ForestPA. It is clear that MH optimization algorithms have shown significant performance in IDS applications; thus, they have been widely adopted, such as whale optimization algorithm [15], particle swarm optimization algorithm [16], Aquila optimization algorithm [17], reptile search algorithm [18], salp swarm algorithm [19], and many others.

Paper Contribution

Following the successful applications of MH optimization algorithms in IDS, we propose an efficient feature selection technique called MGO. This method is developed based on two aspects; the first is to utilize the power of the Growth Optimizer (GO) in the exploration phase of the search process. The second aspect is to employ the integration between GO and WOA in the exploitation phase. The main objective of this study can be simplified as the following points:
  • Suggest a different method for securing IoT by combining DL and feature selection techniques.
  • Use a CNN model to analyze network traffic records and identify complex feature representations.
  • Create a modified version of Growth Optimizer (GO) for improved intrusion detection in IoT environments. The modification uses the operators of the Whale Optimization Algorithm (WOA). The proposed method, called MGO, is employed to address the issue of discrete feature selection.
  • Evaluate the performance of the MGO against established methods using four actual intrusion datasets.
The paper is structured as follows: Section 2 explains the employed methods, Section 3 outlines the proposed IoT security system, Section 4 assesses the system, and Section 5 concludes the results.

2. Background

2.1. Growth Optimizer

In this section, the Growth Optimizer (GO) simulates how people train and reflect as they progress in society. In the learning phase, the information is collected from the environment, whereas the reflection aims to examine the shortcomings and improve the learning method.
In general, the GO starts by using Equation (1) to generate the population X which stands for the solutions for the tested problem.
X i = r × ( U L ) + L , i = 1 , . . . , N
where r is the random value and the limits of the search domain of the problem are represented using U and L. N refers to the total number of solutions in X.
Following [20], X is divided into three parts according to the parameters named P 1 = 5 . The first part comprises the leader and the elites (varying from 2 to P 1 ). The second part contains the middle level (i.e., from P 1 + 1 to N P 1 ) and the third part contains the bottom level (i.e., N P 1 + 1 and N), whereas the best solution is the leader of the upper level.

2.1.1. Learning Stage

By confronting disparities between people, examining the causes of those differences, and learning from them, individuals can be greatly helped in their progress. The GO’s learning stage simulates four key gaps that are formulated:
G 1 = X b X b t G 2 = X b X w G 1 = X b t X w G 1 = X r 1 X r 2
where X b , X b t , X w indicate best, better, and worst solution, respectively; in addition, X r 1 , and X r 2 are two random solutions. G k ( k = 1 , 2 , 3 , 4 ) stands for the gap used to improve the skills learned and decrease the difference between them. Moreover, to reflect the variation between the groups, the parameter named learning factor (LF) is applied and its formulation is given as:
L F k = | | G k | | k = 1 4 | | G k | | , k = 1 , 2 , 3 , 4
Following [20], the individual can assess his learned knowledge using the parameter ( S F i ):
S F i = G R i G R m a x
where G R m a x and G R i represent the maximum growth resistance of X and the growth of X i , respectively.
According to the information collected from L F k and S F i each X i can receive new knowledge from the solution belonging to each gap G k using the knowledge acquisition ( K A k ) that is defined as:
K A k = S F i × L F k × G k , k = 1 , 2 , 3 , 4
After that, the solution X i can improve its information using the following formula:
X i ( t + 1 ) = X i ( t ) + k = 1 4 K A k
The quality of the updated version of X i is computed and compared with the previous one to determine whether there is a significant difference between them.
X i ( t + 1 ) = X i ( t + 1 ) i f f ( X i ( t + 1 ) ) f ( X i ( t ) ) X i ( t + 1 ) i f r 1 < P 2 , i n d ( i ) = i n d ( 1 ) X i ( t ) o t h e r w i s e o t h e r w i s e ,
where r 2 stands for a random number and P 2 = 0.001 is the probability retention. i n d ( i ) refers to the ranking of X i based on the ascending order X using the fitness value.

2.1.2. Reflection Stage

The solution must develop their ability to reflect on the knowledge they have learned, meaning that X must identify all of their areas of weakness, make up for them, and retain their information. They ought to adopt the undesirable attributes of successful X while retaining their outstanding qualities. When the lesson of a specific aspect cannot be mended, the prior information should be abandoned and systematic learning should resume. Equations (8) and (9) can be used to mathematically model this process.
X i ( t + 1 ) = r 4 × ( U L ) i f r 3 < A F X i ( t ) + r 5 × ( X R X i ( t ) ) o t h e r w i s e o t h e r w i s e X i ( t ) o t h e r w i s e ,
A F = 0.01 + 0.99 × ( 1 F E s m a x F E )
where r 3 , r 4 , and r 5 are random values. X R refers to a solution defined as the top P 1 + 1 solutions in X. A F refers to the attenuation factor which depends on function evaluation F E and the total number of functions evaluations m a x F E .
After the complete reflection stage, X i should evaluate its growth, similar to the learning phase. Therefore, Equation (7) is also applied to achieve this task.

2.2. Whale Optimization Algorithm

The WOA [21] draws inspiration from the unique hunting strategy used by a particular species of killer whale known as humpback, whose hunting style is bubble-net feeding. WOA’s mathematical formulation depends on how it behaves when hunting. Each whale’s location can be represented by the solution X b , which can be updated depending on how the whale behaves when attacking its prey. The whales can attack their prey using two different methods. The humpback whale locates its prey and encircles it using the first strategy, known as encircling prey. WOA presupposes that the target prey is the best option ( X b ( t ) ). The other whales attempt to update their locations in the direction of X b ( t ) after it has been identified (found), as in Equation (10):
D i s i = | B X b ( t ) X i ( t ) | , B = 2 r
X i ( t + 1 ) = X b ( t ) A D i s i , A = 2 a r a
where D i s i stands for the distance between X i ( t ) and X b ( t ) . r [ 0 , 1 ] refers to a random value. In addition, a denotes a parameter that decreases from 2 to 0 during the process of updating the solution, formulated as a = a t a t m a x ( t m a x is the total of iterations).
The second strategy is called the bubble-net attack. This phase has two main steps: spiral updating location and shrinking encircling mechanism, and reducing the value of a in Equation (11) for satisfying the process of shrinking encircling. The whale’s locations, X i and X b , are separated by the following distance, which is calculated by the spiral updating position method [21]:
X ( t + 1 ) = D i s e b l c o s ( 2 π l ) + X b ( t )
In Equation (12), l stands for a constant value which represents the shape of the logarithmic spiral.
The whales can also swim simultaneously around the X b utilizing a spiraling path and a contracting circle. The following equation depends on integrating Equations (10)–(11) and Equation (12) [21]; therefore, X can be enhanced as:
X ( t + 1 ) = X b ( t ) A D i s i f p 0.5 D i s e b l c o s ( 2 π l ) + X b ( t ) i f p < 0.5 .
In Equation (13), p [ 0 , 1 ] refers to a probability value used to identify the strategy of updating. In addition, X i can be enhanced using a random selecting solution X r instead of X b as represented using Equation (14) [21]:
X ( t + 1 ) = X r A D i s
D i s = | B X r a n d X ( t ) |

3. Proposed Method

The steps of the developed IoT security are introduced in this section. The developed technique depends on improving the performance of the Growth Optimizer using the Whale Optimization Algorithm (WOA).

3.1. Prepare IoT Dataset

The developed MGO starts by preparing the IoT dataset by normalizing it. This is performed using the min−max technique that applied to the IoT data D S [22], which is represented as
D S = d s 11 d s 12 . . . d s 1 d d s 21 d s 22 . . . d s 2 d . . . . . . . . . . . . d s n 1 d s n 2 . . . d s n d
where D S i = [ d s i 1 , d s i 2 , . . . , d s i d ] denotes the features of traffic i, whereas n and d are sample and feature numbers, respectively. The normalized version of D S based on min−max technique is represented as [22]:
N D S i j = D s i j m i n ( D S j ) m a x ( D S j ) m i n ( D S j )

3.2. CNN for Feature Extraction

Convolutional neural networks (CNNs) are widely used in computer vision as they are robust feature extraction modules, especially when employing pre-trained models alongside transfer learning methods. Meanwhile, CNN is also used in applications where the data are one-dimensional such as in natural language processing. Our study aims to train a DL model that benefits from the big data generated from IoT devices to perform network intrusion detection and reduce processing complexity and inference time. Thus, this section proposes a light CNN model to automatically learn helpful patterns/representations rather than relying on the raw data collected from experimental and real network intrusion detection experiments. In addition, we extract the learned features for further processing (feature selection) to improve the overall framework performance (detection accuracy) and reduce the dimensionality space of the represented feature to accelerate the inference process.
The proposed CNN architecture receives a set of samples, X, where each row is a one-dimensional raw sample representing a network traffic record which includes several network attributes (columns) related to the possible attack class, such as flags related to the IP address, TCP flags, destination, source information, type of service, communication protocols, and protection protocols. The CNN architecture, as shown in Figure 1, is composed of two convolution blocks (ConvBlock) to learn spatial relations between raw attributes and generate new representations as output (feature maps). Each ConvBlock comprises a convolutional layer with a one-dimensional kernel k, activation function, and pooling operation. Each ConvBlock uses a kernel of size 1 × 3 and 64 output channels to produce the output feature maps out ( t ) which is a new transformation of the input raw data at a certain timestamp t where i is the input channel. A non-linear activation function name rectified linear unit (ReLU) after each convolution operation is followed by a max-pooling with size two to output the final feature maps.

3.3. Feature Selection-Based MGO Approach

We developed an alternative FS approach based on a modified version of GO algorithm using WOA as given in Figure 2. This algorithm allocates the relevant features from those extracted using the CNN model.
The first step in MGO as FS approach is to split the data into training and testing sets, which represent 80% and 20%, respectively. Then, the initial solution X is built as given in Equation (18).
X i = L B + r a n d ( 1 , D ) × ( U B L B ) , i = 1 , 2 , . . . , N
where N stands for the total number of solutions and D the number of features. L B and U B are the limits of the search domain. r a n d ( 1 , D ) stands for the random value with D values.
The next step is to generate the Boolean version of X i using the following formula:
B X i j = 1 i f X i j > 0.5 0 o t h e r w i s e
We select only the features corresponding to ones in B X i and remove the other features. Then, we compute the fitness value of X i as:
F i t i = ( 1 λ ) × ( | B X i | D ) + λ × γ i
where γ i stands for the classification error based on KNN using the training sets, whereas, λ [ 0 , 1 ] is the weight used to control the balancing between γ i and the ( | B X i | D ) , which represents the ratio of relevant features.
Thereafter, we determined the best solution X b and used it to enhance the current solutions by combining GO and WOA. This was conducted using GO operators in the exploration phase, while the following integration schema was used during the exploitation phase.
X i = X W O A i f P r > r s X G O o t h e r w i s e
where X G O refers to using the operators of GO that were used to update X i and X W O A is the operator of WOA defined in Equations (10)–(15). P r is the probability of each X i and it is defined as:
P r = F i t i i = 1 N X i
where F i t i stands for the fitness value of X i . In addition, the value of r s is updated using the following formula.
r s = m i n ( P r ) + r a n d × ( m a x ( P r ) m i n ( p r ) )
where m i n and m a x are the minimum and maximum functions, respectively.
Then, the stop condition is checked and in case they are met, the update process is stopped. Otherwise, we repeat it again. After that, X b is used to remove irrelevant features from the testing set and evaluate this process using different performance criteria.
The time complexity of the developed MGO as FS method depends on some factors such as (1) the size of population N, (2) the dimension of features D, and (3) the number of iterations t m a x . So, the complexity of MGO is formulated as:
O ( M G O ) = N + t × ( N × D + ( O ( W O A ) + O ( G O R e f l e c t i o n ) ) × D )
O ( M G O ) = N × D + t × ( N × D + ( K w × D + ( N K w ) × D ) )
So,
O ( M G O ) = N × D + t × ( N × D ) = O ( N × D × t )
where K w refers to the number of solutions that will be updated using WOA.

4. Experimental Series and Results

The section uses a set of experimental series to assess the developed IoT security based on a modified version of the GO algorithm using WOA. These experimental series are implemented using a set of real-world IoT datasets.

4.1. Evaluation Measures

The effectiveness of the suggested technique and all comparing methodologies is evaluated using several indicators.
  • Average accuracy ( A V A c c : This measure stands for the rate of a correct intrusion detected using the algorithm and it is represented as:
    A V A c c = 1 N r k = 1 N r A c c B e s t k ,
    A c c B e s t = T P + T N T P + F N + F P + T N
    in which N r = 30 indicates the iteration numbers.
  • Average Recall ( A V r e c a l l ) : is the percentage of intrusion predicted positively (it can be called true positive rate (TPR)). It can be computed as:
    A V R e c a l l = 1 N r k = 1 N r R e c a l l B e s t k , R e c a l l B e s t = T P T P + F N
  • Average Precision ( A V P r e c ) : stands for the rate of TP samples of all positive cases with the formulation:
    A V P r e c = 1 N r k = 1 N r P r e c B e s t k , P r e c B e s t = T P F P + T P
  • Average F1-measure ( A V F 1 ) : can be computed as:
    A V F 1 = 2 R e c a l l × P r e c i s i o n P r e c i s i o n + R e c a l l
  • Average G-mean ( A V G M ) : can be computed as:
    A V G M = R e c a l l × P r e c i s i o n

4.2. Experiments Setup

In our experiments, we trained the CNN model on each dataset record for 100 epochs with early stopping and Adam with a 0.005 learning rate to update the network parameters. A batch of size 2024 is used to iterate over the data samples. In addition, batch normalization and dropout with a 0.38 ratio were used as regularization techniques to prevent overfitting, increase generalization, and accelerate the training. The network hyper-parameters were selected based on several experiments with different setups where the best hyper-parameters were used in our experiments that maximize the detection accuracy. The CNN was developed using Pytorch framework (https://pytorch.org/, accessed on 15 January 2023) and the training was performed using Nvidia GTX 1080.
To test the performance of the developed MGO, we compared it to several optimizers, namely, the traditional WOA [21], the traditional GO, grey wolf optimizer (GWO) [23], Transient Search Optimization (TSO) [24], firefly algorithm (FFA) [25], and moth flame optimization (MFO) [26]. We set the parameters of each algorithm based on its original implementation, whereas the iterations number is set to 50, and the agent number is 20.

4.3. Experimental Datasets

To validate the proposed framework, we used four well-known datasets for network intrusion detection, which are publicly available. The datasets used to train and test the proposed framework are KDDCup-99, NSL-KDD, BoT-IoT, and CICIDS-2017. Figure 3 shows the corresponding statistics of the datasets used in our experiments. The KDD (Knowledge Discovery and Data Mining) Cup 1999 dataset (KDDCup-99) was created in 1999 for the KDD cup competition organized by the Defence Advanced Research Project Agency (DARPA). The KDDCup-99 collects TCP/IP dump files containing network traffic records recorded for two months. The total number of records is around five million, with 41 features. We used the 10% version of the KDDCup-99 dataset, which contains less than one million records with four attack types, including user-to-root (U2R), probing, remote-to-user (R2L), and denial-of-service (DoS) besides normal traffic. The NSL-KDD dataset is a distilled version of the KDDCup-99 dataset with 41 features. The CICIDS-2017 [27] dataset provides more realistic network traffic records with 79 features collected by the CICFlowMeter tool focusing on SSH, HTTP, HTTPS, email, and FTP protocols. The dataset is presented in several CSV files, where we used four files to train and test the framework. The total used network records are around 600 thousand, with seven attack types and normal traffic. For the Bot-IoT dataset [28], we used the 5% version with 3.5 million records collected from various IoT devices.

4.4. Result and Discussions

This section discusses the comparison results between the developed MGO and other methods to improve the quality of IoT security. Table 1 shows the average over 25 independent runs for each algorithm using the performance measures.
In the multi-classification analysis, the MGO algorithm demonstrates superior efficiency compared to other algorithms during the learning phase on the datasets (i.e., KDD99, NSL-KDD, BIoT, and CIC2017). However, it falls behind the RSA regarding performance on the BIoT dataset. Furthermore, MGO excels in detecting attack types using testing samples on all four datasets compared to other methods.
In addition, the accuracy value of RSA is better than other methods, followed by MFO which allocates the third rank overall to the other methods during the training stage, whereas the Accuracy of RSA based on the testing set is the best after the developed MGO algorithm. Based on Precision, F1-Measure, and Recall, the FFA, GWO, and MFO is the best algorithm that allocates the second rank among the tested algorithms within the testing phase.
Figure 4 illustrates the average performance of each method across various datasets for various measures. The MGO method has the highest average performance for training and testing in multi-classification, followed by the Accuracy method which has better accuracy. The RSA method has better Recall in both training and testing, and GWO has a better F1-Measure in both training and testing. MFO and FFA have higher Precision in training and testing sets, respectively.
To further analyze the results, we applied the Friedman test to determine if there was a significant difference between the different methods. The Friedman test gave us the mean rank for each method, as seen in Table 2. From these mean ranks, we can see that MGO has the highest mean rank across all training and testing set performance measures, followed by RSA overall performance measures in case of using the training set. Meanwhile, the mean rank of the accuracy, precision, F1-measure, recall, and GM for the GO, FFA, MFO, TSO, and FFA, respectively, is the second rank after the developed MGO in the case of the testing set.

4.5. Comparison with Existing Methods

This section compares the results of the developed MGO with other techniques as given in Table 3. Most of those methods may use either one or two datasets. From these results, we can observe that in the case of KDD99, the accuracy of MGO is better than the method applied in [29]; however, the method presented in [30] has better performance than MGO. In the case of the BIoT, we can observe that the developed MGO performs better than other methods mentioned in this study, followed by the method introduced in Churcher et al. [31] ( KNN) which is superior to other methods. For the CICIDS2017 dataset, we noticed that MGO provided better results than the competitive methods.
From the previous results, it is clear that the developed method has a high potential to improve the prediction of attacks in IoT environments. However, the method has some limitations, such as being time-consuming due to the model learning process. These limitations can be addressed by using transfer learning techniques. In addition, the MGO still requires handling the imbalanced datasets in IoT and this can be handled by using the mechanism mentioned in [40].

5. Conclusions and Future work

Our study investigated the development of a two-phase framework to improve the detection accuracy over existing intrusion detection systems (IDS). In addition, the developed framework integrates a deep learning (DL) model and swarm intelligence (SI) technique to combine both techniques’ advantages and facilitate the deployment of the framework in the Internet of Things (IoT) system. We implemented a convolutional neural network architecture as a core feature extraction module to learn and extract new feature representation from the raw input data (network traffic records). In addition, we proposed a novel feature selection (FS) approach based on a modified variant of the Growth Optimizer (GO) algorithm to reduce the extracted feature representation space, speed up the inference, and improve the overall framework performance on IDS. The proposed FS method relies on applying the GWO to boost the search process of the traditional GO algorithm. Thus, the results show that the suggested method performed best compared to several optimization techniques using different evaluation indicators with several public IDS datasets. In future work, the developed MGO can be extended and experimented with in different applications such as healthcare, human activity recognition, fake news detection, and others.

Author Contributions

Conceptualization: A.F., A.D. and M.A.A.A.-q.; Data curation: A.F., M.A.A.A.-q., S.A.A. and S.S.A.; Formal analysis: M.A.A.A.-q., M.A.E., S.A.A. and S.S.A.; Funding acquisition: A.F., Investigation: A.F.; Methodology: A.F. and A.D.; Resources: A.F.; Project administration: A.F. and S.L.; Software: M.A.A.A.-q.; Supervision: A.F. and S.L.; Validation: S.L., S.A.A. and S.S.A.; Visualization: A.D. and M.A.A.A.-q.; Writing—original draft preparation: A.F., A.D. and M.A.A.A.-q.; Writing—review and editing: M.A.A.A.-q., S.L., S.A.A. and S.S.A. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Key R&D Program of China under Grand No. 2021YFB2012202 and the Hubei Provincial Science and Technology Major Project of China under Grant No. 2020AEA011 and the Key Research & Development Plan of Hubei Province of China under Grant No. 2021BAA171, 2021BAA038, 2020BAB100 and the project of Science, Technology and Innovation Commission of Shenzhen Municipality of China under Grant No. JCYJ20210324120002006 and JSGG20210802153009028, the 2021 Industrial Technology Basic Public Service Platform Project of the Ministry of Industry and Information Technology of PRC under Grand No. 2021-0171-1-1.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All the datasets are public, as we described in the main text.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Debicha, I.; Bauwens, R.; Debatty, T.; Dricot, J.M.; Kenaza, T.; Mees, W. TAD: Transfer learning-based multi-adversarial detection of evasion attacks against network intrusion detection systems. Future Gener. Comput. Syst. 2023, 138, 185–197. [Google Scholar] [CrossRef]
  2. Lata, S.; Singh, D. Intrusion detection system in cloud environment: Literature survey & future research directions. Int. J. Inf. Manag. Data Insights 2022, 2, 100134. [Google Scholar]
  3. Schueller, Q.; Basu, K.; Younas, M.; Patel, M.; Ball, F. A hierarchical intrusion detection system using support vector machine for SDN network in cloud data center. In Proceedings of the 2018 28th International Telecommunication Networks and Applications Conference (ITNAC), Sydney, Australia, 21–23 November 2018; pp. 1–6. [Google Scholar]
  4. Wei, J.; Long, C.; Li, J.; Zhao, J. An intrusion detection algorithm based on bag representation with ensemble support vector machine in cloud computing. Concurr. Comput. Pract. Exp. 2020, 32, e5922. [Google Scholar] [CrossRef]
  5. Peng, K.; Leung, V.; Zheng, L.; Wang, S.; Huang, C.; Lin, T. Intrusion detection system based on decision tree over big data in fog environment. Wirel. Commun. Mob. Comput. 2018, 2018. [Google Scholar] [CrossRef]
  6. Modi, C.; Patel, D.; Borisanya, B.; Patel, A.; Rajarajan, M. A novel framework for intrusion detection in cloud. In Proceedings of the Fifth International Conference on Security of Information and Networks, Jaipur, India, 25–27 October 2012; pp. 67–74. [Google Scholar]
  7. Kumar, G.R.; Mangathayaru, N.; Narasimha, G. An improved k-Means Clustering algorithm for Intrusion Detection using Gaussian function. In Proceedings of the International Conference on Engineering & MIS 2015, Istanbul, Turkey, 24–26 September 2015; pp. 1–7. [Google Scholar]
  8. Zhao, X.; Zhang, W. An anomaly intrusion detection method based on improved k-means of cloud computing. In Proceedings of the 2016 Sixth International Conference on Instrumentation & Measurement, Computer, Communication and Control (IMCCC), Harbin, China, 21–23 July 2016; pp. 284–288. [Google Scholar]
  9. Hodo, E.; Bellekens, X.; Hamilton, A.; Dubouilh, P.L.; Iorkyase, E.; Tachtatzis, C.; Atkinson, R. Threat analysis of IoT networks using artificial neural network intrusion detection system. In Proceedings of the 2016 International Symposium on Networks, Computers and Communications (ISNCC), Yasmine Hammamet, Tunisia, 11–13 May 2016; pp. 1–6. [Google Scholar]
  10. Almiani, M.; AbuGhazleh, A.; Al-Rahayfeh, A.; Atiewi, S.; Razaque, A. Deep recurrent neural network for IoT intrusion detection system. Simul. Model. Pract. Theory 2020, 101, 102031. [Google Scholar] [CrossRef]
  11. Wu, K.; Chen, Z.; Li, W. A novel intrusion detection model for a massive network using convolutional neural networks. IEEE Access 2018, 6, 50850–50859. [Google Scholar] [CrossRef]
  12. Alazab, M.; Khurma, R.A.; Awajan, A.; Camacho, D. A new intrusion detection system based on moth–flame optimizer algorithm. Expert Syst. Appl. 2022, 210, 118439. [Google Scholar] [CrossRef]
  13. Samadi Bonab, M.; Ghaffari, A.; Soleimanian Gharehchopogh, F.; Alemi, P. A wrapper-based feature selection for improving performance of intrusion detection systems. Int. J. Commun. Syst. 2020, 33, e4434. [Google Scholar] [CrossRef]
  14. Zhou, Y.; Cheng, G.; Jiang, S.; Dai, M. Building an efficient intrusion detection system based on feature selection and ensemble classifier. Comput. Netw. 2020, 174, 107247. [Google Scholar] [CrossRef]
  15. Mojtahedi, A.; Sorouri, F.; Souha, A.N.; Molazadeh, A.; Mehr, S.S. Feature Selection-based Intrusion Detection System Using Genetic Whale Optimization Algorithm and Sample-based Classification. arXiv 2022, arXiv:2201.00584. [Google Scholar]
  16. Talita, A.; Nataza, O.; Rustam, Z. Naïve bayes classifier and particle swarm optimization feature selection method for classifying intrusion detection system dataset. J. Phys. Conf. Ser. 2021, 1752, 012021. [Google Scholar] [CrossRef]
  17. Fatani, A.; Dahou, A.; Al-Qaness, M.A.; Lu, S.; Elaziz, M.A. Advanced feature extraction and selection approach using deep learning and Aquila optimizer for IoT intrusion detection system. Sensors 2021, 22, 140. [Google Scholar] [CrossRef]
  18. Dahou, A.; Abd Elaziz, M.; Chelloug, S.A.; Awadallah, M.A.; Al-Betar, M.A.; Al-qaness, M.A.; Forestiero, A. Intrusion Detection System for IoT Based on Deep Learning and Modified Reptile Search Algorithm. Comput. Intell. Neurosci. 2022, 2022, 6473507. [Google Scholar] [CrossRef]
  19. Karuppusamy, L.; Ravi, J.; Dabbu, M.; Lakshmanan, S. Chronological salp swarm algorithm based deep belief network for intrusion detection in cloud using fuzzy entropy. Int. J. Numer. Model. Electron. Netw. Devices Fields 2022, 35, e2948. [Google Scholar] [CrossRef]
  20. Zhang, Q.; Gao, H.; Zhan, Z.H.; Li, J.; Zhang, H. Growth Optimizer: A powerful metaheuristic algorithm for solving continuous and discrete global optimization problems. Knowl.-Based Syst. 2023, 261, 110206. [Google Scholar] [CrossRef]
  21. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  22. Abualigah, L.; Abd Elaziz, M.; Sumari, P.; Geem, Z.W.; Gandomi, A.H. Reptile Search Algorithm (RSA): A nature-inspired meta-heuristic optimizer. Expert Syst. Appl. 2022, 191, 116158. [Google Scholar] [CrossRef]
  23. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  24. Fatani, A.; Abd Elaziz, M.; Dahou, A.; Al-Qaness, M.A.; Lu, S. IoT Intrusion Detection System Using Deep Learning and Enhanced Transient Search Optimization. IEEE Access 2021, 9, 123448–123464. [Google Scholar] [CrossRef]
  25. Yang, X.S.; He, X. Firefly algorithm: Recent advances and applications. Int. J. Swarm Intell. 2013, 1, 36–50. [Google Scholar] [CrossRef]
  26. Mirjalili, S. Moth-flame optimization algorithm: A novel nature-inspired heuristic paradigm. Knowl.-Based Syst. 2015, 89, 228–249. [Google Scholar] [CrossRef]
  27. Sharafaldin, I.; Lashkari, A.H.; Ghorbani, A.A. Toward generating a new intrusion detection dataset and intrusion traffic characterization. ICISSp 2018, 1, 108–116. [Google Scholar]
  28. Koroniotis, N.; Moustafa, N.; Sitnikova, E.; Turnbull, B. Towards the development of realistic botnet dataset in the internet of things for network forensic analytics: Bot-iot dataset. Future Gener. Comput. Syst. 2019, 100, 779–796. [Google Scholar] [CrossRef]
  29. Wu, P. Deep learning for network intrusion detection: Attack recognition with computational intelligence. Ph.D. Thesis, UNSW Sydney, Kensington, Australia, 2020. [Google Scholar]
  30. Farahnakian, F.; Heikkonen, J. A deep auto-encoder based approach for intrusion detection system. In Proceedings of the 2018 20th International Conference on Advanced Communication Technology (ICACT), Chuncheon, Republic of Korea, 11–14 February 2018; pp. 178–183. [Google Scholar]
  31. Churcher, A.; Ullah, R.; Ahmad, J.; Ur Rehman, S.; Masood, F.; Gogate, M.; Alqahtani, F.; Nour, B.; Buchanan, W.J. An experimental analysis of attack classification using machine learning in IoT networks. Sensors 2021, 21, 446. [Google Scholar] [CrossRef]
  32. Ma, T.; Wang, F.; Cheng, J.; Yu, Y.; Chen, X. A hybrid spectral clustering and deep neural network ensemble algorithm for intrusion detection in sensor networks. Sensors 2016, 16, 1701. [Google Scholar] [CrossRef]
  33. Javaid, A.; Niyaz, Q.; Sun, W.; Alam, M. A deep learning approach for network intrusion detection system. In Proceedings of the 9th EAI International Conference on Bio-Inspired Information and Communications Technologies (Formerly BIONETICS), New York, NY, USA, 3–5 December 2015; pp. 21–26. [Google Scholar]
  34. Tang, T.A.; Mhamdi, L.; McLernon, D.; Zaidi, S.A.R.; Ghogho, M. Deep learning approach for network intrusion detection in software defined networking. In Proceedings of the 2016 International Conference on Wireless Networks and Mobile Communications (WINCOM), Fez, Morocco, 26–29 October 2016; pp. 258–263. [Google Scholar]
  35. Imamverdiyev, Y.; Abdullayeva, F. Deep learning method for denial of service attack detection based on restricted boltzmann machine. Big Data 2018, 6, 159–169. [Google Scholar] [CrossRef]
  36. Alkadi, O.; Moustafa, N.; Turnbull, B.; Choo, K.K.R. A deep blockchain framework-enabled collaborative intrusion detection for protecting IoT and cloud networks. IEEE Internet Things J. 2020, 8, 9463–9472. [Google Scholar] [CrossRef]
  37. Vinayakumar, R.; Alazab, M.; Soman, K.; Poornachandran, P.; Al-Nemrat, A.; Venkatraman, S. Deep learning approach for intelligent intrusion detection system. IEEE Access 2019, 7, 41525–41550. [Google Scholar] [CrossRef]
  38. Laghrissi, F.; Douzi, S.; Douzi, K.; Hssina, B. Intrusion detection systems using long short-term memory (LSTM). J. Big Data 2021, 8, 65. [Google Scholar] [CrossRef]
  39. Alkahtani, H.; Aldhyani, T.H. Intrusion detection system to advance internet of things infrastructure-based deep learning algorithms. Complexity 2021, 2021, 5579851. [Google Scholar] [CrossRef]
  40. Luque, A.; Carrasco, A.; Martín, A.; de Las Heras, A. The impact of class imbalance in classification performance metrics based on the binary confusion matrix. Pattern Recognit. 2019, 91, 216–231. [Google Scholar] [CrossRef]
Figure 1. The CNN architecture employed for feature extraction.
Figure 1. The CNN architecture employed for feature extraction.
Sensors 23 04430 g001
Figure 2. The workflow structure of the MGO feature selection technique.
Figure 2. The workflow structure of the MGO feature selection technique.
Sensors 23 04430 g002
Figure 3. Datasets statistics where the Test set is the outer circle and the Train set is the inner circle. (a) KDDCup-99. (b) NSL-KDD. (c) CICIDS-2017. (d) Bot-IoT.
Figure 3. Datasets statistics where the Test set is the outer circle and the Train set is the inner circle. (a) KDDCup-99. (b) NSL-KDD. (c) CICIDS-2017. (d) Bot-IoT.
Sensors 23 04430 g003
Figure 4. The average overall of the tested sets for multi-classification. (a) Train Set. (b) Testing set.
Figure 4. The average overall of the tested sets for multi-classification. (a) Train Set. (b) Testing set.
Sensors 23 04430 g004
Table 1. The performance of MGO using the IoT security datasets. (Note: bold indicate best results).
Table 1. The performance of MGO using the IoT security datasets. (Note: bold indicate best results).
TrainTest
AccuracyPrecisionF1RecallGMAccuracyPrecisionF1RecallGM
KDD99GO98.51593.48392.65291.83592.65590.61584.24983.79783.35083.798
WOA92.27592.41497.30493.12692.76984.37582.50187.35185.22583.852
TSO95.43991.02797.43794.91992.95387.53680.79187.47987.01683.846
GWO95.51394.06298.48292.38393.21987.61884.13188.53384.48884.310
MFO96.07397.63198.37197.12397.37788.17587.76388.42089.22588.491
FFA91.98897.32891.53893.36895.32884.31891.60984.28585.69888.604
RSA99.91099.90999.90699.91099.91092.04089.68489.98592.04090.855
MGO99.91099.95999.94699.93399.94692.04090.84190.94191.04090.941
NSL-KDDGO97.10895.10493.01791.02093.04070.22472.20069.79467.54469.833
WOA91.94792.08096.96892.79792.43867.95171.13168.90768.80169.956
TSO95.07890.65797.06794.55892.58771.33071.29869.69770.81071.053
GWO95.18293.72498.14392.05292.88471.06672.15169.94867.93670.012
MFO95.74597.29798.03596.79597.04671.62676.12269.84472.67674.379
FFA91.66096.99191.20193.04094.99567.43775.87362.94468.81772.259
RSA99.20199.15899.14899.20199.18076.10782.17171.73176.10779.081
MGO99.21499.45899.43799.41699.43776.72583.10579.75976.67279.824
BIoTGO99.06899.10799.07699.04599.07699.14198.10098.37198.64498.372
WOA99.47299.47299.47299.47299.47298.95698.95799.00598.96498.960
TSO99.46099.45999.45999.46099.46098.98698.94199.00598.98198.961
GWO99.47799.47699.47699.47799.47798.99098.97599.01998.95998.967
MFO99.48099.48099.48099.48099.48098.99899.01399.02099.00999.011
FFA99.47999.47899.47899.47999.47898.95499.00798.94998.96898.987
RSA98.82998.82998.82998.82998.82999.02099.09899.07099.03899.068
MGO99.62999.52999.62999.72999.62999.22099.18899.21899.24899.218
CIC2017GO99.13099.23999.20499.17099.20499.17099.02099.21599.41099.215
WOA99.69099.49099.45099.69099.59099.43099.24099.19099.43099.335
TSO99.68099.75099.68099.71099.73099.42099.48099.42099.45099.465
GWO99.37099.43099.38099.56099.49599.11099.18099.12099.30099.240
MFO99.36099.37099.48099.43099.40099.10099.12099.22099.17099.145
FFA99.45099.48099.60099.74099.61099.20099.22099.35099.49099.355
RSA99.91199.91099.88999.91199.91099.91199.90799.88899.91199.909
MGO99.94199.92099.92699.93199.92699.94199.94799.94299.93699.942
Table 2. Results of Friedman test.
Table 2. Results of Friedman test.
GOWOATSOGWOMFOFFARSAMGO
TrainingAccuracy3.75003.50003.50004.00004.75003.00005.62507.8750
Precision2.50003.25002.75003.75005.25005.00005.50008.0000
F1-Measure1.75003.25004.25004.75005.25003.25005.50008.0000
Recall1.25003.50004.50003.00005.25005.00005.50008.0000
GM2.00002.75003.50003.75005.25005.25005.50008.0000
TestingAccuracy4.75003.00004.00003.50004.25001.75006.87507.8750
Precision2.50002.75002.75003.25004.75005.50006.75007.7500
F1-Measure2.25002.62504.12504.50005.00002.50007.00008.0000
Recall1.50003.25005.00002.00004.75004.50007.25007.7500
GM1.25002.75003.75003.50004.50005.25007.00008.0000
Table 3. Comparison with other methods.
Table 3. Comparison with other methods.
DatasetWorkAccuracy
KDD Cup 99Wu [29]85.24
Farahnakian et al. [30]96.53
MGO0.9204
NSL-KDDMa et al. [32] SCDNN72.64
Javaid et al. [33] STL74.38
Tang et al. [34] DNN75.75
Imamverdiyev et al. [35] Gaussian–Bernoulli RBM73.23
MGO76.725
BIoT[36] (BiLSTM)98.91
Alkadi et al. [36] (NB)97.5
Alkadi et al. [36] (SVM)97.8
Churcher et al. [31] (KNN)99
Churcher et al. [31] (SVM)79
Churcher et al. [31] (DT)96
Churcher et al. [31] (NB)94
Churcher et al. [31] (RF)95
Churcher et al. [31] (ANN)97
Churcher et al. [31] (LR)74
MGO99.22
CICIDS2017Vinayakumar [37]94.61
Laghrissi et al. [38]85.64
Alkahtani et al. [39]80.91
MGO99.941
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fatani, A.; Dahou, A.; Abd Elaziz, M.; Al-qaness, M.A.A.; Lu, S.; Alfadhli, S.A.; Alresheedi, S.S. Enhancing Intrusion Detection Systems for IoT and Cloud Environments Using a Growth Optimizer Algorithm and Conventional Neural Networks. Sensors 2023, 23, 4430. https://doi.org/10.3390/s23094430

AMA Style

Fatani A, Dahou A, Abd Elaziz M, Al-qaness MAA, Lu S, Alfadhli SA, Alresheedi SS. Enhancing Intrusion Detection Systems for IoT and Cloud Environments Using a Growth Optimizer Algorithm and Conventional Neural Networks. Sensors. 2023; 23(9):4430. https://doi.org/10.3390/s23094430

Chicago/Turabian Style

Fatani, Abdulaziz, Abdelghani Dahou, Mohamed Abd Elaziz, Mohammed A. A. Al-qaness, Songfeng Lu, Saad Ali Alfadhli, and Shayem Saleh Alresheedi. 2023. "Enhancing Intrusion Detection Systems for IoT and Cloud Environments Using a Growth Optimizer Algorithm and Conventional Neural Networks" Sensors 23, no. 9: 4430. https://doi.org/10.3390/s23094430

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop