Next Article in Journal
A Novel and Safe Approach to Simulate Cutting Movements Using Ground Reaction Forces
Next Article in Special Issue
A Hypergraph-Based Blockchain Model and Application in Internet of Things-Enabled Smart Homes
Previous Article in Journal
Automated Methodology for Dependability Evaluation of Wireless Visual Sensor Networks
Previous Article in Special Issue
A Collaborative Data Collection Scheme Based on Optimal Clustering for Wireless Sensor Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

TLTD: A Testing Framework for Learning-Based IoT Traffic Detection Systems

1
School of Information and Software Engineering, University of Electronic Science and Technology of China, Chengdu 610054, China
2
Cyberspace Security Research Center, University of Electronic Science and Technology of China, Chengdu 611731, China
3
Department of Electrical and Computer Engineering, Purdue University, West Lafayette, IN 47907, USA
4
Department of Computer and Information Sciences, Temple University, Philadelphia, PA 19122, USA
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(8), 2630; https://doi.org/10.3390/s18082630
Submission received: 27 June 2018 / Revised: 3 August 2018 / Accepted: 3 August 2018 / Published: 10 August 2018
(This article belongs to the Special Issue Sensor Networks for Collaborative and Secure Internet of Things)

Abstract

:
With the popularization of IoT (Internet of Things) devices and the continuous development of machine learning algorithms, learning-based IoT malicious traffic detection technologies have gradually matured. However, learning-based IoT traffic detection models are usually very vulnerable to adversarial samples. There is a great need for an automated testing framework to help security analysts to detect errors in learning-based IoT traffic detection systems. At present, most methods for generating adversarial samples require training parameters of known models and are only applicable to image data. To address the challenge, we propose a testing framework for learning-based IoT traffic detection systems, TLTD. By introducing genetic algorithms and some technical improvements, TLTD can generate adversarial samples for IoT traffic detection systems and can perform a black-box test on the systems.

1. Introduction

In recent years, with the extensive applications of the Internet of Things (IoT) and the continuous development of machine learning algorithms, many researchers have proposed a number of IoT malicious traffic detection techniques based on machine learning algorithms [1,2,3,4,5,6]. However, most researchers only care about the performance of the detection model but ignore the vulnerability and robustness of the model. This makes the existing model vulnerable to “adversarial samples” [7]. It can make the model misjudgde and then enable the attacker to achieve the purpose of bypassing the model detection [8]. Such a model cannot be applied in practice. The adversarial sample is a special sample deliberately designed by the attacker. When it is input into the machine learning model, it can cause model classification errors. It is just like the visual illusion of the model. This type of adversarial sample designed for machine learning models has recently attracted the attention of many researchers [9,10,11]. For example, in the image classification system, by adding a slight perturbation to the original image, the change of the image classification result can be achieved with a high probability, and even the attacker can make the image be classified as an arbitrarily designated label [8]. Goodfellow et al. proposed a method based on fast gradient symbol algorithm to generate an adversarial sample [12]. Papernot et al. used the Jacobian matrix to determine which dimensions of information need to be modified when generating an adversarial sample [13]. In fact, the Jacobian matrix-based algorithm is also a gradient algorithm. Grosse et al. used a gradient-based algorithm to generate an adversarial sample of Android malware [14]. They assumed that the attacker can know the parameters of the malware detection model. For different neural networks, the model’s misclassification rate was shown to be 40% to 84% after processing the adversarial samples.
At present, most adversarial sample generation methods are designed for image classification models [15,16,17]. However, almost all machine learning models can be attacked by the adversarial samples. This means that there are also security risks in other areas of machine learning algorithm applications [18,19]. On the other hand, most of the current methods use gradient information to transform the original data into adversarial samples. If an attacker only knows what features the model uses, and he knows nothing about the parameters of the model, he cannot produce an effective adversarial sample [15,20].
This paper proposes a testing framework for learning-based IoT traffic detection system (TLTD). The main challenge of TLTD is the method of generating adversarial samples for the IoT traffic detection model based on a genetic algorithm. Without having to know the model parameters, TLTD uses the original data as the input of the algorithm, and produces the counter sample of the specific tag, and the only information used is the classification probability of the tag’s output by the model.
We hope TLTD can be a benchmark learning-based IoT traffic detection model. Our contribution is mainly reflected as follows:
  • We migrate the application scenarios of adversarial samples from the image recognition domain to the IoT malicious traffic detection field. This migration cannot be achieved simply by replacing the model’s training data from pictures to traffic. We need to do specific technical processing on the traffic data to ensure the validity of the adversarial sample.
  • We introduce the genetic algorithm into the method of generating the adversarial sample and realize the black-box attack against the machine learning model.
  • Our approach is equally valid for networks that have difficulty computing gradients or expressing mathematically.
The rest of the paper is organized as follows. Section 2 introduces the related work of adversarial samples. Section 3 presents TLTD (testing framework for learning-based IoT traffic detection system). Section 4 presents and discusses our experimental results. Finally, further discussions and conclusions are accomplished in Section 5.

2. Related Work

Given a primitive input (X), a target tag (t) and L ( x ) ! = t , a similar input X is found to make L ( X ) = t . The special sample ( x ) with this characteristic is called an adversarial sample of the target attack. Similarity can be measured according to a distance algorithm [21]. Here are some common generation methods of adversarial samples and their application.

2.1. Fast Gradient Sign

Ian J. Goodfellow proposed a fast gradient sign to generate adversarial samples in 2015 [9]. The idea is to move every dimension of the sample to a small step toward decreasing confidence. Since input samples are usually multi-dimensional and activation functions ReLU (Rectified Linear Unit) are highly linear, such changes can affect the classification results.
The disturbance function is as follows:
η = ϵ s i g n ( x J ( θ , x , y ) )
where θ is the parameter of the model, x is the input of the model, y is the output class relative to x, and J ( θ , x , y ) is the loss function in the neural network.
So, the perturbed adversarial samples are
x + ϵ s i g n ( x J ( θ , x , y ) ) ,
where ϵ is set to be small enough to be difficult to distinguish. From an intuitive point of view, the fast gradient sign uses the gradient of the loss function to determine which direction each pixel should ultimately change in order to minimize the loss function. In the end, all pixels change in a certain direction by a certain size.

2.2. One Pixel Attack

Su et al. proposed one-pixel attack, which is an adversarial sample generation method based on differential evolution [22]. The typical adversarial sample generation method allows perturbation of all pixels; however, the approach considered by this method is reversed, focusing only on the number of modified pixels without limiting the size of a single variation.
m i n l o s s F , t ( x ) , x [ 0 , 1 ] n
s u b j e c t t o x x 0 d
where d is an integer— d = 1 in the case of a single pixel attack.
Evolutionary algorithms do not require the gradient of the model to be solved. This is an advantage of this type of algorithm. However, this method is actually focused on solving the problem of single-pixel adversarial sample generation. The visual effect of the sample is not optimized. Adversarial samples have significant noise compared to the original sample.

2.3. Application of Adversarial Samples in Malware Detection

The way to generate adversarial samples by adding disturbance cannot only be applied in the fields of computer vision, speech, and natural language, but also in the field of network security. For example, when the adversarial sample is applied to the malware classification model based on machine learning, a slight disturbance is added to the malicious software without changing the attackability of the malware so that the classification model in the machine learning is misjudged to be the normal software.
Kathrin Grosse and Nicolas Papernot et al. applied the method of generating adversarial samples in the field of computer vision to malware classification [14]. In their method, the feature of the input sample is represented by a binary vector. Given a number of behavioral features ( 1 , , M ), a software application can be represented by a set of binary vectors ( X { 0 , 1 } M ), where X i indicates whether the ith behavior is allowed. For a single input sample X, the classifier returns a two-dimensional vector F ( X ) = [ F 0 ( X ) , F 1 ( X ) ] , where F 0 ( X ) indicates that the software is normal software probability, and F 1 ( X ) indicates that the software is a malware probability and satisfies the constraint F 0 ( X ) + F 1 ( X ) = 1 .
The gradient of the input sample is calculated by the classification result to find the direction that most likely causes the classification result to change, as follows:
J F = F ( X ) X = [ F i ( X ) X j ] i 0 , 1 , j [ 1 , m ] .
The method in this paper enables the classifier to misjudge malware as normal software, and its success rate is about 85%.
Compared to image data, there are the following differences in adversarial samples against malicious traffic or software:
(1)
The input samples in the image are all pixels and the values of the pixels are continuous. However, in the field of network security, input characteristics are usually discrete and the range of values of different features is usually different.
(2)
The pixels in the image can be freely changed within the value range. The restrictions on the modification of traffic or software are much more demanding. Arbitrary modifications may result in traffic or software not working properly.

3. Methodology

3.1. Framework

TLTD can detect the captured IoT traffic. The overview of TLTD is shown in Figure 1. The captured flow data and detection results are entered into our testing framework as test samples. Through multiple iterations, TLTD generates adversarial samples for the system, and then the system is tested for security. It can be seen that our testing framework can be well integrated with machine learning-based IoT traffic detection systems. In fact, our testing framework is suitable for the testing of a wider range of machine learning-based network traffic detection model systems, but we only use IoT traffic detection as an example. When the test results show that the detection system cannot resist the attack against the adversarial sample, this indicates that the system has potential safety hazards and it is necessary to implement such reinforcement measures as a distillation defense on the detection system. As we can see, determining how to generate an adversarial sample against IoT traffic is the main challenge of this testing framework. Therefore, we describe, in detail, the algorithm for generating an adversarial sample for IoT traffic.

3.2. Algorithm

The goal of the adversarial sample generation algorithm is to add a slight disturbance to the captured malicious IoT traffic so that the previously trained detection model misjudges it as normal traffic. Similar to the aforementioned adversarial sample for the malware detection model, for one input sample (X), the classifier returns a two-dimensional vector F ( X ) = [ F 0 ( X ) , F 1 ( X ) ] , where F 0 ( X ) indicates the probability that the traffic is normal traffic, and F 1 ( X ) indicates the probability that the traffic is malicious traffic and satisfies the constraint F 0 ( X ) + F 1 ( X ) = 1 . The aim is to add a perturbation ( δ ), so that the classification result F 0 ( X + δ ) of malicious samples (X) is less than F 1 ( X + δ ) , and the smaller the perturbation δ is, the better. This is equivalent to minimizing the number of disturbances and minimizing the degree of perturbation of each feature. In order to ensure the effectiveness of the traffic after the disturbance, we limited the type and amplitude of disturbance features. Details are described in Section 4. The difference is that we used genetic algorithms to select the degree of perturbation of the feature and thus realized black-box attacks against the machine learning model. The pseudo code of our algorithm is shown in Algorithm 1.
Algorithm 1 Generating an adversarial sample.
Require: Population Size p o p _ s i z e , Number of features f e a t _ n u m , Original sample X i
     P i i n i t i a l i z a t i o n ( )
    for i = 0 p o p _ s i z e do
        P i C r o s s o v e r _ O p e r a t o r ( )
        P i M u t a t i o n _ O p e r a t o r ( )
       for j = 0 f e a t _ n u m do
          if P i j > 0 then
             Compute δ i = P i j ( u p p e r i j X i j )
         else
             Compute δ i = P i j ( X i j l o w e r j )
         end if
       end for
       Compute X i + 1 j = X i j + δ i
       if F ( X i + 1 ) < 1 F ( X i + 1 ) then
          Continue
       else
          Output P i
       end if
    end for
For example, the original sample is
X i = [ X i 0 , X i 1 , , X i n ] .
One of the individuals in the genetic algorithm is
P i = [ P i 0 , P i 1 , , P i n ] .
The range of values for each dimension in the original data is
R j = [ l o w e r j , u p p e r j ] , j [ 0 , n ] .
The range of values for each dimension in P i j is [ 1 , 1 ] are now specified. When P i j is 0, it means that the i-th feature of the original sample does not mutate. When P i j is negative, the original sample changes toward the smaller direction, and the degree of change is δ :
δ i = P i j ( X i j l o w e r j ) , j [ 0 , n ] .
When P i j is positive, the original sample changes in the direction of increasing, and its degree of change is δ :
δ i = P i j ( u p p e r i j X i j ) , j [ 0 , n ] .
So, the formula for the perturbed data is
X i j = X i j + P i j ( X i j l o w e r j ) P i j > 0 X i j + P i j ( u p p e r i j X i j ) P i j 0 .
Finally, we get the formula for the fitness function in the genetic algorithm:
F ( P i ) = ω d D ( P i ) + ω e ( 1 E ( P i ) ) ,
where ω d and ω e are two parameters. D ( P i ) is the degree of difference between the original sample and the adversarial sample. E ( P i ) is the success rate of adversarial samples in learning-based models.
Through the preceding fitness function, the population (P) is actually divided into two sections, as shown in Figure 2. The whole optimization process can be divided into three steps.
Step 1.
At this time, the adversarial sample cannot successfully mislead the classifier. Individuals at the top of section A gradually approach the bottom through crossover and mutation operators.
Step 2.
The individuals move from Section A to Section B, indicating that E ( P i ) = 1 , i.e., the adversarial samples generated at this time can successfully mislead the classifier.
Step 3.
Individuals at the top of Section B gradually approach the bottom, indicating the improvement of the similarity between the adversarial traffic and the original traffic.
Eventually, the bottom individual of Section B becomes the optimal individual in the population, and the information that it carries is the adversarial sample being sought out.

4. Experiments

4.1. Data Set and Environment

The dataset used in our experiments was KDD99 (Knowledge Discovery and Data Mining). Each sample in KDD99 contains 41 features. The 41 features can be divided into three categories:
  • Basic features of individual TCP connections;
  • Content features within a connection suggested by domain knowledge;
  • Traffic features computed using a two-second time window.
The hardware environment and software environment for all experiments are shown in Table 1.

4.2. IoT Traffic Detection Model

First, the detection model is trained to determine whether the traffic generated by the IoT device is malicious traffic. When the detection model reaches a certain accuracy, it combines with the method proposed in this paper to generate an adversarial sample and test the security of the model. Because the amount of data between different categories in KDD99 has changed significantly, we selected 4 categories with a large amount of data for perturbation, as shown in Table 2.
Neural network models are established for each of satan, ipsweep, portsweep, and nmap to find the recognition results. The detection model uses a fully connected network with a network structure of 32 × 64 × 2 . The last layer output 0 indicates that the data is identified as a normal traffic type, and output 1 indicates the specified traffic type.
Finally, the detection rate of the detection model is as shown in the Table 3.

4.3. Simulation Experiments

We conducted a total of three sets of simulation experiments. There were slight differences in the technical implementation methods for generating adversarial samples in the simulation experiments. In each set of simulation experiments, we tested the four types of traffic detection models respectively. We hoped to select the best technical implementation by comparing the test results.
The experimental parameters are shown in Table 4.
The tournament selection strategy takes a certain number of individuals from the population each time, and then selects the best one of them into the offspring population. This operation is repeated until the new population size reaches the original population size. The specific steps are as follows:
STEP 1.
Determine the number of individuals selected each time;
STEP 2.
Choose individuals randomly from the population and select the individuals with the best fitness values to enter the offspring population;
STEP 3.
Repeat STEP 2 for several times and the resulting individuals constitute a new generation of the population.

4.3.1. TLTD-I

(A) Preprocessing
The numerical distribution of 13 features in the data set is more concentrated and less variable. If we disturb these features, we will destroy the inherent distribution of traffic characteristics. This makes the generated adversarial sample clearly distinguishable from normal traffic. In order to ensure the effectiveness of the countermeasures, we chose to eliminate these 13 data features; that is, the values of these 13 features were not changed when disturbed. The 13 features which were not used in TLTD-I are as follows: hot, num_failed_logins, logged_in, num_compromised, root_shell, su_attempted, num_root, num_file_creations, num_shells, num_access_files, num_outbound_cmds, is_hot_login, is_guest_login.
(B) Fitness Function
The fitness function in TLTD-I is shown in Equation(12). We used the Euclidean distance to describe the degree of difference between the original sample and the adversarial sample:
D ( P i ) = s u m j n ( X i j X i j ) 2 , j [ 0 , n ] .
(C) Experimental Results
The experimental results are shown in the Figure 3 and Table 5. According to the success rate of adversarial samples, the success rate of the other three classes was over 95% except for the low success rate of the adversarial samples on nmap traffic. However, there were very large changes in the values of some features in the disturbed flow data. Combining with actual scenarios, when these values change drastically, it is very likely to undermine the effectiveness of IoT traffic. Therefore, TLTD-I may not be suitable for practical applications.

4.3.2. TLTD-II

(A) Data Preprocessing
In TLTD-II, a zero-mean normalization method was used to normalize all original data sets to a data set with a mean of 0 and a variance of 1.
The standardized formula is
z = x μ σ ,
where z is the normalized value, μ is the mean of the original data, and σ is the standard deviation of the original data. In addition, in order to simulate a more realistic IoT traffic detection environment, we also removed some of the data features that are difficult to obtain in real-world scenarios. Finally, we perturbed 22 features of the data.
(B) Fitness Function
In TLTD-II, we improved Equation (12) as the following equation:
F ( P i ) = e D ( P i ) ω d + e E ( P i ) ω e .
When the original sample is quite different from the disturbed sample, the left part has a larger value. In order to reduce the value of F ( P i ) , the left part is the leading factor in the optimization of the whole objective function. When the classification results of the neural network model do not change, the value of the right half will be very large. In order to reduce the value of the fitness function, the right half becomes the leading factor. In this case, the fitness function will be optimized in the direction of less difference from the original sample, and at the same time, the recognition result of the classifier will be changed. In brief, this improved fitness function satisfies the requirements of evolution orientation mentioned in the Section 3.2. On the other hand, its gradient is dynamic in the process of optimization, and it can find the optimal solution faster.
(C) Experimental Results
The experimental results of TLTD-II are shown in Figure 4 and Table 6. Compared with TLTD-I, the average D ( P i ) and the number of modified features were significantly reduced in TLTD-II. This means that we generated adversarial samples with fewer perturbations and fewer feature values. However, it also led to a bad result, that is, the success rate of the adversarial sample became very low. So, we propose TLTD-III and hope that the success rate can be further increased on the basis of TLTD-II.

4.3.3. TLTD-III

In TLTD-III, the data pre-processing process and the fitness function were all those conceived in TLTD-II. We believe that the very low success rate of TLTD-II may be related to the large range of feature disturbances. In addition, in order to make the adversarial samples more similar to the original data, we defined the maximum range of variation for each type of feature as [−50%, 50%]. The experimental results are shown in Figure 5 and Table 7.
From the experimental results, it can be seen that the success rate of TLTD-III was almost 100% of the average of D ( P i ) and the number of modified features was also small. This shows that TLTD-III has a certain application value in the actual scene.

4.4. Discussion

We compared the average success rate of adversarial samples and the perturbation size in three experiments. The results re shown in Figure 6. Compared with TLTD-I and TLTD-II, by modifying the normalization method, removing the unrelated feature data, and limiting the range of disturbance, we realized the dual purpose of a high success rate and low disturbance in TLTD-III. Although TLTD-III is still unable to ensure that the generated adversarial samples maintain the functionality of the original samples, we can still say that TLTD-III is an effective testing framework for learning-based IoT traffic detection systems.

5. Conclusions

To address the challenge of a lack of the testing framework for learning-based IoT traffic detection systems, we developed TLTD. Our experimental results show that our approach generates high-quality adversarial samples with the success rate closed to 100%. In the technical implementation of the TLTD algorithm, the selection of normalization method, features, and the range of disturbance are the keys to a good result. We hope TLTD can be a benchmark learning-based IoT traffic detection model. Our future work includes designing an effective defense approach to reinforce the traffic detection model.

Author Contributions

Data curation, J.L.; Formal analysis, X.L.; Investigation, X.L.; Methodology, X.L.; Supervision, Q.Z.; Validation, N.G.; Writing—original draft, X.L.; Writing—review & editing, X.Z. and X.D.

Funding

This research was funded by the National Natural Science Foundation of China under Grant CNS 61572115, in part by the National Key R&D Plan under Grant CNS 2016QY06X1205, and in part by the Science and Technology Project of State Grid Corporation of China under Grant CNS 522722180007.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Doshi, R.; Apthorpe, N.; Feamster, N. Machine Learning DDoS Detection for Consumer Internet of Things Devices. arXiv 2018, arXiv:1804.04159. [Google Scholar]
  2. Hodo, E.; Bellekens, X.; Hamilton, A.; Dubouilh, P.L.; Iorkyase, E.; Tachtatzis, C.; Atkinson, R. Threat analysis of IoT networks using artificial neural network intrusion detection system. In Proceedings of the 2016 International Symposium on Networks, Computers and Communications (ISNCC), Yasmine Hammamet, Tunisiam, 11–13 May 2016; IEEE: Piscataway, NJ, USA, 2016. [Google Scholar] [Green Version]
  3. Bull, P.; Austin, R.; Popov, E.; Sharma, M.; Watson, R. Flow based security for IoT devices using an SDN gateway. In Proceedings of the 2016 IEEE 4th International Conference on Future Internet of Things and Cloud (FiCloud), Vienna, Austria, 22–24 August 2016; IEEE: Piscataway, NJ, USA, 2016. [Google Scholar]
  4. Nobakht, M.; Sivaraman, V.; Boreli, R. A host-based intrusion detection and mitigation framework for smart home IoT using OpenFlow. In Proceedings of the 2016 11th International Conference on Availability, Reliability and Security (ARES), Salzburg, Austria, 31 August–2 September 2016; IEEE: Piscataway, NJ, USA, 2016. [Google Scholar]
  5. Du, X.J.; Yang, X.; Mohsen, G.Z.; Chen, H.H. An effective key management scheme for heterogeneous sensor networks. Ad Hoc Netw. 2007, 5, 24–34. [Google Scholar] [CrossRef]
  6. Du, X.J.; Mohsen, G.Z.; Yang, X.; Chen, H.H. A routing-driven elliptic curve cryptography based key management scheme for heterogeneous sensor networks. IEEE Trans. Wireless Commun. 2009, 8, 1223–1229. [Google Scholar] [CrossRef]
  7. Barreno, M.; Nelson, B.; Joseph, A.D.; Tygar, J. The security of machine learning. Mach. Learn. 2010, 81, 121–148. [Google Scholar] [CrossRef] [Green Version]
  8. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. arXiv 2013, arXiv:1312.6199. [Google Scholar]
  9. Dezfooli, S.M.M.; Fawzi, A.; Fawzi, O.; Frossard, P. Universal adversarial perturbations. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. Number EPFL-CONF-226156. [Google Scholar]
  10. Fawzi, A.; Dezfooli, S.M.M.; Frossard, P. A Geometric Perspective on the Robustness of Deep Networks; Technical Report; Institute of Electrical and Electronics Engineers: Piscataway, NJ, USA, 2017. [Google Scholar]
  11. Gu, S.; Rigazio, L. Towards deep neural network architectures robust to adversarial examples. arXiv 2014, arXiv:1412.5068. [Google Scholar]
  12. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. arXiv 2014, arXiv:1412.6572. [Google Scholar]
  13. Papernot, N.; McDaniel, P.; Jha, S.; Fredrikson, M.; Celik, Z.B.; Swami, A. The limitations of deep learning in adversarial settings. In Proceedings of the 2016 IEEE European Symposium on Security and Privacy (EuroS&P), Saarbrücken, Germany, 21–24 March 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 372–387. [Google Scholar]
  14. Grosse, K.; Papernot, N.; Manoharan, P.; Backes, M.; McDaniel, P. Adversarial perturbations against deep neural networks for malware classification. arXiv 2016, arXiv:1606.04435. [Google Scholar]
  15. Hu, W.; Ying, T. Generating adversarial malware examples for black-box attacks based on GAN. arXiv 2017, arXiv:1702.05983. [Google Scholar]
  16. Kurakin, A.; Goodfellow, I.; Bengio, S. Adversarial examples in the physical world. arXiv 2016, arXiv:1607.02533. [Google Scholar]
  17. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
  18. Yang, W.; Kong, D.; Xie, T.; Gunter, C.A. Malware detection in adversarial settings: Exploiting feature evolutions and confusions in android apps. In Proceedings of the 33rd Annual Computer Security Applications Conference, Orlando, FL, USA, 4–8 December 2017; ACM: New York, NY, USA, 2017. [Google Scholar]
  19. Demontis, A.; Melis, M.; Biggio, B.; Maiorca, D.; Arp, D.; Rieck, K.; Corona, I.; Giacinto, G.; Roli, F. Yes, machine learning can be more secure! A case study on Android malware detection. IEEE Trans. Dependable Secur. Comput. 2017. [Google Scholar] [CrossRef]
  20. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2014; pp. 2672–2680. [Google Scholar]
  21. Carlini, N.; Wagner, D. Towards evaluating the robustness of neural networks. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–24 May 2017; IEEE: Piscataway, NJ, USA, 2017. [Google Scholar]
  22. Su, J.; Vargas, D.V.; Kouichi, S. One pixel attack for fooling deep neural networks. arXiv 2017, arXiv:1710.08864. [Google Scholar]
Figure 1. Overview of our testing framework for learning-based IoT traffic detection systems.
Figure 1. Overview of our testing framework for learning-based IoT traffic detection systems.
Sensors 18 02630 g001
Figure 2. Individuals’ Distribution Diagram.
Figure 2. Individuals’ Distribution Diagram.
Sensors 18 02630 g002
Figure 3. Comparison of the number of perturbation features and the size of the disturbance in TLTD-I.
Figure 3. Comparison of the number of perturbation features and the size of the disturbance in TLTD-I.
Sensors 18 02630 g003
Figure 4. Comparison of the number of perturbation features and the size of the disturbance in TLTD-II.
Figure 4. Comparison of the number of perturbation features and the size of the disturbance in TLTD-II.
Sensors 18 02630 g004
Figure 5. Comparison of the number of perturbation features and the size of the disturbance in TLTD-III.
Figure 5. Comparison of the number of perturbation features and the size of the disturbance in TLTD-III.
Sensors 18 02630 g005
Figure 6. Comparison of the results of three experiments.
Figure 6. Comparison of the results of three experiments.
Sensors 18 02630 g006
Table 1. The environment of all experiments.
Table 1. The environment of all experiments.
CPU (Central Processing Unit)Inter(R) Core(TM) i507400 CPU @ 3.00 GHz
Memory8 GB
Video CardInter(R) HD Graphics 630
Operating SystemWindows 10
Programming LanguagePython 3.6
Development PlatformJupyter Notebook
DependenceTensorflow, numpy etc.
Table 2. The categories we selected in KDD99 (Knowledge Discovery and Data Mining).
Table 2. The categories we selected in KDD99 (Knowledge Discovery and Data Mining).
CategorySatanIpsweepPortsweepNmap
Amount15,89212,38110,4132316
Table 3. The detection rates of the models.
Table 3. The detection rates of the models.
CategorySatanIpsweepPortsweepNmap
Detection Rate0.99400.98050.99310.9330
Table 4. The parameters of TLTD.
Table 4. The parameters of TLTD.
PopulationCross ProbabilityMutation ProbabilitySelectionIterations ω d ω e
3000.50.3Tournament2001000150
Table 5. The results of TLTD-I. The data in the table is the average of the 1000 sample test results.
Table 5. The results of TLTD-I. The data in the table is the average of the 1000 sample test results.
CategorySuccess RateAverage of P i Average of D ( P i ) The Number of Modified Features
satan0.953−0.139102,729.9821.493
ipsweep0.9860.35292,384.8421.975
portsweep0.993−0.11782,101.0522.459
nmap0.140−0.0721337.4718.918
Table 6. The results of TLTD-II. The data in the table is the average of the 1000 sample test results.
Table 6. The results of TLTD-II. The data in the table is the average of the 1000 sample test results.
CategorySuccess RateAverage of P i Average of D ( P i ) The Number of Modified Features
satan0.185−0.1771479.3717.441
ipsweep0.8260.3093322.1320.431
portsweep0.564−0.1975341.4818.841
nmap0.190−0.148186.9516.807
Table 7. The results of TLTD-III. The data in the table is the average of the 1000 sample test results.
Table 7. The results of TLTD-III. The data in the table is the average of the 1000 sample test results.
CategorySuccess RateAverage of P i Average of D ( P i ) The Number of Modified Features
satan1−0.0621888.2319.765
ipsweep10.3791868.6319.943
portsweep0.949−0.1184622.6319.554
nmap10.0983010.9418.276

Share and Cite

MDPI and ACS Style

Liu, X.; Zhang, X.; Guizani, N.; Lu, J.; Zhu, Q.; Du, X. TLTD: A Testing Framework for Learning-Based IoT Traffic Detection Systems. Sensors 2018, 18, 2630. https://doi.org/10.3390/s18082630

AMA Style

Liu X, Zhang X, Guizani N, Lu J, Zhu Q, Du X. TLTD: A Testing Framework for Learning-Based IoT Traffic Detection Systems. Sensors. 2018; 18(8):2630. https://doi.org/10.3390/s18082630

Chicago/Turabian Style

Liu, Xiaolei, Xiaosong Zhang, Nadra Guizani, Jiazhong Lu, Qingxin Zhu, and Xiaojiang Du. 2018. "TLTD: A Testing Framework for Learning-Based IoT Traffic Detection Systems" Sensors 18, no. 8: 2630. https://doi.org/10.3390/s18082630

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop