Next Article in Journal
A Review of the Piezoelectric Electromechanical Impedance Based Structural Health Monitoring Technique for Engineering Structures
Next Article in Special Issue
A Personalized QoS Prediction Approach for CPS Service Recommendation Based on Reputation and Location-Aware Collaborative Filtering
Previous Article in Journal
Spoofing Detection Using GNSS/INS/Odometer Coupling for Vehicular Navigation
Previous Article in Special Issue
Clustering and Flow Conservation Monitoring Tool for Software Defined Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning Perfectly Secure Cryptography to Protect Communications with Adversarial Neural Cryptography

by
Murilo Coutinho
1,†,
Robson De Oliveira Albuquerque
1,†,
Fábio Borges
2,†,
Luis Javier García Villalba
3,*,† and
Tai-Hoon Kim
4,†
1
Cybersecurity INCT Unit 6, Decision Technologies Laboratory—LATITUDE, Electrical Engineering Department (ENE), Technology College, University of Brasília (UnB), 70.910-900 Brasília-DF, Brazil
2
National Laboratory for Scientific Computing; 25.651-075 Petrópolis-RJ, Brazil
3
Group of Analysis, Security and Systems (GASS), Department of Software Engineering and Artificial Intelligence (DISIA), Faculty of Computer Science and Engineering, Office 431, Universidad Complutense de Madrid (UCM), Calle Profesor José García Santesmases, 9, Ciudad Universitaria, 28040 Madrid, Spain
4
Department of Convergence Security, Sungshin Women’s University, 249-1 Dongseon-Dong 3-ga, Seoul 136-742, Korea
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2018, 18(5), 1306; https://doi.org/10.3390/s18051306
Submission received: 25 March 2018 / Revised: 13 April 2018 / Accepted: 18 April 2018 / Published: 24 April 2018
(This article belongs to the Special Issue Advances on Resources Management for Multi-Platform Infrastructures)

Abstract

:
Researches in Artificial Intelligence (AI) have achieved many important breakthroughs, especially in recent years. In some cases, AI learns alone from scratch and performs human tasks faster and better than humans. With the recent advances in AI, it is natural to wonder whether Artificial Neural Networks will be used to successfully create or break cryptographic algorithms. Bibliographic review shows the main approach to this problem have been addressed throughout complex Neural Networks, but without understanding or proving the security of the generated model. This paper presents an analysis of the security of cryptographic algorithms generated by a new technique called Adversarial Neural Cryptography (ANC). Using the proposed network, we show limitations and directions to improve the current approach of ANC. Training the proposed Artificial Neural Network with the improved model of ANC, we show that artificially intelligent agents can learn the unbreakable One-Time Pad (OTP) algorithm, without human knowledge, to communicate securely through an insecure communication channel. This paper shows in which conditions an AI agent can learn a secure encryption scheme. However, it also shows that, without a stronger adversary, it is more likely to obtain an insecure one.

1. Introduction

A significant improvement in Artificial Intelligence (AI) has been achieved. Nowadays, several systems overcame human capabilities in human-like tasks as image recognition [1,2], speech recognition [3], driving cars [4] and playing intuitive games [5]. A natural question to be asked is whether AI will be, someday, better than humans to design or break cryptography.
There are some papers in the literature trying to use machine learning techniques to design new cryptographic algorithms. Most of these works proposed encryption schemes designed with Neural Networks (NN) as a tool to create non-linearity [6,7,8,9]. It could be argued that these proposals are not Artificial Intelligence as there is not an agent learning based on security concepts. Also, in the hands of experienced crypto experts, theses algorithms turned out to be badly broken [10,11,12,13].
More recently [14], Abadi and Andersen suggested a different approach, namely Adversarial Neural Cryptography (ANC), in which three agents, Alice, Bob and Eve, compete in a dispute. Basically, Eve is a NN that tries to eavesdrop on Alice and Bob’s communication. Alice and Bob, which also are NN, try to learn how to protect their communication from Eve. Their idea is different since the agents are learning about security by themselves. However, a possible critic to their work is that they used complex Convolutional Neural Networks and did not show what cryptosystem their system had learned. Naturally, being secure against another neural network means nothing in terms of real security.
In this work, we analyze the algorithms generated by the ANC technique to understand its security. To do this, we designed a small NN capable of generalizing binary operations to a continuum space allowing the back-propagation algorithm to work properly. We designed this NN in a way that would be possible, but not necessary, to learn the One-Time Pad (OTP), which is well-known for being information-theoretically secure [15]. In other words, the OTP is unbreakable under some assumptions. Using this neural network, we show that the ANC model is not good enough to generate secure cryptosystems even using a simple NN.
To overcome these limitations, we propose an improvement to the ANC methodology using the concept of the Chosen-Plaintext Attack (CPA) [16], leading to what we called CPA-ANC. Since our NN is very simple, we were capable of reasoning when the learned model was in fact secure and why. With the proposed CPA-ANC methodology our NN learned the OTP, a secure cryptosystem.
The main contribution of this work is to demonstrate that an Artificial-Intelligent agent can learn a secure encryption algorithm without human knowledge. Previous work had similar contribution for games [5]. However, the conditions to achieve a secure algorithm are difficult to obtain and being more likely to get an insecure encryption algorithm if the agent is not in a well-crafted environment.
This paper is organized as follows: in Section 2 we presented related work. With more detail in Section 2.1 we review Neural Cryptography and in Section 2.2, we present the ANC methodology of [14].
In Section 3 we present our main contributions. In particular, in Section 3.1 we propose an improvement to the ANC methodology called CPA-ANC which uses the Chosen-Plaintext Attack to improve the security of the algorithms that the agents can learn. Also, in Section 3.2 we proposed a simple NN capable of learning the One-Time Pad.
In Section 4 we test the proposed methodology against the traditional ANC. Specifically, in Section 4.2 we trained the proposed NN without an adversary showing the Alice and Bob usually will communicate properly but without any form of encryption. In Section 4.3 we trained the proposed NN under the ANC methodology showing Alice and Bob can protect their communication from Eve but with an insecure encryption scheme. In Section 4.4, we train the proposed NN under the CPA-ANC methodology leading Alice and Bob to protect their communication from Eve generating a secure encryption scheme with very high probability, namely the OTP.
In Section 5, we show a concise comparison with related work and, finally, in Section 6 we present the conclusions, open questions and directions to future research.

2. Related Work

In this section, we review some works that used Neural Networks in Cryptography. More importantly, we explain how Adversarial Neural Cryptography works as proposed in [14].

2.1. Neural Cryptography

In 2002, Kanter et al. [6] proposed a new key exchange protocol between two parties using the notion of chaotic synchronization, which makes it possible for two weakly interacting chaotic systems to converge even though each one of them continues to move in a chaotic way. In Kanter’s protocol, each party has a NN that starts in a random state and at each round it updates itself and then reveals one bit of information about its state to the other party. They also show in [6] that an attacker who uses an identical neural network with the same learning procedure is extremely unlikely to synchronize his network with the other parties. In the same year, however, Shamir et al. [10] broke the system using three different techniques.
Afterwards, several papers used chaotic NN to propose encryption and hash algorithms [7,8,9]. All these papers used the chaotic NN as a tool for randomness only, then we do not consider these works as AI. Nevertheless, all these algorithms were also broken [11,12,13].
NNs were also used to develop pseudo-random number generators (PRNG) [17,18,19]. In these works the parameters of the NNs were considered the seed of the PRNG and the randomness were tested with tools like the NIST random number generator test suit.

2.2. Adversarial Neural Cryptography

A recent work defined the concept of ANC [14]. Abadi and Andersen proposed a system in which 2 NNs named Alice and Bob tried to exchange a message while limiting what a third NN, named Eve, could learn from eavesdropping the communication.
In their work, they did not prescribe specific cryptographic algorithms to these neural networks. In fact, they showed that Alice and Bob could learn how to perform forms of encryption and decryption successfully denying Eve any information about the message. However, in their work, they did not show any solution learned by the networks. In this section, we describe their system to give a general understanding of the mechanics of ANC.

2.2.1. System Organization

The system of ANC is designed with the classical cryptographic scenario involving three parties (Alice, Bob and Eve) showed in Figure 1. Alice and Bob share a secret key K and wish to communicate the plaintext P securely. Eve is a passive attacker that wishes to eavesdrop on their communications trying to gain information about P through the ciphertext C.
In the ANC model, Alice, Bob, and Eve are all neural networks. As is common for neural networks, Alice, Bob, and Eve work over tuples of floating-point numbers, rather than sequences of bits. In other words, K, P, P Bob , P Eve , and C are all tuples of floating-point numbers.

2.2.2. Methodology

As neural networks, Alice, Bob and Eve have parameters that will be denoted by θ A , θ B and θ E , respectively. Also, let’s define E A ( θ A , P , K ) as Alice’s output (maybe an encryption) on inputs P and K, D B ( θ B , C , K ) as Bob’s output (decryption) on inputs C and K, and D E ( θ E , C ) as Eve’s output on input C.
To measure the distance between the true plaintext and an estimated value, the L1 distance is used, the L1 distance is defined by:
d ( P , P ) = 1 N i = 1 N | p i p i | ,
where N is the length of plaintexts.
Eve’s goal is simple: to reconstruct P accurately. To achieve this goal, Eve’s loss function per-example is defined by
L E ( θ A , θ E , P , K ) = d ( P , D E ( θ E , E A ( θ A , P , K ) ) ) .
Intuitively, L E ( θ A , θ E , P , K ) represents how much Eve is wrong when the plaintext is P and the key is K. The loss function for Eve over the distribution of plaintexts and keys is defined by taking an expected value:
L E ( θ A , θ E ) = E P , K L E ( θ A , θ E , P , K )
Thus, one can find the “optimal Eve” by minimizing the loss:
O E ( θ A ) = argmin θ E ( L E ( θ A , θ E ) ) .
Similarly, one can define the per-example loss for Bob and extend it to the distribution on plaintexts and keys:
L B ( θ A , θ B , P , K ) = d ( P , D B ( θ B , E A ( θ A , P , K ) , K ) )
L B ( θ A , θ B ) = E P , K L B ( θ A , θ B , P , K ) .
Alice and Bob want to communicate clearly, but they also want to hide their communication from Eve. Thus, it is useful to define a loss function for Alice and Bob by combining L B and the optimal value of L E :
L A B ( θ A , θ B ) = L B ( θ A , θ B ) L E ( θ A , O E ( θ A ) ) .
Finally, one can obtain the “optimal Alice and Bob” by minimizing L A B ( θ A , θ B ) :
( O A , O B ) = argmin ( θ A , θ B ) ( L A B ( θ A , θ B ) ) .

2.2.3. Training

ANC is an unsupervised learning technique. Training begins with all NN (Alice, Bob and Eve) initialized randomly and without any data. For random examples of plaintexts P, Alice and Bob try to learn to communicate minimizing the error between the original plaintext P and the output from Bob’s network P Bob and, at the same time, making Eve guesses as random as possible. Afterwards, fixing Alice and Bob networks, Eve tries to decrypt C without knowing the key K. This process is repeated thousands of times until, hopefully, Alice and Bob find a way to communicate without giving Eve any information.
When training Alice and Bob, one should not attempt to maximize Eve’s reconstruction error. If Eve was completely wrong, then Eve could be completely right in the next iteration by simply flipping all output bits. A more realistic and useful goal for Alice and Bob is to make Eve produce answers indistinguishable from a random guess.
For more details on the ANC technique, please refer to the original paper [14].

3. Improvement to the ANC Methodology

In this section, we propose an improvement to the ANC methodology using the Chosen-Plaintext Attack (CPA) which we call CPA-ANC. Additionally, we present a simple NN capable of learning the One-Time Pad which will be used to test this new methodology against the traditional ANC.

3.1. Chosen-Plaintext Attack Adversarial Neural Cryptography

As we will see in the experiments of Section 4, the problem with the approach proposed in the original ANC work [14] is that Eve’s job is too hard. It must decrypt a random message having access only to the ciphertext. The consequence is that, under this methodology, Alice and Bob learn to communicate with cryptosystems that are not truly secure. Therefore, one can conclude that Alice and Bob do not have to do much effort to protect themselves against Eve, leading in insecure cryptosystems.
It is possible to improve ANC considering a more robust model of security for Alice, Bob and Eve. Namely, we will let Eve to mount a CPA. Therefore, to be protected against Eve, Alice and Bob will have to find a system secure against the CPA.
In this new setup Eve will choose two messages P 0 and P 1 and send them to Alice. Alice will choose one of these messages randomly, encrypt it with the NN obtaining the ciphertext C and send it to Eve and Bob. As usual, Bob will decrypt the message with a NN. However, Eve will not try to decrypt C, but will only output 0 if it believes P 0 was encrypted or 1 if it believes P 1 was encrypted. We call this the CPA-ANC setup and it is illustrated in Figure 2.
In this scenario of CPA-ANC, Alice and Bob will have to find a much better cryptosystem to communicate securely. In Section 4, we will show that this approach can really improve the quality of the solution and that a NN can, in fact, produce secure cryptosystems.

3.2. A Simple Neural Network Capable of Learning the One-Time Pad

In this section, we propose a NN complex enough to be able to learn some form of cryptography but simple enough to allow us to reason about its security. To do this, we used a continuous generalization for the operator XOR, which is a well-known binary and non-differentiable operation that happens to be used a lot in cryptography.
Thus, if we want a NN that can perform the XOR operation internally, we need a generalization of the operation. It is possible to generalize the XOR operation using the unit circle by mapping the bit 0 to the angle 0 and the bit 1 to the angle π . In this way, the XOR is equivalent to the sum of the angles.
Note, however, that the sum is a continuous operation. Thus, it is possible to work with angles different from 0 or π , generalizing bits to a continuous space. The following equation defines the mapping of a bit b into an angle:
f ( b ) = arccos ( 1 2 b ) .
The inverse of f provides the mapping of an angle a to a “continuous bit”:
f 1 ( a ) = 1 cos ( a ) 2 .
With this operator, we can introduce a small NN in Figure 3 for learning an OTP. We will refer to this NN as CryptoNet.
Basically, CryptoNet receives as input the plaintext and the key and, for each bit received, applies the transformation defined in Equation (9), resulting in angles. The next step is a standard matrix multiplication followed by the inverse transformation defined in Equation (10) resulting in the ciphertext. Note that the ciphertext is not composed of bits but by floating number between 0 and 1.
Mathematically, the fully connected layer of CryptoNet, is doing the following operation:
h 0 h 1 h n 1 = a 0 a n 1 a n a 2 n 1 W 2 n , n ,
where W 2 n , n is a matrix of weights with 2 n rows and n columns, a 0 , , a 2 n 1 are angles obtained from the plaintext and the key, and h 0 , , h n 1 are hidden variables.
Through the remainder of the paper, we will denote the CryptoNet mathematically as the function
C = ζ n ( W , P , K ) ,
where W is the matrix of weights defined above and P , K , C are vectors of n bits of input, key and output, respectively.
The CryptoNet can learn to combine the input in many ways. Since the input bits are mapped to the angles 0 or π , if all the weights (elements of W 2 n , n ) of the connections are integers, the result would be equivalent to XOR operations of the input bits.
Therefore, it is possible to learn an OTP with this network. It should be noted, however, that it is very unlikely to find an OTP randomizing the weights as real numbers since the probability of such event tends to 0.
With this model, we can train the network using ANC and reason about the results. Our goal is to understand if a NN can learn an OTP by itself. Moreover, we want to study and define conditions on the ANC model that can lead to a secure cryptosystem consistently.

4. Results

In this section, we test whether the proposed NN can learn a secure encryption system under three different scenarios. First, we train Alice and Bob to communicate without any adversary, showing that they will learn to communicate properly but without any encryption scheme. Then, we train Alice and Bob to communicate under the ANC setup showing that they will learn to communicate with a weak encryption scheme. Finally, we train Alice and Bob under the proposed CPA-ANC setup showing that they learn to communicate securely with high probability using the One-Time Pad.

4.1. Method

In all our tests, we trained two agents, Alice and Bob, using the same CryptoNet. The goal for Alice and Bob, in this case, was to learn to communicate using a single network. Thus, suppose Alice use an encryption function E ( B , K ) and Bob use a decryption function D ( B , K ) , we defined these functions as a single CryptoNet:
D ( B , K ) = E ( B , K ) = ζ n ( W , B , K )
Thus, Alice and Bob must work together to find a single CryptoNet such that its inverse is itself. Therefore, to communicate properly, it should be the case that
ζ n ( W , ζ n ( W , P , K ) , K ) = P ,
where P is the plaintext and K is the key.
In our tests, we initialized the weight matrix W randomly. The training, as in [14], relies on estimated values calculated over “minibatches” of hundreds or thousands of examples alternating the training of Eve with that of Alice and Bob. Since this process is intensive and time consuming for large keys, we trained the model with small plaintexts and keys: 4-bit ( n = 4 ), 8-bit ( n = 8 ) and 16-bit ( n = 4 ).
Unlike the standard training process of NN, here we do not have a clear concept of convergence. This is because when Eve changes its network, then Alice and Bob objective function also change as consequence. Therefore, we used two stopping criteria:
  • If the decryption error of Bob is very close to zero and Eve’s attack are as bad as random guesses, then we stop. In this case, we say we had convergence or a success.
  • If the first stop criterion is not reached in 100.000 rounds, we stop. Here a round is completed when Alice and Bob are trained and then Eve is trained. If this happens we say we did not have convergence or a failure.
We trained this model using Tensorflow [20], a machine learning framework in python. Also, we used Adam optimizer [21] with a learning rate of 0.0002 . We chose the learning rate and other hyperparameters adjusting the values empirically to avoid overfitting and underfitting as explained in the textbook [22]., which was defined empirically.
To obtain a meaningful solution in terms of XOR operations, we define Algorithm 1:
Algorithm 1: Testing a discrete CryptoNet
Sensors 18 01306 i001

4.2. Training without an Adversary

Our first approach was to train two agents, Alice and Bob, to communicate without an adversary. In other words, we want Alice to communicate with Bob without considering Eve. Note that, in this case, we are not using the ANC technique, just a straight forward communication.
In this case, being M random examples of plaintexts [ P ( 0 ) , P ( 1 ) , , P ( M 1 ) ] and keys [ K ( 0 ) , K ( 1 ) , , K ( M 1 ) ] , we trained CryptoNet by minimizing the loss function
L A B = 1 M i = 0 M 1 d P ( i ) , ζ n W , ζ n ( W , P ( i ) , K ( i ) ) , K ( i )
where the function d is the L1 distance defined by
d ( P , P ) = 1 N i = 1 N | p i p i | .
We tested several models using Algorithm 1, the results are presented in Table 1. All the trials were successful since Alice and Bob were able to communicate without errors. However, all the trials resulted in an insecure algorithm. As an example, the following algorithm was found in one of the trials:
ζ n ( W round , P , K ) = p 1 p 0 p 6 p 3 p 4 p 5 p 2 p 7 ,
where P = [ p 0 , p 1 , , p 7 ] and K = [ k 0 , k 1 , , k 7 ] . Clearly, this solution respects Equation (14).
Note that the key was not used in the learned algorithm of Equation (17). None of the trials resulted in a secure cryptosystem. Naturally, as Alice and Bob were trained to learn to communicate through a CryptoNet, without any safety concerns, there were no expectations that the resulting function would have been secure. In the next section, we will introduce CryptoNet in the context of ANC to understand if Eve can help Alice and Bob to define a secure system to communicate.

4.3. Training the network with Adversarial Neural Cryptography

In Section 4.2, we demonstrated that Alice and Bob could not find a secure network to communicate without any notion of what security means. One way to give Alice and Bob this kind of notion is to introduce an adversary, named Eve, who wants to gain knowledge about the communication over the channel. To do this we use the ANC model.
We will not define Eve explicitly here because it can be any sufficiently complex NN (for example, the one proposed in [14]), as we will show in the results. Alice and Bob will communicate through a single CryptoNet, see Equation (13). Eve will try to learn a NN in which receives as input the ciphertext C and outputs the plaintext P:
P = D E ( θ , C ) ,
where θ are the learned parameters of the NN.
It follows that given M examples of plaintexts [ P ( 0 ) , P ( 1 ) , , P ( M 1 ) ] and ciphertexts [ C ( 0 ) , C ( 1 ) , , C ( M 1 ) ] , Eve can learn by minimizing the loss
L E = 1 M i = 0 M 1 d P ( i ) , D E θ , C ( i ) ,
where d is given in Equation (16).
Note that in the ANC method, Eve uses the plaintext as a measure of performance when training the NN. However, Eve’s NN does not receive the plaintext as input. Therefore, Eve is applying a ciphertext-only attack.
Alice and Bob want to minimize Eve’s knowledge about the plaintext P. In practice, Alice and Bob will learn by minimizing the loss
L = L A B γ min ( L E , 0.5 )
where L A B is given in Equation (15), L E is given in Equation (19) and γ is a hyperparameter. The minimum function is used to guarantee that Alice and Bob would not try to maximize Eve’s error, since, in this case, Eve could just flip all bits in the next round, achieving a correct guess. If Eve is just guessing the plaintext, it would be expected to get about half of the bits right, resulting in an average error of 0.5 .
As in Section 4.2, we train the networks with M random examples of plaintexts [ P ( 0 ) , P ( 1 ) , , P ( M 1 ) ] and keys [ K ( 0 ) , K ( 1 ) , , K ( M 1 ) ] . In our tests, we used a “minibatch” of M = 128 entries. We used γ = 5 in Equation (20) (intuitively this means that is more important to keep Eve’s error high than to reduce Bob’s decryption error). Also, we used the L2 regularization with α = 0.1 (4-bits), α = 0.05 (8-bits) and α = 0.01 (16-bits) (see [22]). All these hyperparameters were determined empirically.
Training alternates between five minibatches for Alice and Bob and 10 minibatches for Eve. The purpose of this ratio is to give a computational edge to Eve without training it so much that it becomes excessively specific to the exact current parameters of Alice and Bob. One example of this process can be seen in Figure 4.
After the training was completed, we used Algorithm 1 to test the learned CryptoNet. Unlike in [14], we did not trained Eve again when the training of Alice and Bob was finished. This is because the CryptoNet obtained is simple enough, so we can easily reason about its security without the aid of a NN.
Table 2 shows a summary of the results. For each trial, the learned model was considered successful if Alice and Bob could communicate without errors when executing Algorithm 1. A learned CryptoNet was considered secure if it learned the OTP. For example, the following CryptoNet was learned by Alice and Bob:
ζ n ( W round , P , K ) = p 0 k 3 p 3 k 0 p 2 k 0 k 3 k 4 k 5 k 6 k 7 p 1 k 0 p 4 k 1 p 5 k 2 p 6 k 2 p 7 k 1 .
This CryptoNet was considered a success since Alice and Bob can communicate without errors. Naturally, this encryption in not secure since several plaintext bits were encrypted with the same key bit. However, this CryptoNet is secure against Eve. Obviously, any adversary with Eve’s knowledge (any human or NN) cannot recover the plaintext since the plaintext itself was generated randomly. For example, suppose that c 1 = c 3 = 0 , then Eve could learn that c 1 c 3 = p 1 p 3 and infer that p 1 = p 3 . However, since the plaintext is random, the best Eve could do is to guess p 1 and p 3 with the same value but, with this strategy, Eve would get both bits right in 50 % of the time and get both bits wrong 50 % of the time. Thus, Eve’s output is not better than random, on average.
This demonstrates that the standard ANC technique proposed in [14] is not good enough because it usually does not lead to a secure cryptosystem even using a very simple NN designed to allow a perfectly secure solution. In the next section, we will use the proposed CPA-ANC giving Eve a lot more power forcing Alice and Bob to a better cryptosystem.

4.4. Learning the One-Time Pad

To test the proposed CPA-ANC setup, as in previous sections, we considered Alice and Bob sharing a single CryptoNet (presented in Figure 3) to encrypt and decrypt. On the other hand, Eve’s network changed a lot. Now Eve is a classifier, receiving P 0 , P 1 and C as input and classifying C as 0 if it believes C was derived from P 0 and 1 if it believes C was derived from P 1 .
We designed Eve with a similar structure as the CryptoNet. It receives two plaintexts P 0 and P 1 , the ciphertext C, and transforms the bits into angles through Equation (9). A fully connected layer, combines the angles into a total of R rules. Each rule is composed by generalized XOR operations and are transformed back into continuous bits through Equation (10). A second fully connected layer combines the results into logits going through a softmax layer resulting in a probability π 0 of C being a ciphertext of P 0 and a probability π 1 of C being a ciphertext of P 1 . Finally, the network outputs 0 if π 0 > π 1 or outputs 1 otherwise. We call this network CPA-CryptoNet and it is represented in Figure 5.
We must change the optimization problem to adapt to this new CPA scenario. To do this, given M examples of plaintexts [ P 0 ( 0 ) , P 0 ( 1 ) , , P 0 ( M 1 ) ] , [ P 1 ( 0 ) , P 1 ( 1 ) , , P 1 ( M 1 ) ] and ciphertexts [ C ( 0 ) , C ( 1 ) , , C ( M 1 ) ] , we define the loss for Eve as the cross-entropy:
L E = 1 M i = 0 M 1 j = 0 1 t j ( i ) log π j ( i )
where t j ( i ) = 1 if C ( i ) is the ciphertext of P j ( i ) , and t j ( i ) = 0 otherwise. Thus, Eve learns by minimizing L E while Alice and Bob learn by minimizing L given by
L = L A B γ min ( E r r , 0.5 )
where L A B is given in Equation (15), E r r is Eve’s classification error and γ is a hyperparameter.
To train the networks, we used a “minibatch” of M = 128 entries. We used γ = 7 in Equation (23). Also, we used the L2 regularization (see [22]) with α = 0.1 for 4-bit ( n = 4 ) and 8-bit ( n = 8 ) key lengths and α = 0.015 for 16-bit ( n = 16 ) key length. Also, for Eve’s network, we defined the number of rules R (see Figure 5) as R = 4 × n . The number of rules defines the number of linear combinations that Eve can analyze together to attack. Eve should need more equations as the key size grows. Therefore, we set it proportionally to the number of bits of the key. We grew the number of linear combinations until reaching a value in which Eve had great power to break the kind of cryptograms learned by Alice and Bob through the Cryptonet reaching the value of 4 × n rules.We chose all these parameters empirically.
Training alternates between the NN of Alice and Bob and the NN of Eve, with Alice and Bob training for 3 “minibatches”, and then Eve training for 60 “minibatches”. The purpose of this ratio is to give a computational advantage to Eve. One example of this process can be seen in Figure 6.
Table 3 shows a summary of the results. For each trial, the learned model was considered successful if Alice and Bob could communicate without errors when executing Algorithm 1. The model was considered a failure otherwise. A learned CryptoNet was considered secure against Eve, if Eve could not extract any information from the ciphertext. For example, the following crypto systems were learned by Alice and Bob:
ζ n ( W round , P , K ) = p 0 k 5 p 1 k 7 p 2 k 1 p 3 k 0 p 4 k 2 p 5 k 6 p 6 k 3 p 7 k 4 and ζ n ( W round , P , K ) = p 0 k 3 p 1 k 10 p 2 k 14 p 3 k 9 p 4 k 5 p 5 k 7 p 6 k 1 p 7 k 8 p 8 k 6 p 9 k 2 p 10 k 12 p 11 k 4 p 12 k 13 p 13 k 15 p 14 k 0 p 15 k 11 .
Note that the learned cryptosystems of Equations (24) are secure, namely, the OTP. In our tests, just one model was a failure as Alice and Bob could not communicate. Almost all successful models trained under the proposed CPA-ANC methodology learned an OTP cryptosystem. Moreover, comparing the results of Table 2 and Table 3, one can note an increased number of successful trials. A plausible reason is that the original ANC methodology leads to a weaker adversary, and because of that, Alice and Bob have too much degrees of freedom with many possible solutions. This may lead to a more complex objective function with many local solutions. With CPA-ANC methodology and with the neural networks used, the only hope for Alice and Bob is to find an OTP solution. Therefore, it is likely that the objective function has a better behavior than the one obtained with the original ANC model.

5. Comparison with Related Work

As aforementioned, several NN cryptosystems were broken. This work reconstructs the technique ANC showing the cryptosystems generated by it. In addition, some of these cryptosystems are indeed Vigenère like ciphers, which can be broken. Furthermore, we show that the learning process from previous work probably will not generate a secure cryptosystem. Moreover, we improved the learning process by means of a stronger adversary, i.e., CPA-ANC and the same NN became able to learn an unbreakable cryptosystem with high probability, namely the OTP. Certainly, a cryptosystem is secure under the OTP assumptions, namely the one-time keys should be truly random and never leaked. The CryptoNet learned how to use the key. Table 4 summarizes the comparison.
Normally, the cryptosystem result from ANC is a type of Vigenère cipher and from CPA-ANC is OTP. We do not need to compare their complexity like in [23], because both cryptosystems require only XOR and have equivalent performance.
Research in the intersection between Security and Artificial Intelligence has a lot of challenges. To clarify our contribution, this work presents the first Artificial Neural Network able to learn an unbreakable cryptographic technique, namely OTP. Even with the advent of quantum computers or any other computing technology [24], we cannot break the OTP. Machines just learned how to use it.

6. Conclusions

In this paper, we have shown that a neural network can learn a perfectly secure cryptosystem in the right circumstances. In addition, we have shown that the original ANC methodology is not good enough to achieve this goal. Moreover, we have presented a new CPA-ANC methodology capable of improving the objective function and the learned model.
In our experiments, we used simple neural networks to better understand the learned model. Using the original ANC methodology, very few of the learned models were truly secure. However, using the proposed CPA-ANC almost all the learned models were an OTP, which is secure. The main message here is that the adversary must be very powerful to force the solution into a strong cryptosystem. In other words, the proposed CPA-ANC methodology alone is not enough to guarantee security, the design of a very strong adversarial capable of breaking cryptosystems is key. In our minimalistic model, this was possible to achieve, however, in general, this is a hard open problem.
For further work, the research will continue to conduct more tests to evaluate more parameters than the current tests. It is necessary to implement a parallel code to increase performance and test the proposed model with larger keys. Additionally, it remains an open problem whether a neural network can learn a secure cryptosystem in which a small key is used to encrypt a very long message, like a block or stream cipher would do.
Of course, we do not recommend using neural networks in real systems. This paper shows that is very likely to get a weak encryption scheme using neural networks. There is a long path ahead to understand in which conditions a complex neural network will learn a secure cryptosystem consistently. Eventually, neural networks might break current cryptosystems and create others more secure. It is a long journey to transform this theoretical result to a practical one.

Author Contributions

M.C., R.d.O.A. and F.B. conceived the general concept and design of the Adversarial Neural Cryptography model. M.C. implemented and trained the neural networks. R.d.O.A. and F.C. reviewed and validated the experiments. M.C., R.d.O.A., F.B., L.J.G.V. and T.H.K. analyzed data resulting from the experiments and the results. All authors contributed equally in reviewing and approving the manuscript.

Acknowledgments

This research work was supported by Sungshin Women’s University. In addition, L.J.G.V. thanks the European Commission Horizon 2020 5G-PPP Programme (Grant Agreement number H2020-ICT-2014-2/671672-SELFNET-Framework for Self-Organized Network Management in Virtualized and Software-Defined Networks). R.d.O.A. thanks the Brazilian research and innovation Agencies CAPES-Coordination for the Improvement of Higher Education Personnel (Grant 23038.007604/2014-69 FORTE-Tempestive Forensics Project), CNPq-National Council for Scientific and Technological Development (Grant 465741/2014-2 Science and Technology National Institute-INCT on Cybersecurity), FAPDF-Research Support Foundation of the Federal District (Grant 0193.001366/2016-UIoT-Universal Internet of Things and Grant 0193.001365/2016-SSDDC-Secure Software Defined Data Center), and the Institutional Security Office of the Presidency of the Republic of Brazil (Grant 002/2017).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
ANCAdversarial Neural Cryptography
OTPOne-Time Pad
CPAChosen-Plaintext Attack
NNNeural Networks
PRNGpseudo-random number generators

References

  1. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv, 2014; arXiv:1409.1556. [Google Scholar]
  2. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the 25th International Conference on Neural Information Processing Systems, Lake Tahoe, NV, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  3. Graves, A.; Mohamed, A.R.; Hinton, G. Speech recognition with deep recurrent neural networks. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Vancouver, BC, Canada, 26–31 May 2013; pp. 6645–6649. [Google Scholar]
  4. Bojarski, M.; Del Testa, D.; Dworakowski, D.; Firner, B.; Flepp, B.; Goyal, P.; Jackel, L.D.; Monfort, M.; Muller, U.; Zhang, J.; et al. End to end learning for self-driving cars. arXiv, 2016; arXiv:1604.07316. [Google Scholar]
  5. Silver, D.; Schrittwieser, J.; Simonyan, K.; Antonoglou, I.; Huang, A.; Guez, A.; Hubert, T.; Baker, L.; Lai, M.; Bolton, A.; et al. Mastering the game of Go without human knowledge. Nature 2017, 550, 354. [Google Scholar] [CrossRef] [PubMed]
  6. Kanter, I.; Kinzel, W.; Kanter, E. Secure exchange of information by synchronization of neural networks. arXiv, 2002; arXiv:cond-mat/0202112. [Google Scholar]
  7. Lian, S.; Chen, G.; Cheung, A.; Wang, Z. A chaotic-neural-network-based encryption algorithm for JPEG2000 encoded images. Lect. Notes Comput. Sci. 2004, 3174, 627–632. [Google Scholar]
  8. Yu, W.; Cao, J. Cryptography based on delayed chaotic neural networks. Phys. Lett. A 2006, 356, 333–338. [Google Scholar] [CrossRef]
  9. Wang, X.Y.; Yang, L.; Liu, R.; Kadir, A. A chaotic image encryption algorithm based on perceptron model. Nonlinear Dyn. 2010, 62, 615–621. [Google Scholar] [CrossRef]
  10. Klimov, A.; Mityagin, A.; Shamir, A. Analysis of neural cryptography. In Advances in Cryptology— ASIACRYPT 2002, Proceedings of the International Conference on the Theory and Application of Cryptology and Information Security, Queenstown, New Zealand, 1–5 December 2002; Springer: Berlin/Heidelberg, Germany, 2002; pp. 288–298. [Google Scholar]
  11. Li, C. Cryptanalyses of Some Multimedia Encryption Schemes. Master’s Thesis, Zhejiang University, Hangzhou, China, May 2005. [Google Scholar]
  12. Qin, K.; Oommen, B.J. On the cryptanalysis of two cryptographic algorithms that utilize chaotic neural networks. Math. Probl. Eng. 2015, 2015, 9. [Google Scholar] [CrossRef]
  13. Zhang, Y.; Li, C.; Li, Q.; Zhang, D.; Shu, S. Breaking a chaotic image encryption algorithm based on perceptron model. Nonlinear Dyn. 2012, 69, 1091–1096. [Google Scholar] [CrossRef]
  14. Abadi, M.; Andersen, D.G. Learning to protect communications with adversarial neural cryptography. arXiv, 2016; arXiv:1610.06918. [Google Scholar]
  15. Shannon, C.E. Communication theory of secrecy systems. Bell Labs Techn. J. 1949, 28, 656–715. [Google Scholar] [CrossRef]
  16. Katz, J.; Lindell, Y. Introduction to Modern Cryptography; Chapman & Hall: London, UK, 2007. [Google Scholar]
  17. Desai, V.; Deshmukh, V.; Rao, D. Pseudo random number generator using Elman neural network. In Proceedings of the IEEE Recent Advances in Intelligent Computational Systems (RAICS), Trivandrum, India, 22–24 September 2011; pp. 251–254. [Google Scholar]
  18. Desai, V.; Patil, R.; Rao, D. Using layer recurrent neural network to generate pseudo random number sequences. Int. J. Comput. Sci. Issues 2012, 9, 324–334. [Google Scholar]
  19. Yayık, A.; Kutlu, Y. Improving Pseudo random number generator using artificial neural networks. In Proceedings of the IEEE 21st Signal Processing and Communications Applications Conference (SIU), Haspolat, Turkey, 24–26 April 2013; pp. 1–4. [Google Scholar]
  20. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M.; et al. TensorFlow: A system for large-scale machine learning. In Proceedings of the 12th USENIX Symposium on Operating Systems Design and Implementation (OSDI), Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  21. Kingma, D.; Ba, J. Adam: A method for stochastic optimization. arXiv preprint, 2014; arXiv:1412.6980. [Google Scholar]
  22. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016; Available online: http://www.deeplearningbook.org (accessed on 19 April 2018).
  23. Borges de Oliveira, F. Analytical comparison. In On Privacy-Preserving Protocols for Smart Metering Systems: Security and Privacy in Smart Grids; Springer International Publishing: Cham, Switzerland, 2017; pp. 101–110. [Google Scholar]
  24. Oliveira, L.B.; Pereira, F.M.Q.; Misoczki, R.; Aranha, D.F.; Borges, F.; Liu, J. The computer for the 21st century: Security privacy challenges after 25 years. In Proceedings of the 26th International Conference on Computer Communication and Networks (ICCCN), Vancouver, BC, Canada, 31 July–3 August 2017; pp. 1–10. [Google Scholar]
Figure 1. Alice, Bob, and Eve, with a symmetric cryptosystem.
Figure 1. Alice, Bob, and Eve, with a symmetric cryptosystem.
Sensors 18 01306 g001
Figure 2. Alice, Bob, and Eve, and the CPA-ANC setup. Alice and Bob share a secret key K. Eve chooses two messages P 0 and P 1 . Alice randomly chooses one message to encrypt producing the ciphertext C. Bob uses the key K to decrypt C producing P Bob . Eve receives the ciphertext C and tries to guess which message was encrypted outputting 0 if believes P 0 was encrypted and 1 if believes P 1 was encrypted.
Figure 2. Alice, Bob, and Eve, and the CPA-ANC setup. Alice and Bob share a secret key K. Eve chooses two messages P 0 and P 1 . Alice randomly chooses one message to encrypt producing the ciphertext C. Bob uses the key K to decrypt C producing P Bob . Eve receives the ciphertext C and tries to guess which message was encrypted outputting 0 if believes P 0 was encrypted and 1 if believes P 1 was encrypted.
Sensors 18 01306 g002
Figure 3. The proposed neural network (CryptoNet). The bits of the plaintext are represented by [ p 0 , p 1 , …, p n 1 ]. The bits of the key are represented by [ k 0 , k 1 , …, k n 1 ]. The function f (see Equation (9)) transforms the bits into angles [ a 0 , a 1 , … a 2 n 1 ]. A fully connected layer combines the angles forming the variables [ h 0 , h 1 , …, h n 1 ]. The function f 1 (see Equation (10)) transforms the combined angles into continuous bits (real numbers in the interval [ 0 , 1 ] ) [ c 0 , c 1 , …, c n 1 ] representing the ciphertext.
Figure 3. The proposed neural network (CryptoNet). The bits of the plaintext are represented by [ p 0 , p 1 , …, p n 1 ]. The bits of the key are represented by [ k 0 , k 1 , …, k n 1 ]. The function f (see Equation (9)) transforms the bits into angles [ a 0 , a 1 , … a 2 n 1 ]. A fully connected layer combines the angles forming the variables [ h 0 , h 1 , …, h n 1 ]. The function f 1 (see Equation (10)) transforms the combined angles into continuous bits (real numbers in the interval [ 0 , 1 ] ) [ c 0 , c 1 , …, c n 1 ] representing the ciphertext.
Sensors 18 01306 g003
Figure 4. The learning curves forThe challenge between Alice and Bob against Eve who tries to minimize its decryption error. Alice and Bob try to minimize Bob’s decryption error while maximizing Eve’s decryption error. Eve is represented in green and Alice and Bob are represented in blue. The number of steps denote the number of “minibatches” on training phase.
Figure 4. The learning curves forThe challenge between Alice and Bob against Eve who tries to minimize its decryption error. Alice and Bob try to minimize Bob’s decryption error while maximizing Eve’s decryption error. Eve is represented in green and Alice and Bob are represented in blue. The number of steps denote the number of “minibatches” on training phase.
Sensors 18 01306 g004
Figure 5. Eve’s neural network (CPA-CryptoNet). Eve receives as input two plaintexts P 0 and P 1 . The bits of each plaintext are represented by [ p 0 , 0 , …, p 0 , n 1 ] for P 0 and [ p 1 , 0 , …, p 1 , n 1 ] for P 1 . Eve also receives the ciphertext C represented by [ c 0 , …, c n 1 ]. The function f (see Equation (9)) transforms the bits into angles [ a 0 , a 1 , … a 3 n 1 ]. A fully connected layer combines the angles generating the hidden variables [ h 0 0 , h 1 0 , …, h R 1 0 ], where R is the number of rules. The function f 1 (see Equation (10)) transforms the combined angles into continuous bits (real numbers in the interval [ 0 , 1 ] ) [ h 0 1 , h 1 1 , …, h R 1 1 ]. Another fully connected layer brings the hidden variables to logits through a softmax layer resulting in a probability π 0 of C being a ciphertext of P 0 and a probability π 1 of C being a ciphertext of P 1 .
Figure 5. Eve’s neural network (CPA-CryptoNet). Eve receives as input two plaintexts P 0 and P 1 . The bits of each plaintext are represented by [ p 0 , 0 , …, p 0 , n 1 ] for P 0 and [ p 1 , 0 , …, p 1 , n 1 ] for P 1 . Eve also receives the ciphertext C represented by [ c 0 , …, c n 1 ]. The function f (see Equation (9)) transforms the bits into angles [ a 0 , a 1 , … a 3 n 1 ]. A fully connected layer combines the angles generating the hidden variables [ h 0 0 , h 1 0 , …, h R 1 0 ], where R is the number of rules. The function f 1 (see Equation (10)) transforms the combined angles into continuous bits (real numbers in the interval [ 0 , 1 ] ) [ h 0 1 , h 1 1 , …, h R 1 1 ]. Another fully connected layer brings the hidden variables to logits through a softmax layer resulting in a probability π 0 of C being a ciphertext of P 0 and a probability π 1 of C being a ciphertext of P 1 .
Sensors 18 01306 g005
Figure 6. The challenge of Alice and Bob against Eve. Eve tries to maximize its classification rate while Alice and Bob try to minimize Eve’s classification rate and minimize Bob’s decryption error. In the figure on the left, Bob’s decryption error decreases over time in blue color. Also, on the right, one can see that Eve increases its classification rate in red color, however, when Alice and Bob learn a secure cryptosystem, in this case the OTP, Eve’s classification rate becomes no better than random. The number of steps denote the number of “minibatches” on training phase. In black, we have a smooth version of the red curve.
Figure 6. The challenge of Alice and Bob against Eve. Eve tries to maximize its classification rate while Alice and Bob try to minimize Eve’s classification rate and minimize Bob’s decryption error. In the figure on the left, Bob’s decryption error decreases over time in blue color. Also, on the right, one can see that Eve increases its classification rate in red color, however, when Alice and Bob learn a secure cryptosystem, in this case the OTP, Eve’s classification rate becomes no better than random. The number of steps denote the number of “minibatches” on training phase. In black, we have a smooth version of the red curve.
Sensors 18 01306 g006
Table 1. 10 networks were learned and tested using Algorithm 1 for each key size. All the trials were successful since Alice and Bob were able to communicate without errors. However, all the trials resulted in an insecure algorithm.
Table 1. 10 networks were learned and tested using Algorithm 1 for each key size. All the trials were successful since Alice and Bob were able to communicate without errors. However, all the trials resulted in an insecure algorithm.
Size of KeyNumber of TrialsSuccessful CommunicationsSecure Algorithm Learned (OTP)
4-bit10100
8-bit10100
16-bit10100
Table 2. 20 networks were learned with ANC and tested using Algorithm 1 for each key size. Not all trials were successful since Alice and Bob were not able to communicate without errors in some cases. The network learned a secure algorithm (OTP) in some rare cases.
Table 2. 20 networks were learned with ANC and tested using Algorithm 1 for each key size. Not all trials were successful since Alice and Bob were not able to communicate without errors in some cases. The network learned a secure algorithm (OTP) in some rare cases.
Size of KeyNumber of TrialsSuccessful CommunicationsSecure Algorithm Learned (OTP)
4-bit20202
8-bit20182
16-bit20110
Table 3. 20 networks were learned CPA-ANC and tested using Algorithm 1 for each key size. Most of the trials were successful since Alice and Bob were able to communicate without errors. Most of the successful networks learned the OTP, which is a secure cryptosystem.
Table 3. 20 networks were learned CPA-ANC and tested using Algorithm 1 for each key size. Most of the trials were successful since Alice and Bob were able to communicate without errors. Most of the successful networks learned the OTP, which is a secure cryptosystem.
Size of KeyNumber of TrialsSuccessful CommunicationsSecure Algorithm Learned (OTP)
4-bit201919
8-bit202020
16-bit202019
Table 4. Comparison with related work.
Table 4. Comparison with related work.
WorkTechniqueConclusion
[6]chaotic synchronization aaBroken in [10]
[7]chaotic NNBroken in [11]
[8]chaotic NNBroken in [12]
[9]chaotic NNBroken in [13]
[14]ANCconverge to OTP with low probability.
This work aaCPA-ANCconverge to OTP with high probability.

Share and Cite

MDPI and ACS Style

Coutinho, M.; De Oliveira Albuquerque, R.; Borges, F.; García Villalba, L.J.; Kim, T.-H. Learning Perfectly Secure Cryptography to Protect Communications with Adversarial Neural Cryptography. Sensors 2018, 18, 1306. https://doi.org/10.3390/s18051306

AMA Style

Coutinho M, De Oliveira Albuquerque R, Borges F, García Villalba LJ, Kim T-H. Learning Perfectly Secure Cryptography to Protect Communications with Adversarial Neural Cryptography. Sensors. 2018; 18(5):1306. https://doi.org/10.3390/s18051306

Chicago/Turabian Style

Coutinho, Murilo, Robson De Oliveira Albuquerque, Fábio Borges, Luis Javier García Villalba, and Tai-Hoon Kim. 2018. "Learning Perfectly Secure Cryptography to Protect Communications with Adversarial Neural Cryptography" Sensors 18, no. 5: 1306. https://doi.org/10.3390/s18051306

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop