Next Article in Journal
REEGAT: RoBERTa Entity Embedding and Graph Attention Networks Enhanced Sentence Representation for Relation Extraction
Previous Article in Journal
A Hybrid GA/ML-Based End-to-End Automated Methodology for Design Acceleration of Wireless Communications CMOS LNAs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Noisy Quantum Channel Characterization Using Quantum Neural Networks

1
School of Artificial Intelligence, Beijing Normal University, Beijing 100875, China
2
Applied Optics Beijing Area Major Laboratory, Beijing Normal University, Beijing 100875, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(11), 2430; https://doi.org/10.3390/electronics12112430
Submission received: 27 January 2023 / Revised: 5 May 2023 / Accepted: 24 May 2023 / Published: 27 May 2023
(This article belongs to the Topic Quantum Information and Quantum Computing)

Abstract

:
Channel noise is considered to be the main obstacle in long-distance quantum communication and distributed quantum networks. Here, employing a quantum neural network, we present an efficient method to study the model and detect the noise of quantum channels. Based on various types of noisy quantum channel models, we construct the architecture of the quantum neural network and the model training process. Finally, we perform experiments to verify the training effectiveness of the scheme, and the results show that the cost function of the quantum neural network could approach above 90% of the channel model.

Graphical Abstract

1. Introduction

Quantum information processing is the transmission or processing of information through a quantum system [1,2], such as quantum communications [3,4,5,6,7], quantum computation [8,9,10] and quantum metrology [11,12]. During recent decades, there have been extensive studies into quantum information processing. Specifically, quantum communications (QC), as the most promising approach for secure information transmission, is becoming increasingly applicable. In the realization of QC, the quantum channels are key ingredients for quantum information and signal transmission. However, there are two main obstacles in the channels for the implementation of long-distance communications and network construction: channel loss and channel noise [13]. There are many related works on eliminating channel noise, such as channel discrimination [14], error-rejection coding [15] and quantum error corrections [16,17,18]. As a priority in the research of noisy quantum communications, noisy quantum channel detection is perceived as a highly versatile task.
Meanwhile, classical and quantum neural networks have become pervasive tools in quantum information processing [19,20]. During recent decades, intensive research work related to quantum neural networks has made great progress, and the applications of quantum neural networks in various quantum information tasks have also gradually increased with their increased performances [21,22,23,24]. For example, in 2010, Bisio et al. focused on storing the unitary operation in a quantum memory with limited resources rather than improving their performance and generalization [25]. In ref. [26], Verdon et al. introduced quantum machine learning of parametrized unitary operations and proved that it is possible to construct a functionally simpler linear quantum perceptron to fit the unitary mapping. Based on this work, Beer et al. used the unitary operator as a quantum perceptron and proposed a corresponding efficient training algorithm [27]. The quantum perceptrons form a quantum neural network through connections, which can achieve forward propagation and back propagation and complete the task of learning arbitrary unknown unitary operators.
In this paper, employing a quantum neural network, we study various noisy channel models of quantum communications to enable a high accuracy information transmission. Meanwhile, the errors caused by channel noise during the information transmission process can be reduced. Numerical experiments are performed to evaluate the performance of the model. By analyzing the experimental results, we conclude that the quantum neural network is effective for both two-dimensional and three-dimensional quantum channels, which can be further generalized to higher-dimensional cases. Among the studies related to improving the performance of the quantum neural network, we find that a simpler structured network works better for simple tasks. Furthermore, the network hyperparameter λ , which is the reciprocal of the learning rate, is selected according to the problem at hand. The random initialization of the network impacts the training performance, which can be reduced by increasing the number of training rounds.
Our paper is organized as follows: in Section 2, we briefly introduce a model of a noisy quantum channel and a quantum neural network. In Section 3, we feature the details of our scheme including task definition, data preparation, network architecture and the training process. In Section 4, we present our experimental results and analyses. In Section 5, we propose some methods to improve the performance of the network. In Section 6, we test the robustness of the network. Finally, we summarize the work in Section 7.

2. Preliminaries

2.1. Noisy Quantum Channel

Theoretically, the channel noise will cause signal distortion, which affects the efficiency and fidelity of qubits during transmission. Here, in this part, we first present the basic models of noisy quantum channels.

2.1.1. Channel Models under Two-Dimensional Hilbert Space

Traditionally, four basic noisy quantum channels are commonly used in the qubit and are expressed by a two-dimensional Hilbert space. These are called the depolarizing channel, amplitude damping channel, phase damping channel and bit flip channel [28].
  • Depolarizing channel.
    The qubits are transmitted through the depolarizing channel, with the probability of the states being unchanged as 1 p , and the probability, p, that they become completely mixed state as I / 2 . The channel operator can be described by a completely positive trace-preserving map as
    ε ( ρ ) = p I 2 + ( 1 p ) ρ .
    where ρ represents the input state and the mixed state I / 2 can be expressed as
    I 2 = ρ + σ 1 ρ σ 1 + σ 2 ρ σ 2 + σ 3 ρ σ 3 4 .
    By substituting I / 2 into the above formula, we get
    ε ( ρ ) = ( 1 3 p 4 ) ρ + p 4 ( σ 1 ρ σ 1 + σ 2 ρ σ 2 + σ 3 ρ σ 3 ) .
    From the above equation, it is obvious that there is a 1 3 p / 4 probability of maintaining the original state, and a p / 4 probability that the input state will undergo the relative σ 1 , σ 2 and σ 3 operations. Here, σ 1 , σ 2 and σ 3 are the Pauli matrices, expressed as
    σ 1 = σ x = 0 1 1 0 , σ 2 = σ y = 0 i i 0 , σ 3 = σ z = 1 0 0 1 .
    The Kraus operators of the depolarizing channel can be obtained by
    E 0 = 1 3 p 4 I , E 1 = p 4 σ x , E 2 = p 4 σ y , E 3 = p 4 σ z .
  • Amplitude damping channel.
    The amplitude damping channel describes the energy dissipation of the qubits system, which can be described as
    ε ( ρ ) = E 0 ρ E 0 + E 1 ρ E 1 .
    Here, the normalization condition satisfies the relation
    E 0 E 0 + E 1 E 1 = 1
    where the Kraus operations E 0 and E 1 can be expressed as
    E 0 = 1 0 0 1 γ , E 1 = 0 γ 0 0 .
  • Phase damping channel.
    The phase damping channel describes the loss of quantum information without the loss of energy; it has the same expression as ε ( ρ ) in the amplitude damping channel, but with different Kraus operations
    E 0 = 1 0 0 1 γ , E 1 = 0 0 0 γ .
    The E 0 operation leaves | 0 unchanged but reduces the amplitude of the | 1 state. Furthermore, the E 1 operation destroys | 0 and reduces the amplitude of the | 1 state.
  • Bit-flip channel.
    The bit-flip channel keeps the original input state unchanged with the probability p, and flips the state of each qubit from | 0 ( | 1 ) to | 1 ( | 0 ) with the probability 1 p , which can be described as
    ε ( ρ ) = p ρ + ( 1 p ) X ρ .
    Furthermore, it has Kraus operations that can be defined as
    E 0 = p I = p 1 0 0 1 , E 1 = 1 p X = 1 p 0 1 1 0 .

2.1.2. Channel Models under Three-Dimensional Hilbert Space

To increase the channel capacity, a higher dimensional coding space can also be considered for qutrit transmission. The noisy quantum channel that is commonly used in the qutrit system can be expressed in the general form as
U m n = j = 0 2 e 2 π i j n / d | ( j + m ) % d j | .
Here, the parameters m , n = 0 , 1 , 2 . Furthermore, the corresponding operators in the above equation can be described as
U 00 = 1 0 0 0 1 0 0 0 1 , U 01 = 1 0 0 0 e 2 π i / 3 0 0 0 e 4 π i / 3 , U 02 = 1 0 0 0 e 4 π i / 3 0 0 0 e 8 π i / 3 , U 10 = 0 0 1 1 0 0 0 1 0 , U 11 = 0 0 e 4 π i / 3 1 0 0 0 e 2 π i / 3 0 , U 12 = 0 0 e 8 π i / 3 1 0 0 0 e 4 π i / 3 0 , U 20 = 0 1 0 0 0 1 1 0 0 , U 21 = 0 e 2 π i / 3 0 0 0 e 4 π i / 3 1 0 0 , U 22 = 0 e 4 π i / 3 0 0 0 e 8 π i / 3 1 0 0 .
Then, according to Euler’s equation, e i x = c o s x + i s i n x , an approximation of U can be obtained.

2.2. Quantum Neural Network

Quantum neural networks realize the basic structure of a neural network by considering the dynamics of quantum mechanics, which uses quantum memory and computing power advantages to achieve more efficient and smarter learning abilities. The quantum neural networks in this work use a general unitary operator as a quantum perceptron, this operator can only process quantum data and the transformations of quantum gates on quantum information that is unitary and linear. Quantum neurons can be mathematically represented as matrices, and the process of training quantum neural networks can be interpreted as iteratively adjusting the matrix values to improve the network’s ability to fit the input data.
A quantum neural network has a similar workflow to a classical neural network in terms of processing data, including task definition, data preparation, data preprocessing, training the network, optimizing the parameters, testing the network performance and selecting the optimal result. To achieve the detection of noisy quantum channels for quantum neural networks, the quantum neural network needs to learn a general detection method, whether for 2D or 3D noisy channels.

3. Methodology

3.1. Task Definition

The aim of this work is to employ quantum neural networks to identify and characterize the impact of noisy quantum channels, enabling the trained networks to replicate the effects of quantum noise channels. Before training starts, we need to finish the preparation steps, including data preparation, model building and parameter initialization. Then, we can realize the transmission of information by performing forward propagation and obtain the output states of the network. Through calculating the fidelity between the network’s output states and the output states in the data, we can obtain the cost function between them. If the current result does not satisfy the end condition, the unitary operation of the network is updated by the back propagation algorithm. The training procedure described above can be depicted in Figure 1. After training is finished, we can use the test data to evaluate the performance of the quantum neural networks. Furthermore, by adjusting the structure and other parameters of the networks, the network structure with the best performance can be selected.

3.2. Data Preparation

The data preparation in this work includes two subsections: preparation of the noisy quantum channel and preparation of the training data. The former outlines the steps taken to simulate a quantum channel with added noise, while the latter describes the generation of the input and target data for the neural network.

3.2.1. Preparation of the Noisy Quantum Channel

In practical applications, the form of the noisy quantum channel is always unknown. Thus, in this work, we simulate the noisy quantum channel by initializing it in a random form. As the form of the generated noisy quantum channel is unknown to the quantum neural network, it will not have any effect on training.
Taking the four channels in two-dimensional Hilbert space as an example, the preparation process for each noisy quantum channel can be described as follows:
  • It is required to generate the parameters p and γ in the expression randomly which take a value in ( 0 , 1 ) .
  • Each Kraus operation of the channel is assigned a weight w with a value in ( 0 , 1 ) , then we obtain the value E of the channel as
    E = i = 1 n w i e i ,
    where w i denotes the probability of occurrence of the i-th Kraus operation, e i is the value of i-th Kraus operation, n represents the number of Kraus operations in the channel and i = 1 n w i = 1 .

3.2.2. Preparation of the Training Data

According to the requirements of the quantum neural network model, the training data consist of pairs of a combination of input states and output states, which can be defined as
D t r a i n = { ( Ψ i n 1 , Ψ o u t 1 ) , ( Ψ i n 2 , Ψ o u t 2 ) , , ( Ψ i n n , Ψ o u t n ) } ,
where Ψ i n represents the input state, which is a normalized pure state generated randomly. The output state through the noisy quantum channel E can be described as | Ψ o u t = E | Ψ i n .
According to the definition of the cost function, the states need to be normalized to ensure that the calculated fidelity value is between 0 and 1. Thus, the output state should be normalized, i.e., | Ψ o u t = N o r m ( | Ψ o u t ) . After that, the training data are divided into two sets: the training set and the test set.

3.3. The Architecture of a Quantum Neural Network

The typical structure of a quantum neural network is similar to classical feed-forward neural networks, consisting of one input layer, one output layer and several hidden layers. For the hidden layers and the output layer, the output of the previous layer can be used as the input of the current layer. Figure 2 illustrates the architecture of a quantum neural network model, comprising two hidden layers, each containing three and four neurons, respectively.
Assume that the network has L hidden layers, the input of network is ρ i n and the output is ρ o u t , then the evolution of the quantum neural network can be expressed by the following equation:
ρ o u t t r i n , h i d ( U ( ρ i n | 0 0 h i d , o u t 0 0 | ) U ) ,
where U U o u t U L U L 1 U 1 represents the quantum circuit and U l denotes the U operator of the l-th layer which consists of neurons acting on the l 1 -th and l-th layers.
Meanwhile, the depth of the network can be changed by changing the number of hidden layers. Furthermore, the width of the network can be changed by changing the number of neurons in each layer.

3.4. Training Process

The training process of the network can be understood as the process of tuning the value of each neuron in the network, and the goal is to find a set of suitable solutions to maximize the cost function. The cost function is defined as the calculation of the average fidelity between all outputs and the expected outputs of the network. The higher the cost value, the closer the output is to the expected value. This reveals the better performance of the training effect of the network. For pure quantum states, it can be described as
C = 1 N x = 1 N ϕ x o u t | ρ x o u t | ϕ x o u t ,
where N represents the number of training data, ρ x o u t is the output of the network and | ϕ x o u t denotes the x-th output in the training data pair which is the expected output. When the output states of the training data are not pure, we use the fidelity for mixed states to describe it:
F ( ρ , σ ) : = [ t r ρ 1 / 2 σ ρ 1 / 2 ] 2 .
After obtaining the values of the cost function, updating the network parameters U is realized by backpropagation, and the network is then revolved with the new parameters U. The value of the cost function becomes larger after the update, i.e., the network fits the training data better. Backpropagation in quantum neural networks is referred to as the corresponding process in classical networks, and a detailed derivation can be found in [27].

4. Experiment Results

Here, in this part, we conduct experiments on 2D and 3D channels, and the results show that the quantum neural network can be generalized to higher dimensional channels.
We use the proposed quantum neural network to train each of the four noisy quantum channels that are commonly used in the qubit transmission system 100 times, and the test set is used to evaluate the performance of the network. The structure of the network is [1, 2, 1], which means it has three layers including one input layer, one hidden layer and one output layer; the number of neurons each layer has is one, two and one. The number of data both in the training set and test set is 10 pairs. The results are shown in Figure 3.
The figure shows that the network is capable of generalizing the available information. As expected, the training costs are higher than the test costs. For different channels, the phase damping channel approaches the highest costs on both the training set and the test set, the amplitude damping channel performs better than the other two channels and the depolarizing channel obtains the lowest costs.
By analyzing the expressions of these four channels, we find that they have different levels of complexity. For example, the phase damping channel has a counter-diagonal value of 0, the amplitude damping channel has a constant value of 0, the value in the bit-flip channel is a real number and the depolarizing channel has an imaginary part. Combined with the experimental results, we believe that the neural network has a better training ability for the channels with simpler structures. Specifically, the channel with a simple structure has higher medians, upper bounds and lower bounds, indicating it has higher costs and more stable results. However, it has more outliers, which are caused by overfitting of the data.
In the following, we use a quantum neural network with the structure of [1, 2, 1] to train a 3D noisy quantum channel 100 times, where the number of data both in the training set and test set is 10 pairs. We record the number of costs in each interval, and organize the results into a histogram, as shown in Figure 4.
The results show that the network could also be suitable for high-dimension channels. Although the cost is lower than that of lower dimensional channels, it also achieves an average of about 91% on the training set and 82.5% on the test set.

5. Methods to Improve the Performance of the Network

To improve the performance of the network, we conducted experiments and summarized all factors that may have effects on the network performance. Note that the conclusions we have drawn are generalized, the results shown below are representative and similar conclusions can be obtained even if different parameters are selected.
During training of a classical neural network, the performance of the model usually becomes better when the network structure becomes complex. We conducted related experiments on quantum neural networks but obtained different results. We trained networks with different architectures for the same amount of rounds, as shown in Table 1 and Table 2. We find that as the network structure becomes more complex, the performance of the network becomes worse. This is caused by the speed of convergence; the simpler the structure of the network, the easier it is to train. As the training rounds increase, their cost values gets closer and the maximum cost value that can be achieved in the end is similar.
The hyperparameter λ of the network is the reciprocal of learning rate η , that is to say, η = 1 / λ . We explored the effect of λ on network performance, and the results are shown in Figure 5.
From the results in the figure, we can conclude that λ affects the convergence rate. With a smaller value of λ , the faster the convergence rate. On the contrary, the convergence rate becomes slower with a larger value. However, this does not mean that a smaller value of λ is better; for networks with more complex structures, too small a λ will lead to abnormal convergence.
From the aspect of cost, we can see that as the number of iterations increases, the difference between the costs of different λ s becomes smaller.
Table 3 shows the results of training a network for 1000 rounds. During the training process, the current optimal model is continuously updated and the results are output every 100 rounds. A visualization of the results is shown in Figure 6.
From the results, we can see that with the increase in the training rounds, the performance of the current optimal model becomes better both on the training set and on the test set.
Moreover, we find there are differences in the results when training the same set of data several times. By analyzing the training process, we hypothesized that this is related to the random initialization of neurons represented by U-matrices. To verify the hypothesis, we performed ten experiments on one network; each experiment had the same parameters except for the U matrices.
The results are shown in Table 4 and Figure 7, from which we can learn that random initialization of U matrices does make a difference and this is uncontrollable.

6. The Robustness of the Network

In this section, we focus on the property of robustness of the quantum neural network. We first trained a network with 100 training pairs and recorded the results, then the training pairs were replaced by noisy pairs. Replacement with noisy pairs was performed in units of 10, the noise pairs used here were randomly generated and there was no relationship between the input and output states in the pairs. The results are shown in Table 5 and Figure 8.
From Figure 8, we can see that as the number of noisy pairs increases, both the training cost and the test cost decrease. When the number of noise pairs is less than 80, the value of the test cost is higher than 90%, which indicates the resistance of the network to noisy data, so we can conclude that the network is robust.

7. Discussion and Summary

In this paper, we use quantum neural networks to approximate the effect of a noisy quantum channel and conduct experiments on coherent noisy quantum channels. In the case of incoherent noisy quantum channels, the outputs are mixed states that can be described by density matrices. To model such channels, we can employ completely positive and trace-preserving (CPTP) maps and use incoherent quantum neural networks to approximate their effects. The cost function is then evaluated using the fidelity formula for mixed states.
In summary, we investigate methods for noisy quantum channel detection with a quantum neural network. The experiments conducted in this work have not yet been carried out on a quantum computer, but instead have been accomplished on a classical simulator. The current work mainly focuses on the study of noisy channels of different dimensions, determining how to improve the performance and testing the robustness of the quantum neural network.
According to the simulated results of 2D channels and 3D channels, we can conclude that using a quantum neural network may provide an efficient way to detect channel noise and it can be extended to higher dimensions. Then, we conducted experiments on all factors that may have impacts on network performance, including the network structure, the hyperparameters, the training rounds and random initialization. The results show that networks with simpler structures perform better when trained for the same amount of time. The hyperparameter λ affects the training speed; a network with a smaller λ has a higher training speed, but too small a λ leads to errors for networks with complex structures. The network structure and hyperparameter λ are set according to the problem at hand. The number of training rounds affects the cost both on the training set and the test set; it is beneficial to train as many times as possible. The random initialization of U matrices could lead to different results of the model and it is uncontrollable, but we can select the optimal model through multiple experiments to reduce its impact. Finally, we investigate the robustness of the network in noisy conditions.

Author Contributions

Software, L.L.; Investigation, B.L.; Writing—original draft, J.S.; Supervision, C.W. All authors have read and agreed to the published version of the manuscript.

Funding

The authors gratefully acknowledge the support from the National Natural Science Foundation of China through grants nos. 62131002 and 62071448 and the Fundamental Research Funds for the Central Universities (BNU).

Data Availability Statement

Data available on request from the corresponding author due to restrictions e.g., privacy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bennett, C.H.; DiVincenzo, D.P. Quantum information and computation. Nature 2000, 404, 247–255. [Google Scholar] [CrossRef] [PubMed]
  2. Knill, E.; Laflamme, R.; Martinez, R.; Tseng, C.H. An algorithmic benchmark for quantum information processing. Nature 2000, 404, 368–370. [Google Scholar] [CrossRef]
  3. Scarani, V.; Bechmann-Pasquinucci, H.; Cerf, N.J.; Dušek, M.; Lütkenhaus, N.; Peev, M. The security of practical quantum key distribution. Rev. Mod. Phys. 2009, 81, 1301. [Google Scholar] [CrossRef]
  4. Deng, F.G.; Li, X.H.; Zhou, H.Y.; Zhang, Z.J. Improving the security of multiparty quantum secret sharing against Trojan horse attack. Phys. Rev. A 2005, 72, 044302. [Google Scholar] [CrossRef]
  5. Long, G.L.; Deng, F.G.; Wang, C.; Li, X.H.; Wen, K.; Wang, W.Y. Quantum secure direct communication and deterministic secure quantum communication. Front. Phys. China 2007, 2, 251–272. [Google Scholar] [CrossRef]
  6. Wang, C. Quantum secure direct communication: Intersection of communication and cryptography. Fundam. Res. 2021, 1, 91–92. [Google Scholar] [CrossRef]
  7. Sheng, Y.B.; Zhou, L.; Long, G.L. One-step quantum secure direct communication. Sci. Bull. 2022, 67, 367–374. [Google Scholar] [CrossRef]
  8. Barz, S.; Fitzsimons, J.F.; Kashefi, E.; Walther, P. Experimental verification of quantum computation. Nat. Phys. 2013, 9, 727–731. [Google Scholar] [CrossRef]
  9. Howard, M.; Wallman, J.; Veitch, V.; Emerson, J. Contextuality supplies the ‘magic’ for quantum computation. Nature 2014, 510, 351–355. [Google Scholar] [CrossRef] [PubMed]
  10. Wu, X.; Liang, X.; Tian, Y.; Yang, F.; Chen, C.; Liu, Y.C.; Tey, M.K.; You, L. A concise review of Rydberg atom based quantum computation and quantum simulation. Chin. Phys. B 2021, 30, 020305. [Google Scholar] [CrossRef]
  11. Giovannetti, V.; Lloyd, S.; Maccone, L. Advances in quantum metrology. Nat. Photonics 2011, 5, 222–229. [Google Scholar] [CrossRef]
  12. Taylor, M.A.; Bowen, W.P. Quantum metrology and its application in biology. Phys. Rep. 2016, 615, 1–59. [Google Scholar] [CrossRef]
  13. Han, C.; Zhou, Z.W.; Guo, G.C. Long distance quantum communication over a noisy channel. J. Phys. B At. Mol. Opt. Phys. 2006, 39, 1677. [Google Scholar] [CrossRef]
  14. ur Rehman, J.; Farooq, A.; Jeong, Y.; Shin, H. Quantum channel discrimination without entanglement. Quantum Inf. Process. 2018, 17, 1–16. [Google Scholar] [CrossRef]
  15. Gao, W.C.; Mi, S.C.; Cao, C.; Liu, X.; Wang, T.J.; Wang, C. Faithful transmission of single-photon qubits using error-rejection coding. IEEE Photonics J. 2019, 11, 1–7. [Google Scholar]
  16. Chiaverini, J.; Leibfried, D.; Schaetz, T.; Barrett, M.D.; Blakestad, R.; Britton, J.; Itano, W.M.; Jost, J.D.; Knill, E.; Langer, C.; et al. Realization of quantum error correction. Nature 2004, 432, 602–605. [Google Scholar] [CrossRef] [PubMed]
  17. Schumacher, B.; Westmoreland, M.D. Approximate quantum error correction. Quantum Inf. Process. 2002, 1, 5–12. [Google Scholar] [CrossRef]
  18. Zhou, S.; Liu, Z.W.; Jiang, L. New perspectives on covariant quantum error correction. Quantum 2021, 5, 521. [Google Scholar] [CrossRef]
  19. Schuld, M.; Sinayskiy, I.; Petruccione, F. The quest for a quantum neural network. Quantum Inf. Process. 2014, 13, 2567–2586. [Google Scholar] [CrossRef]
  20. Shen, H.; Zhang, P.; You, Y.Z.; Zhai, H. Information scrambling in quantum neural networks. Phys. Rev. Lett. 2020, 124, 200504. [Google Scholar] [CrossRef]
  21. Shi, J.; Chen, S.; Lu, Y.; Feng, Y.; Shi, R.; Yang, Y.; Li, J. An approach to cryptography based on continuous-variable quantum neural network. Sci. Rep. 2020, 10, 1–13. [Google Scholar] [CrossRef]
  22. Qiu, P.H.; Chen, X.G.; Shi, Y.W. Detecting entanglement with deep quantum neural networks. IEEE Access 2019, 7, 94310–94320. [Google Scholar] [CrossRef]
  23. Nagy, A.; Savona, V. Variational quantum Monte Carlo method with a neural-network ansatz for open quantum systems. Phys. Rev. Lett. 2019, 122, 250501. [Google Scholar] [CrossRef] [PubMed]
  24. Lu, B.; Fan, C.R.; Liu, L.; Wen, K.; Wang, C. Speed-up coherent Ising machine with a spiking neural network. Opt. Express 2023, 31, 3676–3684. [Google Scholar] [CrossRef]
  25. Bisio, A.; Chiribella, G.; D’Ariano, G.M.; Facchini, S.; Perinotti, P. Optimal quantum learning of a unitary transformation. Phys. Rev. A 2010, 81, 032324. [Google Scholar] [CrossRef]
  26. Verdon, G.; Pye, J.; Broughton, M. A universal training algorithm for quantum deep learning. arXiv 2018, arXiv:1806.09729. [Google Scholar]
  27. Beer, K.; Bondarenko, D.; Farrelly, T.; Osborne, T.J.; Salzmann, R.; Scheiermann, D.; Wolf, R. Training deep quantum neural networks. Nat. Commun. 2020, 11, 1–6. [Google Scholar] [CrossRef] [PubMed]
  28. Nielsen, M.A.; Chuang, I. Quantum computation and quantum information. Am. J. Phys. 2002, 70, 558–559. [Google Scholar] [CrossRef]
Figure 1. The basic principle of Quantum Neural Networks.
Figure 1. The basic principle of Quantum Neural Networks.
Electronics 12 02430 g001
Figure 2. The architecture of quantum neural network models.
Figure 2. The architecture of quantum neural network models.
Electronics 12 02430 g002
Figure 3. The results of training on four different noisy channels.
Figure 3. The results of training on four different noisy channels.
Electronics 12 02430 g003
Figure 4. The training results on 3D channels. Here, the height of the column indicates the number of times the accuracy is within the corresponding interval in 100 training instances.
Figure 4. The training results on 3D channels. Here, the height of the column indicates the number of times the accuracy is within the corresponding interval in 100 training instances.
Electronics 12 02430 g004
Figure 5. The effect of hyperparameter λ on the quantum neural network.
Figure 5. The effect of hyperparameter λ on the quantum neural network.
Electronics 12 02430 g005
Figure 6. Visualization of Table 3.
Figure 6. Visualization of Table 3.
Electronics 12 02430 g006
Figure 7. Visualization of Table 4.
Figure 7. Visualization of Table 4.
Electronics 12 02430 g007
Figure 8. Visualization of Table 5.
Figure 8. Visualization of Table 5.
Electronics 12 02430 g008
Table 1. Performance of networks with different structures on a 2D channel.
Table 1. Performance of networks with different structures on a 2D channel.
Network Structure[1, 1][1, 2, 1][1, 2, 2, 1][1, 2, 4, 2, 1]
Training cost0.95980.95780.95420.9518
Test cost0.92950.92860.92720.9259
Table 2. Performance of networks with different structures on a 3D channel.
Table 2. Performance of networks with different structures on a 3D channel.
Network Structure[1, 1][1, 2, 1][1, 2, 2, 1]
Training cost0.94960.94830.9317
Test cost0.88250.90220.8958
Table 3. The relation between training rounds and cost.
Table 3. The relation between training rounds and cost.
Training Rounds1002003004005006007008009001000
Training cost0.94850.95380.95500.95540.95570.95580.95590.95600.95610.9561
Test cost0.89630.90350.90480.90520.90540.90560.90570.90570.90580.9058
Table 4. The relation between random initialization and cost.
Table 4. The relation between random initialization and cost.
Number of Times12345678910
Training cost0.94930.95060.94720.94780.94610.94640.94550.94830.95170.9502
Test cost0.90150.90170.90490.89740.90630.90480.90490.90750.89890.9042
Table 5. The influence of the noisy pairs on cost.
Table 5. The influence of the noisy pairs on cost.
Number of Noisy Pairs0102030405060708090100
Training cost0.96860.91390.89070.82240.79740.77340.69040.62860.60100.56230.5538
Test cost0.94490.94470.94430.94010.94260.94140.93510.91050.87960.59240.4884
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Song, J.; Lu, B.; Liu, L.; Wang, C. Noisy Quantum Channel Characterization Using Quantum Neural Networks. Electronics 2023, 12, 2430. https://doi.org/10.3390/electronics12112430

AMA Style

Song J, Lu B, Liu L, Wang C. Noisy Quantum Channel Characterization Using Quantum Neural Networks. Electronics. 2023; 12(11):2430. https://doi.org/10.3390/electronics12112430

Chicago/Turabian Style

Song, Junyang, Bo Lu, Lu Liu, and Chuan Wang. 2023. "Noisy Quantum Channel Characterization Using Quantum Neural Networks" Electronics 12, no. 11: 2430. https://doi.org/10.3390/electronics12112430

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop