Next Article in Journal
Elegante+: A Machine Learning-Based Optimization Framework for Sparse Matrix–Vector Computations on the CPU Architecture
Previous Article in Journal
Real-Time Electrical Energy Optimization in E-Commerce Systems Based on IoT and Mobile Agents
Previous Article in Special Issue
QSA-QConvLSTM: A Quantum Computing-Based Approach for Spatiotemporal Sequence Prediction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust and Scalable Quantum Repeaters Using Machine Learning

1
Department of Aerospace Engineering, Wichita State University, Wichita, KS 67260, USA
2
Wyant College of Optical Sciences, University of Arizona, Tucson, AZ 85719, USA
3
Department of Mathematics, Statistics, and Physics, Wichita State University, Wichita, KS 67260, USA
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Information 2025, 16(7), 552; https://doi.org/10.3390/info16070552 (registering DOI)
Submission received: 8 May 2025 / Revised: 23 June 2025 / Accepted: 25 June 2025 / Published: 28 June 2025
(This article belongs to the Special Issue Quantum Information Processing and Machine Learning)

Abstract

Quantum repeaters are integral systems to quantum computing and quantum communication as they allow the transfer of information between qubits, particularly over long distances. Because of the “no-cloning theorem,” which says that general quantum states cannot be directly copied, one cannot perform signal amplification in the usual way. The standard approach uses entanglement swapping, in which quantum states are teleported from one (short) segment to the next, using at each step a shared entangled pair. This is the job of the repeater. In general, this requires reliable quantum memories and shared entanglement resources, which are vulnerable to noise and decoherence. It is also difficult to manually create and implement the quantum algorithm for the swap circuit as the size of the system increases. Here, we propose a different approach: to use machine learning to train a repeater node. To demonstrate the feasibility of this method, the system is simulated in MATLAB 2022a. Training is conducted for a system of 2 qubits. It is then scaled up, with no additional training, to systems of 4, 6, and 8 qubits using transfer learning. Finally, the systems are tested in noisy conditions. The results show that the scale-up is very effective and relatively easy, and the effects of noise and decoherence are r e d u c e d as the size of the system increases.

1. Introduction and Literature Review

Large-scale quantum communication networks face a unique (quantum) problem: the standard (classical) amplification of a signal is impossible because a quantum state cannot be copied [1,2]. Thus, the attenuation of a signal from the scattering or absorption of the photons as they travel through optical fibers (or air) needs to be addressed in a different way. Quantum repeaters provide one solution for dealing with this. The standard approach uses entanglement swapping, in which quantum states are teleported from one (short) segment to the next, using at each step a shared entangled pair [3,4,5]. The usual approach also necessitates a reliable set of quantum memories at each node to store the entangled pairs used to perform the teleportation. There is the alternative of “all-photonic” quantum repeaters [6], but these require instead a very large overhead of photons and entangled resources [7]. Great progress has been made in all of the contributing components to this protocol: researchers have demonstrated entanglement swapping operations, quantum memories, and, even on a small scale, full quantum networks. But the field is still in flux in the sense that no one technology or physical implementation has been settled on [8,9,10] and many difficulties remain. A major issue is noise and decoherence effects. To deal with these, one can use some form of error correction, such as entanglement purification [3,11]. Substantial progress still needs to be made, and new approaches [12,13,14] need to be tried. For example, Liu et al. [12] propose returning to the old idea of vacuum beam guides in place of optical fibers, as they offer orders of magnitude lower loss. Mastriani [14] presents a simplified version of entanglement swapping that is less expensive for quantum repeaters. Another avenue being explored is quantum zero repeaters [15], which offer high-fidelity entanglement swapping and which require fewer computational resources. Azuma et al. [7] have a good comprehensive recent review of the various approaches.
In our research presented here, building on our prior quantum machine learning work, we propose a different kind of simplification: to use quantum machine learning [16] to perform all the necessary operations. In this way, we bypass the quantum memory and its attendant measurements. Scale-up is easy, as we use a quantum version of the well-known technique of transfer learning [17] to leverage knowledge of the larger system from the smaller system. Transfer learning is now in widespread use in classical machine learning; according to a survey paper from 2016 [18], over 700 papers in the previous 5 years had used the method and in a vast array of (classical) contexts. We applied this in earlier work on entanglement [19]. This transfer learning could be called “bootstrapping”, a term we coined when we first applied this to quantum computing [20] when thinking about how a classical computer uses a small “boot sector” program to launch a much larger OS. Most important, because of the parallel and distributed nature of the trained quantum neural network architecture, the resultant operation is robust to both noise and decoherence [21], analogously to the way classical neural networks have long been successfully used to deal with incomplete or damaged data. In fact, the network becomes more robust as the system size increases [22], which is a major result of our work presented here.
In general, using this kind of approach for a number of problems, we have found that the tedious complexity of developing quantum algorithms by hand can be forgone, handing the task over to a machine learning system. It also frequently shortens [23] and/or simplifies [24] a sequence from one found by the so-called “lego” method, using gates as building blocks. This is because instead of a sequence of steps, the problem is considered as a whole: to find the desired output from a given input. This kind of approach may have important and far reaching applications, for example, in quantum secure direct communication [25] or quantum key distribution (QKD) [26].

2. Methodology

2.1. System Description

We use a standard qubit Hamiltonian [27]. We label the amplitude for a qubit to flip between the states | 0 and | 1 as K (also called [27] the “tunneling” amplitude), ϵ for bias or relative energy of the two states, and ζ for coupling between pairs of qubits. The Hamiltonian for n qubits is then
H = i = 1 n K i σ x , i + i = 1 n ϵ i σ z , i + 1 2 i j n ζ i , j σ z , i σ z , j
where K i is the tunneling amplitude for qubit i, ϵ i is the bias for qubit i, ζ i , j is the coupling coefficient between the i j pair of qubits in the system, and σ x and σ z are the Pauli spin matrices.
The parameters K, ϵ , and ζ of the Hamiltonian are time-dependent, as then so is H. The quantum system state variable is the density matrix, ρ , which evolves from an initial ρ 0 at t = 0 to a ρ f at the final time t f according to the Schrodinger equation
d ρ d t = 1 i h [ H , ρ ]
The quantum parameters are trained using machine learning to achieve a final target density matrix at the final time ρ f = T . So, an initial ρ 0 paired with a target ρ f = T represents the required quantum state change. More details about this are given later. The time dependence of each parameter is represented as a finite Fourier series, and the Fourier coefficients are trained using an adjoint Levenberg–Marquardt machine learning method also described below.

2.2. Cost Function with Frobenius Output Measure

The quantum system evolves in time according to the Schrodinger equation, from an initial density matrix ρ 0 to a final density matrix ρ f at the final time t f . Recall that the Hamiltonian, H, is time varying. A cost function, L, is defined as a positive definite error based on the Frobenius norm of a matrix. T is the target output density matrix for the quantum system and ρ at t f is the actual quantum system output density matrix. The system dynamics are enforced as a constraint with Lagrange multiplier matrix γ where ⊙ is the element-by-element dot product of two matrices.
L = ( T i j ρ i j ) ( T i j ρ i j ) | t f + 0 t f γ d ρ d t 1 i h [ H , ρ ] d t + d ρ d t 1 i h [ H , ρ ] γ d t
Taking the variation with respect to γ and setting it equal to zero enforces the Schrodinger equation dynamics. Integrating the first term in each integral by parts and taking the variation with respect to ρ and setting that equal to zero at any time t < t f gives
d γ d t = 1 i h [ H , γ ]
and its complex conjugate equation. Note that since ρ is specified at t = 0, its variation at that initial time is zero, and that initial condition term vanishes. At t f , we have (and its conjugate):
γ i j = ( T i j ρ i j )
The Hamiltonian, H, contains parameters mentioned above, which describe how H varies with time. Let w be one of these parameters. Taking the variation in L with respect to w gives a gradient of L that can be used in various gradient-based adaptation methods to determine the parameter.
δ L δ w = 0 t f 2 R e 1 i h γ t d H d w , ρ
Following mainly [28], which is based on the Levenberg–Marquardt method of nonlinear optimization presented in [29,30,31], we derive a quasi-second order method for learning the quantum parameters in H. Levenberg–Marquardt outperforms gradient descent and conjugate gradient methods for medium-sized problems.
Let W be a vector of the quantum parameters. L is a function of W , so we define L W as a vector of partials of L with respect to each element of W and then define L ^ as a quadratic (2nd-order) local approximation to L about W 0 (the current value of the quantum parameter vector) as
L ^ ( W ) = L ( W 0 ) + L w W 0 T ( W W 0 ) + 1 4 L ( W 0 ) ( W W 0 ) L W W 0 T L W W 0 ( W W 0 ) T
notice that
L ^ ( W 0 ) = L ( W 0 )
and
L ^ W W 0 = L W W 0
and
L ( W 0 ) = ( T i j ρ i j ) t ( T i j ρ i j ) | t f
as the forward dynamics equation constraint being enforced means the other terms in L are zero. This means L ( W 0 ) is the squared “output” error at the final time with the current parameters W 0 . Then, setting the variation of L with respect to W equal to 0 gives
0 = L ^ W W = L W W 0 + 1 L ( W 0 ) ( W W 0 ) L W W 0 T L W W 0 .
Then define a Hessian by
H e s s = 1 L ( W 0 ) L W W 0 L W W 0 T 1 L ( W 0 )
giving
W = W 0 H e s s 1 L W W 0 T
The Levenberg–Marquardt algorithm modifies this by combining this Hessian update rule with a gradient descent update rule that is weighted by parameter λ that is updated dynamically during the learning process. Also, the Hessian and the gradient of L become averages over all training pairs in the training set, or in a mini-batch subset, and L ( W 0 ) becomes the average squared error over the entire set, by necessity, since it could be zero for any single training data point. For details on the LM process, including updating λ , see [32].
The combined weight update rule is
W = W 0 η [ H e s s + λ I ] 1 L w W 0 T
where I is the identity matrix and η is a constant that controls the rate of the update “learning” of the quantum parameters. The standard LM algorithm is
W = W 0 η [ H ^ e s s + λ I ] 1 E O w W 0 T
If we absorb L ( W 0 ) into η and λ and define the output O in the standard LM algorithm above as the squared error L, and have its target value as zero, our update using L looks much like the standard LM formulation. Because the error at the final time
γ i j = ( T i j ρ i j )
is propagated backward through the quantum system dynamics in Equation (3), the error is contained in the gradient calculated via Equation (5). Therefore, “error” E, which multiplies the gradient in the standard LM algorithm, is set to a vector of ones in our LM formulation.
The motivation for the approximation L ^ is as follows. Let a function of one variable, f(x), be approximated close to x = 0 by the form
f ( x ) ^ = 1 2 a ( 1 + b x ) 2
Note this form matches the form of the first term of L above. Then, choose a = 2f(0) to make f(0) = f ^ ( 0 ) and choose ab = f x 0 to make f x 0 = f ^ x 0 . Then,
f ^ ( x ) = f ( 0 ) + f x 0 x + 1 4 f ( 0 ) x 2 L x 0 T f x 0
Quoting from our previous paper [32], the following description of the LM procedure we use is as follows: “For small λ , the update rule is similar to the Gauss-Newton algorithm, allowing larger steps when the error is decreasing rapidly. For larger λ , the algorithm pivots to be closer to gradient descent and makes smaller updates to the weights. This flexibility is the key to L.M.’s efficacy, changing λ to adapt the step size to respond to the needs of convergence: moving quickly through the parameter space where the error function is steep and slowly when near an error plateau and thereby finding small improvements. Our implementation is a modified L.M. algorithm following several suggestions in [33]. One epoch of training consists of the following: (1) Compute the Jacobian L W and then the Hessian (11) with current weights w (2) Update the damping parameter λ (3) Calculate a potential parameter update (4) Find if RMS error has decreased with new parameters, or if an acceptable uphill step is found (5) If neither condition in step 3 is satisfied, reject the update, increase λ , and return to step 2 (6) For an accepted parameter change, keep the new parameters and decrease λ , ending the epoch. The identity matrix I that multiplies λ can be replaced by a scaling matrix D T D which serves the primary purpose of combating parameter evaporation [34], which is the tendency of the algorithm to push values to infinity while somewhat lost in the parameter space. Following [33], we can choose D T D to be a diagonal matrix with entries equal to the largest diagonal entries of H e s s yet encountered in the algorithm, with a minimum value of 10 6 . Updates to the damping factor may be done directly or indirectly; our results here use a direct method.”

2.3. System Training and Training Pairs

A list of valid quantum states and the resulting paired swap must be produced to create a set of input/target output pairs to train the network. Each input, with the exception of the charge basis inputs, results from creating a ket of random, uniformly distributed probability amplitude values via a random number generator and subsequently normalizing the ket. The output is then found by applying the appropriate swap operation matrix to the input. An example for the 2 qubit system is as follows: if the (random) input ket is | ψ | ζ , then the target output ket is | ζ | ψ , where ⊗ represents the tensor product. In other words, the receiving state is prepared in a random state | ζ ; after the operation, the incoming state | ψ is “transmitted” to the receiver. In total, 74 training pairs are generated, with 70 being generated via a random number generation and the remaining 4 training pairs are the standard charge basis [27], which, for a single qubit, is the set { | 0 , | 1 } . For an n qubit (4, 6 and 8) system, the incoming states | ψ and | ζ are n 2 qubit states (and can be entangled); the input state for all the qubits would then be | ψ | ζ and the output or target state would be | ζ | ψ . For testing the transfer learning scaled up from 2 qubits, we use 70 randomly generated inputs and the 2 n charge basis.
The input/target output for each training pair can then be represented in their density matrix forms T = ρ f = | Ψ f Ψ f | and ρ o = | Ψ o Ψ o | where | Ψ 0 and | Ψ f are the kets of the initial “input” and target final “output”. The system Hamiltonian parameters are then trained by presenting it with an initial “input” state from the training set, the system is evolved to the final time according to the Schrodinger equation, i ρ t = [ H , ρ ] , the co-state γ λ is calculated backward in time according to Equation (3) and the associated element of the gradient vector is calculated according to (5). Averaging over all training pairs, the Hessian is then calculated using (11) and the parameters updated with (13). After the completion of the training of the 2-qubit system, the system is then scaled up to 4, 6, and 8 qubits through transfer learning [19,20]. This was carried out by copying the tunneling, K, and bias, ϵ , from the original 2 qubits, respectively, to the added 2 qubits. The coupling, ζ , between the first 2 qubits is copied, respectively, to the coupling between the added 2 qubits. There is no coupling between the old 2 and the added 2 qubits. The results of the scale-up are obtained by testing on 70 randomly generated states and 2 n charge basis corresponding to the system’s size.

2.4. Noise Simulation

Once the system is trained, the task turns to proving the system’s resilience to noise. Noise is perhaps the most important obstacle to building large-scale quantum computers and networks, and a great deal of research has been devoted to this problem. One approach is error correction [35], but to be most effective, one needs to know ahead of time what the most likely sources of error will be [36]. These will vary from platform to platform. A full treatment would include both unital and non-unital interactions of the quantum system with its environment, which cause not only errors but also decoherence [37]. Here, in this initial work, we do not assume any particular physical implementation of the quantum repeater or model for the environmental effects, but confine ourselves to a simple model of random noise: physically, that there are (small) factors in the time evolution of the system that we do not know. Following [38], we imagine that the Hamiltonian for the noisy system consists of two parts: H t o t a l = H + H n , where H is the ideal Hamiltonian and H n the noise term, equal to some random Hermitian matrix. We do, however, split our random perturbations into real and imaginary parts in order to see, in the simulations, what specifically the resilience is to unitary errors and to dephasing. In line with our previous published work [22], we refer to “pure noise”, which is the addition of random magnitudes to the system, and to “decoherence”, which is the addition of random phases to the system. We performed simulations in which each was present alone and then when both were present (which we called “complex noise”). All three of these were used to test the system.
To add noise to the system, a noise matrix of small random values, that is, Hermitian, positive semi-definite, and possessing a trace of 1 (the required properties of a valid density matrix), is produced at each timestep for each training pair. This noise matrix is then added to the input state’s density matrix and normalized to produce a valid density matrix containing the desired type of noise. The trained system is then tested on specific swap tasks, that is, swap qubit A with qubit B and qubit C with qubit D, etc., depending on the size of the system. The inputs to which the noise is added for the 2-qubit system is identical to the inputs of the training set. For the 4-, 6-, and 8-qubit systems, the inputs to which the noise is added is the same as the ones used to test the system after scale-up.
In the Results Section below, we report an RMS error for the training and testing datasets as well as a measure of the average amount of noise introduced for the testing results. These are calculated as follows. The RMS error for the entire testing dataset is calculated by
R M S = i p a i r | | T ρ f | | F R N s i z e 2 2 N p a i r s
The noise level reported below is an average over the elements of the noise matrix, all the timesteps, and all the training pairs calculated by
A v e r a g e N o i s e = i p a i r i t i m e | | ρ n o i s e | | F R N s i z e 2 2 n t i m e s t e p s N p a i r s
For both equations above, | | A | | F R is the Frobenius norm of the matrix A. The system is tested with varying levels of average noise and the results are presented in the following section.

3. Results

To first establish a baseline for training effectiveness, a two-qubit system was trained on the dataset described above to sufficiently lower errors, and those parameters were then transferred [19,20] to the four-, six-, and eight-qubit cases with no additional training following the procedure described above. The baseline results are summarized in Table 1 for a testing set consisting of the charge basis states and 70 randomly generated quantum states for each system size. A jump is seen in the RMS error when scaling up from 2 to 4 qubits, but stabilizes from 4 to 6 and 6 to 8 qubits.
Next, noise was added to the system to test the network’s resiliency. Tests were conducted with pure noise and decoherence separately, as well as together, all at various levels, and the errors in the presence of noise are reported in the Table 2, Table 3, Table 4 and Table 5. RNP is a parameter that we use to generate the level of random noise numbers. The testing RMS is reported for each noise type at each RNP level. We find that the network’s noise tolerance improves as the system’s size increases, as demonstrated by the figures. Therefore, larger systems have the potential to be more noise-tolerant without the need for additional training.
Plots of testing RMS vs. increasing amounts of the average complex noise level are shown in the plots in Figure 1 for 2, 4, 6, and 8 qubits. These plots show the same scale for the vertical axis in order to clearly show the decrease in testing RMS with increasing numbers of qubits. Since 6 and 8 qubits shower lower effect of noise on the testing RMS, the maximum noise level tested was extended to better show the trend. Plots of testing RMS vs. increasing amounts of average complex noise level are shown in the plots in Figure 2 for 6 and 8 qubits.

4. Conclusions and Future Research

Machine learning can be enormously helpful in both quantum computing and quantum communication for several reasons: (1) it bypasses the difficult tasks of both algorithm construction and breaking down a desired unitary into a sequence of gates; (2) scale-up is relatively easy; and (3) the multiple inter-connectivity of the architecture leads to robustness to both noise and decoherence. In the present study, we see both of those advantages: we obtain low-error swap results in the presence of noise and decoherence, and the tolerance to that noise increases as the system size increases.
Robustness to incomplete and/or damaged data, and to environmental noise, is well known in classical neural networks [39]. The fact that this translates to the quantum regime is not only interesting but of great possible practical importance in quantum computing [21] as well as quantum communications. In an upcoming paper [40], we provide a mathematical justification for this in the 2-qubit case.
Since full quantum simulation goes up exponentially with the size of the system, given our computational resources, we were only able to do this exploratory study up to eight qubits. But we hope, in the near future, to be able to perform direct measurements using quantum hardware, which will not suffer from this limitation.
In the meantime, tolerance to all the different specific types of noise tolerance could be explored in simulations. Cheng et al. showed that noise can be modeled by using matrix product density operators, which can produce effective noise models while reducing the computational power required for a simulation on classical computers [41]. Noise might be added during training in addition to or instead of after training. And in the future, it will be interesting to see what kinds of improvements could be made were the training to be carried out online, on an actual quantum hardware system: we might expect to see robustness to actual physical noise and decoherence as the network could learn to take into account the physical flaws, crosstalk, or environment actually present.

Author Contributions

Conceptualization and methodology, J.S. and E.B.; software, J.S., D.F. and J.D.; simulations and validations, J.D. and D.F.; original draft preparation, J.D. and D.F.; writing, review, and editing, D.F., J.D., J.S. and E.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Requests for Matlab code, further data and information can be made via email to the authors.

Acknowledgments

We thank the entire research group for valuable discussions: Nathan Thompson, William Ingle, Anusha Krishna Murthy, and Nam Nguyen.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Dieks, D. Communication by EPR devices. Phys. Lett. A 1982, 92, 271–272. [Google Scholar] [CrossRef]
  2. Wootters, W.K.; Zurek, W.H. A single quantum cannot be cloned. Nature 1982, 299, 802–803. [Google Scholar] [CrossRef]
  3. Briegel, H.J.; Dür, W.; Cirac, J.I.; Zoller, P. Quantum repeaters for communication. arXiv 1998, arXiv:quant-ph/9803056. [Google Scholar]
  4. Pan, J.W.; Bouwmeester, D.; Weinfurter, H.; Zeilinger, A. Experimental entanglement swapping: Entangling photons that never interacted. Phys. Rev. Lett. 1998, 80, 3891. [Google Scholar] [CrossRef]
  5. Shi, Y.; Patil, A.; Guha, S. Measurement-Based Entanglement Distillation and Constant-Rate Quantum Repeaters over Arbitrary Distances. arXiv 2025, arXiv:2502.11174. [Google Scholar]
  6. Azuma, K.; Tamaki, K.; Lo, H.K. All-photonic quantum repeaters. Nat. Commun. 2015, 6, 6787. [Google Scholar] [CrossRef]
  7. Azuma, K.; Economou, S.; Elkouss, D.; Hilaire, P.; Jiang, L.; Lo, H.K.; Tzitrin, I. Quantum repeaters: From quantum networks to the quantum internet. Rev. Mod. Phys. 2023, 95, 045006. [Google Scholar] [CrossRef]
  8. Zhang, Y.L.; Jie, Q.X.; Li, M.; Wu, S.H.; Wang, Z.B.; Zou, X.B.; Zhang, P.F.; Li, G.; Zhang, T.; Guo, G.C.; et al. Proposal of quantum repeater architecture based on Rydberg atom quantum processors. arXiv 2024, arXiv:2410.12523. [Google Scholar]
  9. Zajac, J.M.; Huber-Loyola, T.; Hofling, S. Quantum dots for quantum repeaters. arXiv 2025, arXiv:2503.13775. [Google Scholar]
  10. Cussenot, P.; Grivet, B.; Lanyon, B.P.; Northup, T.E.; de Riedmatten, H.; Sørensen, A.S.; Sangouard, N. Uniting Quantum Processing Nodes of Cavity-coupled Ions with Rare-earth Quantum Repeaters Using Single-photon Pulse Shaping Based on Atomic Frequency Comb. arXiv 2025, arXiv:2501.18704. [Google Scholar]
  11. Chelluri, S.S.; Sharma, S.; Schmidt, F.; Kusminskiy, S.V.; van Loock, P. Bosonic quantum error correction with microwave cavities for quantum repeaters. arXiv 2025, arXiv:2503.21569. [Google Scholar]
  12. Gan, Y.; Azar, M.; Chandra, N.K.; Jin, X.; Cheng, J.; Seshadreesan, K.P.; Liu, J. Quantum repeaters enhanced by vacuum beam guides. arXiv 2025, arXiv:2504.13397. [Google Scholar]
  13. Mor-Ruiz, M.F.; Miguel-Ramiro, J.; Wallnöfer, J.; Coopmans, T.; Dür, W. Merging-based quantum repeater. arXiv 2025, arXiv:2502.04450. [Google Scholar]
  14. Mastriani, M. Simplified entanglement swapping protocol for the quantum Internet. Sci. Rep. 2023, 13, 21998. [Google Scholar] [CrossRef] [PubMed]
  15. Bayrakci, V.; Ozaydin, F. Quantum Zeno repeaters. Sci. Rep. 2022, 12, 15302. [Google Scholar] [CrossRef] [PubMed]
  16. Behrman, E.C.; Steck, J.E.; Kumar, P.; Walsh, K.A. Quantum algorithm design using dynamic learning. Quantum Inf. Comput. 2008, 8, 12–29. [Google Scholar] [CrossRef]
  17. Zhu, Z.; Lin, K.; Jain, A.K.; Zhou, J. Transfer Learning in Deep Reinforcement Learning: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 13344–13362. [Google Scholar] [CrossRef]
  18. Weiss, K.; Khoshgoftaar, T.M.; Wang, D. A survey of transfer learning. J. Big Data 2016, 3, 9. [Google Scholar] [CrossRef]
  19. Thompson, N.; Nguyen, N.; Behrman, E.; Steck, J. Experimental pairwise entanglement estimation for an N-qubit system: A machine learning approach for programming quantum hardware. Quantum Inf. Process. 2020, 19, 394. [Google Scholar] [CrossRef]
  20. Behrman, E.; Steck, J. Multiqubit entanglement of a general input state. Quantum Inf. Comput. 2013, 13, 36–53. [Google Scholar] [CrossRef]
  21. Behrman, E.; Nguyen, N.; Steck, J.; McCann, M. Quantum neural computation of entanglement is robust to noise and decoherence. In Quantum Inspired Computational Intelligence; Bhattacharyya, S., Maulik, U., Dutta, P., Eds.; Morgan Kaufmann: Boston, MA, USA, 2017; pp. 3–32. [Google Scholar] [CrossRef]
  22. Nguyen, N.H.; Behrman, E.C.; Steck, J.E. Quantum Learning with Noise and Decoherence: A Robust Quantum Neural Network. Quantum Mach. Intell. 2020, 2, 1–15. [Google Scholar] [CrossRef]
  23. Rethinam, M.; Javali, A.; Hart, A.; Behrman, E.; Steck, J. A genetic algorithm for finding pulse sequences for nmr quantum computing. Paritantra—J. Syst. Sci. Eng. 2011, 20, 32–42. [Google Scholar]
  24. Nola, J.; Sanchez, U.; Murthy, A.K.; Behrman, E.; Steck, J. Training microwave pulses using machine learning. Acad. Quantum 2025. [Google Scholar] [CrossRef]
  25. Pan, D.; Lin, Z.; Wu, J.; Zhang, H.; Sun, Z.; Ruan, D.; Yin, L.; Long, G.L. Experimental free-space quantum secure direct communication and its security analysis. Photonics Res. 2020, 8, 1522–1531. [Google Scholar] [CrossRef]
  26. Cao, Y.; Zhao, Y.; Zhang, J.; Wang, Q.; Niyato, D.; Hanzo, L. From single-protocol to large-scale multiprotocol quantum networks. IEEE Netw. 2022, 36, 14–22. [Google Scholar] [CrossRef]
  27. Chuang, I.; Nielsen, M. Quantum Computation and Quantum Information; Cambridge University Press: Cambridge, UK, 2000. [Google Scholar]
  28. Roweis, S. Levenberg-Marquardt Optimization. Available online: https://people.duke.edu/~hpgavin/SystemID/References/lm-Roweis.pdf (accessed on 23 June 2025).
  29. Levenberg, K. A Method for the Solution of Certain Non-Linear Problems in Least Squares. Q. Appl. Math. 1944, 2, 164–168. [Google Scholar] [CrossRef]
  30. Marquardt, D. An Algorithm for Least-Squares Estimation of Nonlinear Parameters. J. Soc. Ind. Appl. Math. 1963, 11, 431–441. [Google Scholar] [CrossRef]
  31. More, J.J. The Levenberg-Marquardt Algorithm: Implementation and Theory; Springer: Berlin/Heidelberg, Germany, 1978; pp. 431–441. [Google Scholar]
  32. Steck, J.E.; Thompson, N.L.; Behrman, E.C. Programming Quantum Hardware via Levenberg-Marquardt Machine Learning. In Intelligent Quantum Information Processing; CRC Press: Boca Raton, FL, USA, 2024. [Google Scholar]
  33. Transtrum, M.K.; Sethna, J.P. Improvements to the Levenberg Marquardt algorithm for nonlinear least- squares minimization. arXiv 2012, arXiv:1201.5885. [Google Scholar]
  34. Transtrum, M.K.; Machta, B.B.; Sethna, J.P. Geometry of nonlinear least squares with applications to sloppy models and optimization. Phys. Rev. E 2011, 83, 036701. [Google Scholar] [CrossRef]
  35. Gottesman, D. An introduction to quantum error correction and fault tolerant quantum computation. Proc. Symp. Appl. Math. 2010, 13. [Google Scholar] [CrossRef]
  36. Knill, E.; Laflamme, R.; Zurek, W. Resilient quantum computation: Error models and thresholds. Proc. R. Soc. Lond. A 1998, 454. [Google Scholar] [CrossRef]
  37. Georgopoulos, K.; Emary, C.; Zuliani, P. Modelling and simulating the noisy behavior of near-term quantum compiuters. Phys. Rev. A 2021, 104, 062432. [Google Scholar] [CrossRef]
  38. Markiewicz, M.; Puchala, Z.; de Rosier, A.; Laskowski, W.; Zyczkowski, K. Quantum noise generated by local random Hamiltonians. Phys. Rev. A 2017, 95, 032333. [Google Scholar] [CrossRef]
  39. Lu, W.; Zhang, Z.; Qin, F.; Zhang, W.; Lu, Y.; Liu, Y.; Zheng, Y. Analysis on the inherent noise tolerance of feedforward network and one noise-resilient structure. Neural Netw. 2023, 165, 786–798. [Google Scholar] [CrossRef]
  40. Rodriguez, R.; Nguyen, N.; Behrman, E.; Li, A.; Steck, J. Existence of a robust optimal control process for efficient measurements in a two-qubit system. arXiv 2025, arXiv:2506.19122. [Google Scholar]
  41. Cheng, S.; Cao, C.; Zhang, C.; Liu, Y.; Hou, S.Y.; Xu, P.; Zeng, B. Simulating noisy quantum circuits with matrix product density operators. Phys. Rev. Res. 2021, 3, 023005. [Google Scholar] [CrossRef]
Figure 1. Plots of RMS vs. average noise level for 2, 4, 6, and 8 qubits with complex noise.
Figure 1. Plots of RMS vs. average noise level for 2, 4, 6, and 8 qubits with complex noise.
Information 16 00552 g001
Figure 2. Plots of RMS vs. average noise level for 6 and 8 qubits with higher complex noise levels.
Figure 2. Plots of RMS vs. average noise level for 6 and 8 qubits with higher complex noise levels.
Information 16 00552 g002
Table 1. RMS values from testing WITHOUT noise.
Table 1. RMS values from testing WITHOUT noise.
n qubits Testing RMS
2 5.38 × 10 8
4 9.02 × 10 6
6 9.62 × 10 7
8 7.32 × 10 8
Table 2. Error induced by noise for 2 qubits.
Table 2. Error induced by noise for 2 qubits.
RNPPure Noise RMSDecoherence RMSComplex Noise RMS
1 × 10 6 1.97 × 10 5 4.91 × 10 5 9.44 × 10 5
1 × 10 5 2.01 × 10 4 4.94 × 10 4 9.44 × 10 4
1 × 10 4 0.00190.00470.0085
Table 3. Error induced by noise for 4 qubits.
Table 3. Error induced by noise for 4 qubits.
RNPPure Noise RMSDecoherence RMSComplex Noise RMS
1 × 10 6 1.42 × 10 5 2.28 × 10 5 8.30 × 10 5
1 × 10 5 1.07 × 10 4 1.98 × 10 4 6.66 × 10 4
1 × 10 4 8.30 × 10 4 0.00120.0021
Table 4. Error induced by noise for 6 qubits.
Table 4. Error induced by noise for 6 qubits.
RNPPure Noise RMSDecoherence RMSComplex Noise RMS
1 × 10 6 1.03 × 10 5 1.94 × 10 5 9.73 × 10 5
1 × 10 5 7.24 × 10 5 1.06 × 10 4 1.74 × 10 4
1 × 10 4 1.72 × 10 4 1.75 × 10 4 1.99 × 10 4
Table 5. Error induced by noise for 8 qubits.
Table 5. Error induced by noise for 8 qubits.
RNPPure Noise RMSDecoherence RMSComplex Noise RMS
1 × 10 6 6.56 × 10 6 9.26 × 10 6 1.36 × 10 5
1 × 10 5 1.34 × 10 5 1.36 × 10 5 1.43 × 10 5
1 × 10 4 2.45 × 10 5 1.39 × 10 5 4.53 × 10 5
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Fuentealba, D.; Dahn, J.; Steck, J.; Behrman, E. Robust and Scalable Quantum Repeaters Using Machine Learning. Information 2025, 16, 552. https://doi.org/10.3390/info16070552

AMA Style

Fuentealba D, Dahn J, Steck J, Behrman E. Robust and Scalable Quantum Repeaters Using Machine Learning. Information. 2025; 16(7):552. https://doi.org/10.3390/info16070552

Chicago/Turabian Style

Fuentealba, Diego, Jackson Dahn, James Steck, and Elizabeth Behrman. 2025. "Robust and Scalable Quantum Repeaters Using Machine Learning" Information 16, no. 7: 552. https://doi.org/10.3390/info16070552

APA Style

Fuentealba, D., Dahn, J., Steck, J., & Behrman, E. (2025). Robust and Scalable Quantum Repeaters Using Machine Learning. Information, 16(7), 552. https://doi.org/10.3390/info16070552

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop