Previous Article in Journal
Processing Maps and Nano-IR Diagnostics of Type I Modifications in Mid-IR Germanate-Based Optical Glass
Previous Article in Special Issue
Learning Framework for Underwater Optical Localization Using Airborne Light Beams
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Communication

Adaptive Routing for Meshed QKD Networks of Flexible Size Using Deep Reinforcement Learning

Chair of Communications, Research Institute of Electrical Engineering and Information Technology, Kiel University, 24143 Kiel, Germany
*
Author to whom correspondence should be addressed.
Photonics 2026, 13(2), 198; https://doi.org/10.3390/photonics13020198
Submission received: 14 January 2026 / Revised: 4 February 2026 / Accepted: 14 February 2026 / Published: 16 February 2026
(This article belongs to the Special Issue Machine Learning and Artificial Intelligence for Optical Networks)

Abstract

Quantum Key Distribution (QKD) networks guarantee information-theoretical security of exchanged keys, but key rates are still limited. This makes efficient and adaptive routing a critical challenge, especially in meshed topologies without quantum repeaters. Conventional shortest path routing approaches struggle to cope with dynamic key store filling levels and changes in network topologies, which leads to load imbalance and blocked connections. In this work, we propose an adaptive routing framework based on Deep Reinforcement Learning (DRL) for hop-wise end-to-end routing in unknown meshed QKD networks. The agent leverages Graph Attention Networks (GATs) to process the network states of varying topologies, enabling generalization across previously unseen meshed networks without topology-specific retraining. The agent is trained on random graphs with 10 to 20 nodes and learns a routing policy that explicitly balances key consumption across the network by utilizing a reward function that is based on the entropy of key store filling levels. We evaluate the proposed approach on the 14-node NSFNET topology under time-varying traffic demands. Simulation results demonstrate that the DRL-based routing significantly outperforms hop-based and weighted shortest path benchmarks, achieving up to a 18.7% increase in mean key store filling levels while completely avoiding key store depletion. These results highlight the potential of graph-based DRL methods for scalable, adaptive, and resource-efficient routing in future QKD networks.
Keywords: Quantum Key Distribution; QKD networks; adaptive routing; Deep Reinforcement Learning; Graph Attention Networks; load balancing; Resource-Aware Optimization Quantum Key Distribution; QKD networks; adaptive routing; Deep Reinforcement Learning; Graph Attention Networks; load balancing; Resource-Aware Optimization

Share and Cite

MDPI and ACS Style

Johann, T.; Kühl, S.; Pachnicke, S. Adaptive Routing for Meshed QKD Networks of Flexible Size Using Deep Reinforcement Learning. Photonics 2026, 13, 198. https://doi.org/10.3390/photonics13020198

AMA Style

Johann T, Kühl S, Pachnicke S. Adaptive Routing for Meshed QKD Networks of Flexible Size Using Deep Reinforcement Learning. Photonics. 2026; 13(2):198. https://doi.org/10.3390/photonics13020198

Chicago/Turabian Style

Johann, Tim, Sebastian Kühl, and Stephan Pachnicke. 2026. "Adaptive Routing for Meshed QKD Networks of Flexible Size Using Deep Reinforcement Learning" Photonics 13, no. 2: 198. https://doi.org/10.3390/photonics13020198

APA Style

Johann, T., Kühl, S., & Pachnicke, S. (2026). Adaptive Routing for Meshed QKD Networks of Flexible Size Using Deep Reinforcement Learning. Photonics, 13(2), 198. https://doi.org/10.3390/photonics13020198

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop