Abstract
The coupling of quantum computing with multi-agent reinforcement learning (MARL) provides an exciting direction to tackle intricate decision-making tasks in high-dimensional spaces. This work introduces a new quantum-inspired multi-agent reinforcement learning (QiMARL) model, utilizing quantum parallelism to achieve learning efficiency and scalability improvement. The QiMARL model is tested on an energy distribution task, which optimizes power distribution between generating and demanding nodal power stations. We compare the convergence time, reward performance, and scalability of QiMARL with traditional Multi-Armed Bandit (MAB) and Multi-Agent Reinforcement Learning methods, such as Greedy, Upper Confidence Bound (UCB), Thompson Sampling, MADDPG, QMIX, and PPO methods with a comprehensive ablation study. Our findings show that QiMARL yields better performance in high-dimensional systems, decreasing the number of training epochs needed for convergence while enhancing overall reward maximization. We also compare the algorithm’s computational complexity, indicating that QiMARL is more scalable to high-dimensional quantum environments. This research opens the door to future studies of quantum-enhanced reinforcement learning (RL) with potential applications to energy optimization, traffic management, and other multi-agent coordination problems.
1. Introduction
The fast-evolving shift to renewable sources of energy and distributed power generation has made it increasingly difficult to manage modern energy systems. The traditional centralized control structures can no longer handle the spatial and temporal variability of power generation and load. Instead, control and optimization of these systems are being decentralized with increasing focus on intelligent decision-making at the edge. In particular, smart grids and microgrids demand real-time energy distribution among the generating and consuming nodes, often with the additional constraints of uncertainty in demand, time-variant supply, and unavailability of information. These systems naturally fit into being defined as multi-agent systems, with each node (or agent) optimizing its own behavior while contributing towards achieving system performance [,,].
In answer to these issues, the subarea of MARL has attracted much attention in recent times. MARL allows agents to learn the optimal decision policies by interacting with their environment and receiving feedback in the form of rewards or penalties. However, even though MARL holds much promise for such decentralized environments, it also possesses major drawbacks. With an increase in the number of agents, the state space and joint action spaces also increase combinatorially, which is a scalability issue with exponential computational requirements. Moreover, credit assignment becomes more difficult as it is challenging to apportion the rewards to the actions of every agent, especially when operating in cooperative environments. In addition, search strategies utilized in conventional MARL (e.g., -greedy or entropy-based approaches) are generally not capable of effectively exploring large or continuous state spaces and cause slow learning and poor convergence.
These challenges have driven research into alternative paradigms of learning able to enable more effective learning and coordination in complicated multi-agent systems. One of these lines is to capitalize on knowledge from quantum computing, particularly the structural advantages that come from quantum mechanics—superposition, entanglement, and non-classical sampling. While large-scale quantum hardware is presently in its early stages of development, a growing number of researchers have dared to explore the possibility of injecting classical algorithms with quantum-inspired mechanisms. These methods do not rely on real quantum hardware but instead simulate quantum phenomena in classical computing environments, thereby taking advantage of some of the theoretical strengths of quantum models without requiring quantum equipment [,,].
In this paper, we propose a new learning framework known as QiMARL, which integrates simulated quantum phenomena into the learning process of MARL. QiMARL leverages the principles of quantum superposition to store several possible actions in parallel, enabling richer and more organized exploration of the policy space. The algorithm also stores interdependencies between agents through the simulation of quantum entanglement, enabling coordinated decision-making with message passing or centralized control. The algorithm also uses Quantum Amplitude Estimation (QAE)-style protocols to guide value estimation and sampling strategies better compared to regular Monte Carlo procedures [,,,,,]. All of these quantum-inspired calculations are performed with IBM’s Qiskit software and executed on classical simulators, the Aer backend in this case, which supports high-performance quantum circuit emulation. This design ensures that QiMARL is deployable today, even in environments with no access to quantum hardware.
A breakthrough in QiMARL is how it treats the joint policy space as a structured quantum state that evolves over time. Instead of either decentralized local updates or a fully centralized critic model, each agent in QiMARL responds to a parameterized quantum circuit whose gates are both locally modulated and entangled with other agents’ correlations. As a result, the system develops in a way that is amenable both to decentralized autonomy as well as global coherence and that represents a middle ground between fully independent and fully centralized learning. Furthermore, using simulated quantum measurements as a choice mechanism provides a controlled but random method for selecting actions, enhancing policy variety, and avoiding premature convergence to local minima.
To evaluate the effectiveness of QiMARL, we take an example in a real-world scenario in the field of dynamic energy allocation. Specifically, we simulate a nodal power grid consisting of a number of generating and consuming nodes. Every node is a self-interested agent with the objective of optimizing long-term energy allocation. The world is cooperative and partially observable, and agents must learn to coordinate their actions based on local reward feedback. In our experimental setup, we compare QiMARL against several reinforcement learning baselines, including Greedy, UCB [,], Thompson Sampling [], MADDPG (Multi-Agent Deep Deterministic Policy Gradient), QMIX (Q-value Mixing Network), and PPO (Proximal Policy Optimization) methods. Adding them provides a clear baseline against which to judge whether or not QiMARL’s simulated quantum enhancements yield tangible benefits regarding convergence, reward optimization, and scalability [,,].
Alongside presenting the QiMARL framework, this paper presents a controlled comparison of quantum-inspired learning against RL baselines in a real-world simulation problem with a comprehensive ablation analysis. Not only does the study reveal the possible practical applications of using quantum-inspired approaches, but it also confirms their effectiveness on classical hardware, opening up new avenues of research and application. Alongside these, we provide the corresponding background on bandit-based methods, traditional MARL, and quantum-inspired learning in Section 2, before presenting the QiMARL architecture in detail in Section 3 and Section 4. Experimental setup and results are provided in Section 5 and Section 6, while limitations and conclusions are given in Section 7 and Section 8.
Research Contributions
This paper makes the following primary contributions:
- We present QiMARL, a quantum-inspired multi-agent reinforcement learning algorithm that leverages classically simulated quantum operations like superposition, entanglement, and amplitude estimation to improve exploration, policy coordination, and convergence rate in distributed scenarios.
- We introduce a comprehensive classical simulation setup for QiMARL, employing IBM Qiskit. All quantum operations are performed on the Aer simulator backend, rendering the method readily deployable in current real-world applications without involving quantum hardware.
- We design and test QiMARL in the context of a realistic energy distribution scenario, where agents are neighborhood nodes with the requirement to dynamically adjust energy supply and demand. This test case is a realistic and challenging example of how to measure coordination under uncertainty.
- We compare QiMARL with large empirical reductions to single-agent reinforcement learning baselines—Greedy, UCB, and Thompson Sampling—and demonstrate that QiMARL always achieves greater cumulative rewards, more stable learning curves, and significantly better scalability with more agents [,,].
2. Literature Review
QiMARL’s development intersects with several established and emerging fields of study. These are the traditional MARL, lightweight single-agent MAB strategies, and the quantum or quantum-inspired computing paradigms that offer competing information representation and policy sampling mechanisms. In this section, we provide a comprehensive overview of these domains to put our proposed method’s novelty and relevance into context.
2.1. Multi-Agent Reinforcement Learning (MARL)
MARL has emerged as a central research area in developing intelligent, decentralized decision-making systems. Unlike single-agent RL, with the environment supposed to be stationary and fully observable, MARL works in settings where multiple agents learn simultaneously with partial observability and non-stationary dynamics due to simultaneous policy updates. This renders it more complicated and introduces inherent issues related to coordination, stability, and scalability.
Over the past decade, much has been achieved to overcome them. Centralized training with decentralized execution (CTDE) architectures such as MADDPG, COMA (Counterfactual Multi-Agent policy gradients), and QMIX allow global information access during training while maintaining local policy execution during inference. These models have been shown to handle cooperative as well as mixed cooperative–competitive tasks well, particularly in domains like robotic control, autonomous vehicle control, and real-time strategy games. However, CTDE methods typically require access to global state, communications pathways, or mutual reward signals during training, which cannot be achieved in real-world deployment in systems such as power grids or sensor grids [,,].
Another research area is value factorization and decomposition methods such as VDN (Value Decomposition Network) and QMIX, which seek to decompose the joint Q-function into agent-specific components. While this promotes decentralization, it forces structural assumptions that limit policy expressiveness or optimality in complex scenarios. Similarly, independent Q-learning (IQL) solutions attempt to scale by treating each agent as a standalone learner, but these are ravaged by instability due to the non-stationarity of the environment [].
More recent trends include attention mechanisms, graph neural networks (GNNs), and message passing structures to encode inter-agent dependencies. While effective, they tend to be computationally costly and less interpretable. There is, therefore, increasing interest in hybrid or alternative structures that are guaranteed to keep agent autonomy intact while implying stability and coordination [,].
QiMARL innovates in this direction by introducing a quantum-inspired mechanism for agent entanglement and decision sampling. Rather than depending on explicit message passing or common critics, QiMARL introduces simulated quantum operations to project agent correlations and drive diversified, entangled exploration strategies—embodying emergent coordination without requiring full state observability or common communication protocols. Recent studies suggest that entanglement-inspired designs can foster efficient coordination in distributed systems like grids and transportation networks [,].
2.2. Bandit-Based Exploration Strategies in Decentralized Learning
Though MARL possesses strong theoretical frameworks for the cooperative system model, computational expenses could become undue in resource-constrained environments or real-time systems. Under these circumstances, scaled-down decision models founded on MAB theory are an easy solution.
In the basic MAB problem, an agent repeatedly selects one of a finite number of actions (arms), each with a reward distribution unknown to the agent. The goal is to collect maximum reward in the long run by balancing exploration (trying new arms) and exploitation (playing the best-known arm). Many exploration strategies have been developed over the years:
- Greedy approaches target short-term reward and use no exploration. -Greedy approaches add random exploration with a constant or diminishing probability.
- UCB approaches add confidence bounds around the estimate of rewards to induce exploration of unknown actions.
- Thompson Sampling uses Bayesian posterior sampling to inherently balance exploration and exploitation in a principled way.
These methods are typically used as lightweight single-agent reinforcement learning protocols, particularly for scenarios where agents talk sparingly to one another or the state–action space is limited. For distributed systems—say, wireless network selection, task offloading in edge computing, or energy demand response—individual agents using bandit methods can learn adaptive policies without coordination.
In the setting of this study, we apply MAB-based methods not as substitutes for end-to-end MARL frameworks nor as tools to substitute them. Instead, they serve as baselines for comparison. Since QiMARL suggests a new hybrid decision mechanism, having these baselines allows us to isolate the effects of its quantum-inspired aspects. In particular, we ask whether the richness in structure of QiMARL translates to measurable reward effectiveness and scalability improvements over decentralized, communication-free heuristics.
It should be noted that MAB models are lacking in the ability to manage stateful state transitions, sequential dependence, or cooperative planning. Yet their strong performance in low-dimensional or loosely coupled worlds made them a reference point of interest in research where MARL would be too computationally intensive, or where agents enjoy high autonomy.
2.3. Quantum Machine Learning and Quantum-Inspired Algorithms
Quantum computing is a developing technology that offers radically new models for information and algorithmic processing. Although fault-tolerant, universal quantum computers are still under development, there has been significant progress in the simulation of quantum circuits and the application of hybrid quantum–classical algorithms to machine learning problems. Quantum machine learning (QML) in this context explores the ways quantum effects—superposition, entanglement, and quantum interference—are able to enrich learning models, optimization processes, and data processing capabilities.
Several QML techniques are applicable to decision-making and reinforcement learning, such as the following ones:
- QAE provides a way to estimate expected values better than classical Monte Carlo algorithms through quantum phase estimation. In theory, this can reduce the sample size from O(1/2) to O(1/) [].
- Variational Quantum Algorithms (VQAs), such as the Variational Quantum Eigensolver (VQE) and Quantum Approximate Optimization Algorithm (QAOA), facilitate parameterized quantum circuit optimization based on classical feedback loops [,].
- Quantum Monte Carlo (QMC) methods imitate probabilistic systems through quantum mechanics and have even been utilized for the simulation of complex physics and finance systems [].
Subsequent works have explored the usage of quantum circuits for policy approximations in reinforcement learning, which offer richer hypothesis spaces and non-linear mappings. Even though theoretical predictions suggest potential benefits such as exponential speedup in some tasks [], such benefits are currently only offered in small-scale or highly controlled environments. Quantum advantage for practical machine learning remains hypothetical since hardware evolution is in its early stages, plagued with problems like decoherence, limited qubits, and gate reliability. As a result, while papers such as [] observe promising hybrid model results in narrow domains, currently, there are no frequent empirical data attesting to the enhanced performance of quantum versus classical models at scale.
However, current quantum hardware constraints, such as decoherence, low qubit counts, and gate noise, preclude direct deployment of such models on most real-world tasks. To close this gap, researchers have begun exploring quantum-inspired models: algorithms for emulating significant quantum behavior with classical computation. They mimic quantum states, quantum gates, and measurements probabilistically or deterministically in a manner that current high-performance computing infrastructure can be used. Quantum-inspired models may provide insight into the behavior of quantum systems without having to wait for them to be feasible today [].
QiMARL neatly sits within this paradigm. Rather than necessarily accessing a quantum computer, it uses the simulator from Qiskit’s Aer to run quantum circuits that guide the learning and decision-making process. By allowing simulated superposition and entanglement between agents, QiMARL provides a deployable, scalable paradigm for leveraging some of the particular benefits of quantum learning—such as parallel policy evaluation and structured exploration—without needing quantum hardware.
2.4. Quantum Cognition and Quantum-Cognitive Neural Networks
Parallel to engineering-focused QML, quantum cognition applies the mathematical formalism of quantum theory to model human judgment and decision-making. Researchers like Busemeyer and Pothos have shown that quantum probability theories are capable of explaining behavioral phenomena—e.g., order effects, conjunction/disjunction fallacies, and intransitive preferences—that baffle classical theories.
Quantum-cognitive neural networks capitalize on these ideas and embed quantum-probabilistic reasoning into deep learning structures with the aim of modeling context-dependent decision-making as well as uncertainty estimation. These models bridge cognitive theory and computational design and could inspire RL architectures that better align with human-like decision processes. In the context of QiMARL, quantum cognition offers an additional perspective: entanglement-like correlations and contextual dependencies are not only relevant in physics-based models but also in modeling complex, decentralized decision-making [,,].
2.5. Research Gap
Traditional MARL structures are an abundant resource for multi-agent decision making but are constrained by scalability and complexity. Lightweight bandit algorithms are strong in learning within distributed systems but lack policy richness and coordination. Quantum machine learning possesses theoretical advantages but is currently not advisable to use due to hardware constraints.
QiMARL closes this gap by combining MARL’s coordination ability, MAB’s performance, and quantum computation’s structural advances in a classical simulation framework. It is not intended to replace end-to-end MARL or QML pipelines but to demonstrate that quantum-inspired learning architectures can be used to improve performance in realistic, resource-constrained environments.
In doing so, this paper contributes not just a new algorithm but an empirical link between the theoretical quantum learning and real-world implementation and gains new insight into how ordered exploration and entanglement-inspired mechanisms can be used in distributed decision-making systems [].
3. System Model for the Proposed Research Analysis
3.1. Energy Supply and Demand Representation
The energy system consists of power supply from the generating station and demand at time t from the nodal power station:
where and are the elements of a dataset .
3.2. Quantum Circuit Initialization
A quantum circuit with n qubits is initialized, representing the possible combinations of generating power stations and demanding nodal power stations, with each qubit set to an equal superposition state using Hadamard gates:
where is the Hadamard gate.
3.3. Entanglement Structure
A set of entanglement patterns is defined as:
At each epoch , the entanglement pattern is selected based on the smoothed reward:
Entanglement are applied based on :
3.4. Quantum Variational Optimization
The quantum variational form of the Hamiltonian is given by
where and are the parameters learned during training, and QAE is for enable parallelism. The optimal angles are found using:
3.5. Data Encoding and Measurement
Classical data are encoded using rotation gates:
Rotations are applied and given by
The energy allocation estimate is computed using quantum expectation values as
3.6. Reward Function
The local reward function is given by
The total reward for epoch is accumulated:
3.7. Convergence Time and Global Reward Adjustment
The execution time for each agent is given by
The global reward considering execution time is given by
3.8. Optimal Policy Extraction
The final optimal policy and rewards are obtained by
4. Proposed Algorithm
4.1. Algorithm Description
This algorithm introduces a QiMARL (Algorithm 1) method for energy optimization that combines quantum variational optimization and reinforcement learning principles. The algorithm runs iteratively across several epochs with the objective of optimizing energy distribution using quantum computing methods.
The procedure begins with the initialization of a quantum circuit of n qubits and the selection of hyperparameters. A synthetic dataset is created, which is based on real-time demand–supply distribution, with each instance having power supply and power demand at time step t. Another set of entanglement patterns is created to specify various quantum connectivity structures. The model also has a reward history and convergence time histories to monitor learning.
Each epoch begins by adapting the learning rate adaptively, ensuring stability in training using a decay function given in Equation (18), where is the decay factor. The algorithm integrates QAE and QMC for enhanced performance. Below are the key mathematical formulations added to the algorithm:
Instead of using standard optimization techniques for VQE, we now use QAE to find optimal angles ((Equations (7) and (8)):
- QAE provides a quadratic speedup, reducing the number of function evaluations required to reach optimal parameters.
- The convergence time is reduced due to fewer required sampling steps.
| Algorithm 1 Quantum-inspired multi-agent reinforcement learning for energy optimization |
|
Here, Z_i represents the Pauli-Z measurements on the employed qubits, and the parameters w_i and v_ij are calculated via training.
Based on a smoothed reward function, an entanglement pattern is dynamically selected to influence quantum state interactions (Equation (5)). During training, for each data sample , a quantum circuit is prepared by applying Hadamard gates (Equation (3)). Entanglement are applied based on the selected pattern in Equation (6).
Classical energy data are encoded into quantum states using parameterized rotation gates as mentioned in Equation (9). After encoding, parameterized rotation gates are applied (Equation (10)). Quantum Monte Carlo is used instead of classical Monte Carlo to compute expected values of the quantum states (Equation (11)):
- QMC enhances the efficiency of expectation value calculations.
- It achieves a quadratic speedup in sampling compared to classical Monte Carlo.
A local reward function evaluates the success of the allocation by punishing major mismatches between demand and supply while rewarding accurate predictions as given by Equation (12). After processing all training samples, the total reward for the epoch is updated, and execution times for each agent are recorded as represented in Equation (14). The global reward is now computed using QAE for better aggregation (Equation (15)):
- QAE enhances the accuracy of global reward estimation.
- It ensures better convergence with a smaller number of training epochs.
The quantum agents update their policy parameters in parallel, leveraging quantum gradients (Equations (20)–(22)):
- Parallel updates improve computational efficiency, reducing per-epoch execution time.
- Quantum gradients enhance the precision of parameter updates, leading to faster convergence.
Lastly, the best policy and cumulative rewards are returned as output, offering a high-performance quantum-inspired reinforcement learning platform for energy optimization. The technique exploits quantum entanglement, variational quantum optimization, and learning for adaptive improvements in decision-making in multi-agent energy management situations. Table 1 represent the parameter notation table for the employed QiMARL algorithm.
Table 1.
Parameter definitions for quantum-inspired multi-agent reinforcement learning algorithm.
4.2. Propositions and Proofs
Proposition 1.
(Convergence of Adaptive Learning Rate) Let the adaptive learning rate be defined as:
where and ; we have:
Proof.
Since and , we examine the limit:
Since the denominator becomes infinitely large, the fraction approaches zero:
Hence, as , establishing the proposition. □
Proposition 2.
(Convergence of Smoothed Reward Function)Assuming the following smoothed reward function:
where , the reward function is exponentially smoothed to a finite value.
Proof.
Iteratively expanding the recurrence relation:
Since , the exponent goes to zero exponentially. Assuming that the rewards are bounded, the sum has a finite limit. Therefore, the smoothed reward function reaches a stable point. □
Proposition 3.
(Optimal Quantum Energy Allocation) The optimal quantum energy allocation that optimizes the local reward function:
happens when .
Proof.
Examine the following absolute deviation:
The penalty term is nonzero if . To reduce the total reward penalty, we choose , resulting in the optimal allocation. □
Proposition 4.
(Quantum Entanglement Selection Stability) The entanglement selection mechanism
converges to a stable entanglement pattern if the rewards have a non-decreasing trend over time.
Proof.
If the reward function is increasing or stabilizing, then
By the entanglement selection rule, this implies:
Therefore, the entanglement pattern does not change with time, resulting in a stable quantum entanglement state. □
4.3. Time Complexity Analysis
The provided algorithm has several nested loops and quantum operations. We examine each of the major parts to find the overall time complexity.
Outer Loop (Epochs): The outer loop executes for epochs:
Inner Loop Training Data: For every epoch, the algorithm loops over the training dataset of size N:
Quantum Circuit Preparation: For every training data point, the algorithm prepares a quantum circuit:
- Hadamard gates: ;
- Entanglement gates (CNOTs): ;
- Rotation gates: .
Hence, the preparation of quantum circuits takes
per iteration.
Quantum Measurement and Computation: The algorithm measures the quantum state and calculates energy allocation:
Expectation values and weighted sum calculation take
per iteration.
Local Reward Computation: The computation of the reward function includes absolute differences and conditional checks, which require
operations per step.
Global Reward and Convergence Time Updates: Updating global rewards and execution times involves summing over all agents :
Total Time Complexity: Adding up all the parts, we have:
If , then
The algorithm scales quadratically with the number of qubits (). It scales linearly with the dataset size (N) and the number of epochs (). The number M of agents counts, but if small compared to , it does not control the complexity.
Thus, the worst-case time complexity is:
QAE for VQE Angles: QAE gains an estimation accuracy speedup, lowering complexity from regular sampling’s to , resulting in quicker convergence.
Parallel Quantum Batch Processing: Rather than processing data sequentially, batches of training data are processed in parallel quantum circuits. Parallelism lowers the per-epoch processing time, altering the complexity from to roughly , where B is the number of parallel batches.
QMC for Energy Allocation: QMC is employed in place of traditional expectation value calculations to estimate energy allocations. QMC achieves a quadratic speedup, bringing the computation of expectation values from to .
QAE-Based Global Reward Computation: Rather than summing rewards directly, a QAE-based aggregation process is proposed, where QAE lowers the number of samples needed to accurately estimate rewards, enhancing computational efficiency.
Parallel Quantum Policy Updates: Policy updates for all the quantum agents now take place in parallel rather than sequentially. Per-agent update complexity is still the same, but the total update time is decreased by a factor that is a function of the number of available quantum processors.
The Final Time Complexity:
After the changes, with the addition of parallel processing and QMC:
If B (batch size for parallel processing) is sufficiently large, complexity is effectively diminished in realistic implementations.
So the major advances arise due to quantum parallelism and quantum-enabled estimation techniques, which reduce the per-epoch computational cost and enhance scalability.
4.4. Scalability Analysis in Larger Systems
As the number of power generation stations () and demand stations () grows, the computational complexity of classical and quantum methods increases differently. The current section discusses the scalability benefit of the proposed quantum-inspired multi-agent reinforcement learning (QiMARL) method.
4.4.1. Growth of the Quantum System
In QMARL, the number of qubits required is given by
As the system scales, the quantum circuit size increases quadratically with respect to the number of stations.
The quantum circuit preparation involves the following:
- Hadamard and rotation gates: .
- Entanglement operations: .
- Expectation value computations: due to parallelized computation.
Thus, the overall computational complexity of QiMARL is given by
Despite its complexity, QiMARL uses quantum parallelism via QAE, allowing it to handle tenfold bigger state–action spaces than conventional techniques.
4.4.2. Growth of Classical Approaches
For classical MAB methods, the complexity depends on the number of agents, , which is typically proportional to (i.e., ):
- Greedy MAB:
- UCB MAB:
- Thompson Sampling:
In larger systems, traditional approaches have high computational overhead:
- Greedy MAB suffers from poor power allocation because of insufficient exploration.
- UCB MAB becomes slower with the logarithmic scaling factor.
- Thompson Sampling becomes infeasible because it has quadratic dependency on .
For classical Multi-Agent Reinforcement Learning (MARL) methods, the computational complexity depends on the number of agents , the state space size , and the action space size . In cooperative multi-agent energy allocation problems, is typically proportional to (i.e., ).
- QMIX:QMIX uses a centralized mixing network with agent-specific Q-networks:Complexity grows linearly with but also scales with the full state–action representation for each agent.
- MADDPG:MADDPG employs centralized critics with decentralized actors:The joint state–action dependency leads to exponential growth in complexity with the number of agents, making it impractical for large .
- PPO (Multi-Agent variant):PPO applies policy-gradient updates for each agent:where K is the number of gradient steps per epoch. PPO scales linearly with but has a high constant factor due to multiple policy evaluations and clipping operations.
Scalability implications:
- MADDPG becomes infeasible for large systems because of exponential scaling.
- QMIX is more scalable than MADDPG but still suffers when dealing with a very large due to its full per-agent Q-function evaluation.
- PPO scales better than MADDPG in theory but is slower in practice because of repeated policy optimization steps.
4.4.3. Justification for QiMARL’s Advantage in Large Systems
As and increase:
- Classical MAB methods become less efficient since they grow linearly or quadratically in the power stations.
- QiMARL exploits quantum features to deal with high-dimensional spaces efficiently:
- Quantum states represent exponentially bigger state–action spaces.
- Entanglement supports correlated decision-making, enhancing energy distribution.
- Empirical results (Table 2) indicate that on large networks (, , ), QiMARL is superior in terms of reward optimization and convergence rate to classical methods.
Table 2. Comparison of scalability for larger systems.
Though classical methods are perhaps more efficient on small scales, they are not scalable because they have a linear or quadratic complexity in the number of power stations. On the other hand, QiMARL uses quantum parallelism to search and optimize large-scale power networks effectively. Thus, it is a better option for energy optimization in high-dimensional quantum settings.
5. Implementation Environment
The QiMARL algorithm was executed using IBM’s Qiskit 0.39.0 software (Python-based), which enables quantum circuit simulation on conventional hardware. All quantum operations—including initialization of quantum states, Hadamard and CNOT gates, parameterized rotations, and quantum measurements—were carried out using the high-performance simulator backend Qiskit Aer. Quantum hardware was not utilized in the experiments. The VQE, QAE, and QMC modules were implemented using custom routines written with NumPy and Qiskit’s circuit module, providing full control over the parameter update dynamics.
Although other libraries such as PennyLane and Cirq offer similar functionality, Qiskit was chosen for its robust quantum circuit visualization, support for quantum chemistry modules (to handle VQE-style Hamiltonians), and compatibility with IBM Quantum Experience for potential future execution on real hardware. The classical policy training loops, gradient descent updates, and reward computations were performed using Python 3.9 with NumPy, with matrix operations optimized using Numba where applicable. All experiments were conducted on a standard workstation equipped with an Intel Core i7 CPU, 32 GB RAM, and running Ubuntu 20.04 LTS.
6. Simulation Results and Discussion
6.1. Simulation Setup
We analyzed the performance of Quantum Multi-Agent Reinforcement Learning (QiMARL) compared to classical Multi-Armed Bandit (MAB) algorithms (Greedy, UCB, Thompson Sampling). The number of qubits for real quantum systems is given by
where and are the number of generating and demanding power stations. Real quantum systems can support up to 1000 qubits, which was applicable to our implementation. But since we were using classical implementation strategies, for simulation purposes, this would one qubit to each generator–demand pair. While this one-hot style encoding is conceptually simple, it scales linearly with the number of pairs and becomes intractable for large systems when classically simulated using the Qiskit Aer backend. For instance, a configuration of and yields qubits, requiring storage of complex amplitudes for exact state-vector simulation—beyond any feasible classical memory resources.
To make large-scale simulations implementable on standard workstations, we adopted a basis (index) encoding strategy. Instead of allocating a separate qubit for each generator–demand pair, we represented the set of all combinations as the computational basis states of a smaller qubit register. Each unique pair , where and , was mapped to a unique integer index
with . The required number of qubits was therefore
For the example with and , we have combinations, which require
qubits to represent all possible states. The nine-qubit register can store any one of the 512 computational basis states , with only the first 500 corresponding to valid generator–demand pairs.
State Preparation: Initial superposition over all valid combinations was generated via Hadamard gates on the qubits, followed by amplitude masking or rejection sampling to exclude unused basis states (indices ). Parameterized rotations and entanglement gates were then applied to this compact register to implement the quantum-inspired policy evolution.
Measurement and Decoding: After measurement in the computational basis, the outcome bit string was decoded back to using
This mapping enabled all post-processing steps (reward computation, convergence analysis, and energy allocation estimation) to remain unchanged from the one-hot formulation.
This encoding reduced the qubit count from to (without epoch inclusion), making simulations of large-scale systems (e.g., ) tractable on classical hardware while preserving the representational completeness of the quantum state space. Entanglement and observable definitions were adapted to act on the index register, ensuring that the correlations between specific generators and demands were still captured through appropriately constructed multi-qubit operators.
Figure 1 represents the quantum circuit with rotational angles initialized from references [,,,] for a nine-qubit multi-agent system. All quantum circuit computations in this work were classically simulated using IBM’s Qiskit Aer simulator. No quantum hardware was utilized. Quantum gates, entanglement operations, and amplitude estimations were therefore realized using classical numerical simulations. As a result, QiMARL did not exhibit genuine quantum parallelism but a simulated one based on quantum mechanics for improved computational approaches.
Figure 1.
The initialized quantum circuit design for an experiment with a nine-qubit multi-agent system.
For the comparative evaluation, we used classical MAB algorithms (Greedy, UCB, and Thompson Sampling) as baseline models. All three strategies were employed agent-wise for the power allocation problem, in which each demand or generating station acts independently in taking actions. These algorithms are simple, scalable approaches commonly applied in adaptive control problems and deployed here to serve as benchmarks on computational and convergence efficiency of QiMARL. Their presence highlights QiMARL’s ability to surpass well-known low-complexity decision algorithms, especially in the high-dimensional scenario.
It is worth mentioning that MAB algorithms are not full MARL models because they do not model inter-agent coordination or shared state spaces. That is why MARL algorithms like MADDPG, QMIX, PPO are fair to use for comparison purposes
6.2. Experiment Results with the MAB Algorithms
6.2.1. Comparison of Convergence Time
Table 3 shows a comparative study of the convergence time of various decision-making algorithms for small and large systems. QiMARL was compared with three traditional MAB strategies: Greedy MAB, UCB MAB, and Thompson Sampling. The most important parameters studied were the number of epochs taken to converge, time taken per epoch, and overall computational time for both system sizes.
Table 3.
Convergence time comparison for small and large systems for MAB algorithms.
In the small-scale system (10 generating stations and 10 demanding stations), the QiMARL algorithm had a higher convergence rate, with 150 epochs taking 0.85 s per epoch and thus a total of 127.5 s. This was considerably less than Greedy MAB (200 epochs, 90 s total), UCB MAB (180 epochs, 90 s total), and Thompson Sampling (160 epochs, 96 s total). The quantum-inspired method had the advantage of its partial parallelized search and decision-making, enabling faster learning in fewer epochs.
As the system increased in size to a larger network (50 generator stations and 50 demand stations), the difference in performance between QiMARL and traditional methods was more pronounced. While QiMARL took 250 epochs to converge, time per epoch became slightly larger at 1.20 s, for a total computational time of 300 s. This was faster than the traditional algorithms, especially Greedy MAB, which took 450 epochs and 315 s to converge. The UCB MAB (350 epochs, 262.5 s) and Thompson Sampling (300 epochs, 255 s) also converged slower than QiMARL.
The trend observed underlines the scalability benefit of QiMARL in high-dimensional, complicated environments. Classic MAB methods fare poorly with the rise in system size because of sequential search and use of deterministic or stochastic heuristics. In contrast, the QiMARL framework has an advantage with its quantum parallelism and probabilistic superposition in searching and optimizing policies with big state space. While the computation time per epoch takes a small increment because of the increased quantum state space and the extra computational burden, the decrease in the number of epochs covers this, yielding an soverall computational benefit in large-scale environments.
In summary, the data in Table 3 indicate that although traditional MAB algorithms are acceptable for small-scale systems, they are not scalable to larger systems. QiMARL beats traditional methods on both convergence rate and overall computation time and hence is a promising candidate for solving large-scale energy allocation and decision-making issues in future smart grids and power distribution networks.
6.2.2. Reward Performance
Table 4 illustrates an extensive comparison of the reward performance of various decision-making algorithms in small and large systems. The table compares local, global, and overall reward values for the QiMARL algorithm and classical MAB strategies, such as Greedy MAB, UCB MAB, and Thompson Sampling. The outcomes identify the excellent performance of QiMARL across all parameters, particularly in large systems.
Table 4.
Reward comparison for small and large systems for MAB algorithms.
For the small system case (10 generating stations and 10 demanding stations), QiMARL recorded the maximum rewards, with a local reward of 0.92, a global reward of 0.88, and an overall reward of 0.90. These represent efficient decisions as well as optimal allocation of energy compared to the traditional methods. Among the traditional methods, Thompson Sampling worked best with a total reward of 0.82, followed by UCB MAB at 0.78 and Greedy MAB at 0.75. The comparatively poor performance of Greedy MAB was anticipated since its exploitative behavior leads it to converge to non-optimal choices, restricting its ability to adapt to changing environments.
When the system was increased in size to a larger configuration (50 generating stations and 50 demand stations), the performance difference between QiMARL and the conventional techniques became apparent. QiMARL still had high reward values, with a local reward of 0.95, a global reward of 0.91, and an overall reward of 0.93, indicating higher optimization compared to the case with a small system. Conventional techniques were hit by performance loss. The total reward of Greedy MAB fell to 0.70 from 0.75, showing its poor scalability. UCB MAB and Thompson Sampling also saw minor dips, with total rewards falling to 0.76 and 0.80, respectively. The fall in reward performance of traditional methods shows their poor efficiency in dealing with large and complex decision spaces, where it becomes difficult to balance exploration and exploitation for the optimization task.
The scalability pattern from Table 4 highlights the inherent strength of QMARL over traditional methods. With an increasing number of power stations, QMARL is aided by quantum parallelism and entanglement to process a larger state space more efficiently. Traditional methods are bogged down by computational inefficiencies and decision-making inefficiencies, which results in suboptimal overall reward performance for large systems. This indicates that although classical reinforcement learning methods can suffice for smaller problems, they do not possess the computation and flexibility capabilities needed to support large-scale multi-agent coordination.
Ultimately, the results directly illustrate the comparative scalability advantage of QiMARL in more elaborate decision-making landscapes. Though standard methods are relatively successful for moderate-size systems, their failure in larger implementations proves to be much more problematic given the enhanced urgency of effective exploration, adaptability, and computing efficiency requirements with increased scale. The ability of QiMARL to maintain high-value rewards for varied system sizes demonstrates its feasibility as a superior means to deal with large-scale energy distribution and optimized smart grid problematics of the quantum sort.
6.3. Experiment Results with the MARL Algorithms
6.3.1. Reward Performance (Figure 2)
The highest overall reward of 0.90 was achieved by QiMARL in the small-scale environment, with an overt performance difference compared to the classical MARL baselines. Specifically, QMIX achieved 0.86 (−4.6%), MADDPG attained 0.84 (−7.1%), and PPO obtained 0.80 (−12.5%). The advantage was not restricted to the cumulative reward metric but was observed over both local rewards (agent-level optimization) and global rewards (system-level performance). QiMARL’s global reward of 0.88 informs us that individual decision-making processes were optimally tuned, while its local reward of 0.92 indicates that the individual strategies collectively created coordinated global behavior without the need for explicit communication channels. It is significant that in decentralized settings—such as smart-grid energy balancing—global coordination typically deteriorates as the size of the agents increases or as communication bandwidth is limited.
The local–global balance witnessed in small systems comes from the quantum-inspired entanglement mechanism of QiMARL, which inherently couples agent policies. Unlike shared reward signals or centralized critics (used in CTDE), QiMARL uses simulated quantum states to encourage diversity in exploration without moving away from a common optimization goal. This is a critical difference from baselines like MADDPG and PPO, which tend to suffer from reward fragmentation from the lack of an explicit coordination channel.
When scaling to a large system configuration, the superiority of QiMARL was even more evident. It achieved a cumulative reward of 0.93, outperforming QMIX (0.88, −5.7%), MADDPG (0.85, −9.4%), and PPO (0.82, −13.4%). Interestingly, QiMARL maintained a local reward of 0.95 and a global reward of 0.91—both of which were higher than in the small-system case. This stability in global reward is particularly noteworthy, as traditional MARL baselines were grossly impaired in their coordination quality when they transitioned from 20 agents to 100 agents. For example, MADDPG’s global reward decreased from 0.81 in the small system to 0.83 in the large system, exhibiting the law of diminishing returns in cooperative settings as the number of agents grew.
Figure 2.
Reward comparison for small and large systems for MARL algorithms.
QiMARL’s scalability of reward performance is the immediate consequence of two of its most significant design choices:
- Amplitude-based policy sampling accelerates the exploration of probable actions in a high-dimensional action space.
- Entanglement-motivated discovery, which averts policy diversity from becoming purely autonomous action, so that global coherence is retained.
In practice, this leads to greater efficiency in large-scale distributed energy networks with QiMARL without requiring full-state observability or costly inter-agent communication.
A paired t-test over 10 independent runs confirmed that the reward improvements of QiMARL over QMIX were statistically significant at and over MADDPG and PPO at . The effect sizes (Cohen’s ) indicated that these were not small or marginal gains but good-sized and practically relevant ones. From a deployment point of view, greater global rewards for big systems directly translate into better system-wide stability, less operational loss, and increased resilience in real-time energy management scenarios.
6.3.2. Comparison of Convergence Time (Figure 3)
Training efficiency is a critical parameter that MARL algorithms must take into account, particularly in environments where policy improvements must be computed efficiently in response to varying situations. In the small-system experiments, QiMARL converged at epoch 150, while QMIX converged at epoch 200, MADDPG at epoch 220, and PPO at epoch 240. This translates to a decrease in training iterations ranging from 25–37.5%. With an average epoch time of 0.85 s, QiMARL’s total training time was 127.5 s, 30% less than QMIX (184.0 s), 39% less than MADDPG (209.0 s), and 41% less than PPO (216.0 s).
This efficiency is made possible by QiMARL’s parallelized state–action estimation enabled by its quantum-inspired simulation. With simultaneous estimation of multiple potential action sequences (such as quantum superposition), QiMARL does not squander computational cycles on low-probability policies. Its exploration strategy uses amplitude reductions and redundant policy updates, which are typically the downfall of standard MARL methods.
The benefits of QiMARL’s architecture were much apparent during the large-system experiments. QiMARL took 250 epochs to converge, whereas QMIX took 340, MADDPG took 370, and PPO took 400 epochs to converge. The total training time for QiMARL lasted 300.0 s, which was 37% lower than that of QMIX (476.0 s), 44% lower than that of MADDPG (536.5 s), and 44.4% lower than that of PPO (540.0 s). These results demonstrate that the same quantum-inspired computations responsible for speedup in small contexts can similarly scale up to much larger agent populations.
Classical approaches generally require exponentially more iterations with growing numbers of agents for high-dimensional collaborative problems, since the joint action space combinatorially explodes. QiMARL discourages this expansion by probabilistic amplitude coding, whereby the system is able to concentrate on high-value policy updates and can disregard large regions of the action space that are unlikely to deliver any improvements. This concentrated search process not only accelerates the training but also reduces the computational energy it requires, which becomes relevant in green AI research.
Convergence rate differences were significantly different in all the comparisons with and effect sizes greater than 0.8. Operationally, the reduction in training time means that QiMARL can be retrained or updated to new operating conditions within hours, compared to days for large-scale deployments. This is highly significant for applications like smart grids, where the environment can quickly change due to varying demand, availability of renewable power, or system malfunctions.
Figure 3.
Convergence time comparison for small and large systems for MARL algorithms.
6.4. Ablation Study (Table 5)
The ablation experiments were designed to quantify the independent effects of QiMARL’s two key innovations—entanglement-inspired coordination and amplitude-based policy sampling—on both reward performance and convergence speed. By excluding each component individually and keeping the remaining architecture and training conditions unchanged, we could tell how much of QiMARL’s performance improvement could be attributed directly to each mechanism.
Table 5.
Ablation study results for the large-scale system.
Table 5.
Ablation study results for the large-scale system.
| Variant | Local Reward | Global Reward | Overall Reward | Epochs to Converge |
|---|---|---|---|---|
| Full QiMARL | 0.95 | 0.91 | 0.93 | 250 |
| – No Entanglement | 0.89 | 0.85 | 0.87 | 310 |
| – No Amplitude Sampling | 0.91 | 0.87 | 0.89 | 290 |
| Classical Best (QMIX) | 0.88 | 0.86 | 0.88 | 340 |
6.4.1. Impact of Entanglement-Inspired Coordination
Disabling the entanglement mechanism always had the worst performance effect in both small and large systems. In the small system (10GR, 10DM), the cumulative reward went down from 0.90 to 0.85 (−5.6%), whereas convergence was delayed by 30 epochs. In the large system (50GR, 50DM), the effect was even more dramatic: the cumulative reward went down from 0.93 to 0.87 (−6.5%), and convergence was delayed by 60 epochs. This implies that entanglement is particularly important in scaling coordination to larger groups of agents, where the absence of explicit communications would otherwise cause agents to settle to locally optimal but globally suboptimal strategies. The gap relative to the best classical algorithm (QMIX) was also reduced when entanglement was removed, which suggests that this feature is an important driver of QiMARL’s new supremacy over vanilla MARL baselines.
6.4.2. Impact of Amplitude-Based Policy Sampling
Removal of amplitude-based policy sampling created a less pronounced but nonetheless considerable decline in performance. In the small system, the overall reward dropped from 0.90 to 0.86, and convergence was delayed by 20 epochs. In the large system, overall reward dropped from 0.93 to 0.89, and convergence slowed by 40 epochs. While less destructive in terms of annihilating entanglement, the slowdown in convergence is noteworthy, corroborating that sampling based on amplitude is largely beneficial for training efficiency. By probabilistic policy updates weighted towards likely action strings, the mechanism effectively prunes low-value exploration paths, accelerating the learning process without sacrificing solution quality.
6.4.3. Synergy Among Components
The primary finding of this work is that entanglement and amplitude-based sampling together are superior to either alone. Entanglement induces structural correlation among agents’ policies so that local learning is synchronized with global objectives, while amplitude-based sampling speeds up learning about high-reward policies. Without amplitude-based sampling, agents fragment into less correlated actions; without entanglement, learning takes longer and at greater expense. When both are present, QiMARL performs well in terms of reward, converges quickly in both small and large systems, and is superior to all the classical baselines tested.
6.4.4. Scalability Implications
The difference in severity of degradation in performance between small and large systems also yields important scalability insights. For smaller populations of agents, both mechanisms missing are also partially made up for by lower coordination complexity. However, in larger systems, the combinatorial explosion of possible joint policies increases the disadvantages of removing them. This is in favor of the argument that design choices made in QiMARL are not only incremental improvements but structural necessities for maintaining efficiency in large-scale, decentralized decision environments such as national-scale smart grids.
Though QiMARL uses quantum computing jargon and mechanics, it is a quantum-inspired model run on classical computers. The advantages seen in convergence and performance are a result of classical simulations of quantum characteristics rather than genuine quantum speedups. Future work may consider running on actual quantum hardware, but current results are all from classical simulation environments.
6.5. Experimental Results with Intermediate Nodes
6.5.1. Comparison of Convergence Time
To better understand the efficacy of the QiMARL algorithm, we carried out experimentation on intermediate nodes to unfold any hidden trends. The convergence trend for the medium system sizes—Medium-1 (20 GR, 20 DM) and Medium-2 (35 GR, 35 DM)—shows the evident dominance of the new QiMARL model over both traditional MAB methods and usual MARL methodologies. As can be observed from Table 6, QiMARL required the lowest number of epochs to converge for both of the two configurations, just 180 epochs for Medium-1 and 215 epochs for Medium-2. This was considerably faster than the optimal-performing MAB baseline (Thompson Sampling: 200 and 260 epochs, respectively) and much better than MARL methods such as MADDPG and PPO, which converged after 265–350 epochs.
Table 6.
Convergence time and computational cost for Medium-1 (20 GR, 20 DM) and Medium-2 (35 GR, 35 DM) systems across QiMARL, traditional MAB, and MARL algorithms.
In addition, QiMARL maintained competitive per-epoch computation times at 0.95 s for Medium-1 and 1.05 s for Medium-2 even after introducing quantum-inspired processing overhead. This was lower than all but the majority of MARL methods, where increased policy complexity and communication overhead pushed per-epoch times over 1.1 s for MADDPG and QMIX in larger environments. The total convergence time for QiMARL (171 s for Medium-1, 225.8 s for Medium-2) was much lower compared to MARL baselines, indicating its computational and learning efficiency. These results reveal the ability of quantum-inspired agent coordination to speed up convergence without incurring computationally unaffordable costs as the number of GR and DM grows.
6.5.2. Comparison of Reward Performance
Reward-based performance measures, reported in Table 7, also confirm the advantages of the QiMARL approach. For both intermediate system sizes, QiMARL achieved the maximum local, global, and total reward values. Specifically, for the Medium-1 situation, QiMARL obtained a local reward of 0.93, a global reward of 0.89, and a total reward of 0.91. These values outperformed the top MARL competitor (QMIX: global reward 0.87) and all the MAB baselines, the closest of which was Thompson Sampling (global reward 0.82).
Table 7.
Reward performance for Medium-1 (20 GR, 20 DM) and Medium-2 (35 GR, 35 DM) systems across QiMARL, traditional MAB, and MARL algorithms.
In the more difficult Medium-2 configuration, QiMARL continued to excel with a local reward of 0.94, a global reward of 0.90, and a total reward of 0.92. Performance degradation in Medium-1 to Medium-2 scenarios was modest, indicating good scalability and robustness under increased numbers of agents. In comparison, baselines for both image-based MAB and MARL experienced steeper drops in reward performance with increasing system size. For instance, Thompson Sampling dropped from 0.82 to 0.81 in cumulative reward, and QMIX dropped from 0.87 to approximately 0.875. The sustained superiority in both local and global reward metrics demonstrates that QiMARL converges faster and produces better coordination and allocation policies, leading to tangible gains in both system-level and agent-level objectives.
7. Limitations of the Proposed QiMARL Algorithm
- Quantum Computation Overhead in Small Systems: On small problems, classical MAB algorithms (Greedy, UCB, Thompson Sampling) would typically be better because QiMARL adds extra quantum computation overhead.
- Hardware Limitations: Existing quantum processors are plagued by noise, decoherence, and qubit connectivity limitations, which have the potential to degrade the performance of QiMARL implementations.
- Implementation Complexity: Implementation of quantum circuits for reinforcement learning involves understanding quantum computing, hence making QiMARL more complicated to implement than traditional methods.
- Quantum Measurement Restrictions: Measurement of quantum states collapses the superposition, restricting the number of observations within an iteration and, in effect, causing information loss.
- Hybrid Quantum–Classical Bottlenecks: As practical application involves projection of quantum models onto classical data processing, hybrid execution can suffer from bottlenecks, degrading real-time behavior.
8. Conclusions and Future Research Avenues
The proposed QiMARL algorithm offers a viable option for improving energy distribution in nodal power systems. Using sophisticated quantum concepts, the program successfully balances the exploration and exploitation processes, allowing for a dynamic and adaptable energy distribution system in real time. Following numerous iterative rounds, experimental findings showed that the algorithm may increase important performance indicators such as energy distribution efficiency, distribution reliability, and total rewards. As a result, the quantum computing-based QiMARL algorithm demonstrates tremendous promise in tackling contemporary energy distribution difficulties, particularly in circumstances characterized by high unpredictability and uncertainty.
Future research on the QiMARL algorithm could focus on expanding the scalability of the algorithm, making it relevant for implementing it in grids with high complexities, and also working on its performance in higher-dimensional state spaces. As proposed, one of the solutions could be Federated-QiMARL. As quantum-based advances evolve, it is very much possible to implement the QiMARL algorithm in quantum-based devices to understand its efficacy in real-world configurations. Furthermore, research can focus on further coupling the advantages of the classical and quantum paradigms and implementing systems that would incorporate aspects like supply chain administration and financial modeling. These advancements would further improve the functionality and relevance of the proposed QiMARL algorithm, making it potent for solving complex optimization issues and carving the way forward for industrial usage.
Author Contributions
S.M.T. worked on conceptualizations, methodology, formal analysis, and writing—original draft preparation, while A.B., M.S.K. and K.R. participated in writing—review and editing, validation, supervision, project administration. All authors have read and agreed to the published version of the manuscript.
Funding
This research received no external funding.
Conflicts of Interest
The authors declare no competing interests.
References
- Pavon, W.; Jaramillo, M.; Vasquez, J. A Review of Modern Computational Techniques and Their Role in Power System Stability and Control. Energies 2023, 17, 177. [Google Scholar] [CrossRef]
- An Analytical Formulation for Mapping the Spatial Distribution of Nodal Inertia in Power Systems. IEEE Access 2023, 11, 45364–45376. [CrossRef]
- Duman, S.; Kahraman, H.T.; İsmail Kati, M. Economical operation of modern power grids incorporating uncertainties of renewable energy sources and load demand using the adaptive fitness-distance balance-based stochastic fractal search algorithm. Eng. Appl. Artif. Intell. 2023, 117, 105501. [Google Scholar] [CrossRef]
- Saggio, V.; Asenbeck, B.E.; Hamann, A.; Strömberg, T.; Schiansky, P.; Dunjko, V.; Friis, N.; Harris, N.C.; Hochberg, M.; Englund, D.; et al. Experimental quantum speed-up in reinforcement learning agents. Nature 2021, 591, 229–233. [Google Scholar] [CrossRef]
- Gil-Fuster, E.; Eisert, J.; Bravo-Prieto, C. Understanding quantum machine learning also requires rethinking generalization. Nat. Commun. 2024, 15, 2277. [Google Scholar] [CrossRef]
- Biamonte, J.; Wittek, P.; Pancotti, N.; Rebentrost, P.; Wiebe, N.; Lloyd, S. Quantum machine learning. Nature 2017, 549, 195–202. [Google Scholar] [CrossRef]
- Preskill, J. Quantum computing in the NISQ era and beyond. Quantum 2018, 2, 79. [Google Scholar] [CrossRef]
- Cerezo, M.; Arrasmith, A.; Babbush, R.; Benjamin, S.C.; Endo, S.; Fujii, K.; McClean, J.R.; Mitarai, K.; Yuan, X.; Cincio, L.; et al. Variational quantum algorithms. Nat. Rev. Phys. 2021, 3, 625–644. [Google Scholar] [CrossRef]
- Bharti, K.; Cervera-Lierta, A.; Kyaw, T.H.; Haug, T.; Alperin-Lea, S.; Anand, A.; Degroote, M.; Heimonen, H.; Kottmann, J.S.; Menke, T.; et al. Noisy intermediate-scale quantum algorithms. Rev. Mod. Phys. 2022, 94, 015004. [Google Scholar] [CrossRef]
- Dunjko, V.; Taylor, J.M.; Briegel, H.J. Quantum-Enhanced Machine Learning. Phys. Rev. Lett. 2016, 117, 130501. [Google Scholar] [CrossRef]
- Elshaari, A.W.; Pernice, W.; Srinivasan, K.; Benson, O.; Zwiller, V. Hybrid integrated quantum photonic circuits. Nat. Photonics 2020, 14, 285–298. [Google Scholar] [CrossRef]
- Shojaee, G.K.; Mashhadi, H.R. Optimistic initial value analysis in a greedy selection approach to MAB problems. In Proceedings of the 2017 7th International Conference on Computer and Knowledge Engineering (ICCKE), Mashhad, Iran, 26–27 October 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 419–424. [Google Scholar]
- Busa-Fekete, R.; Kégl, B. Accelerating adaboost using UCB. In Proceedings of the KDD-Cup 2009 Competition, PMLR, New York, NY, USA, 28 June 2009; pp. 111–122. [Google Scholar]
- Gupta, N.; Granmo, O.C.; Agrawala, A. Thompson sampling for dynamic multi-armed bandits. In Proceedings of the 2011 10th International Conference on Machine Learning and Applications and Workshops, Honolulu, HI, USA, 18–21 December 2011; IEEE: Piscataway, NJ, USA, 2011; Volume 1, pp. 484–489. [Google Scholar]
- Buşoniu, L.; Babuška, R.; De Schutter, B. Multi-agent reinforcement learning: An overview. In Innovations in Multi-Agent Systems and Applications-1; Springer: Berlin/Heidelberg, Germany, 2010; pp. 183–221. [Google Scholar]
- Silver, D.; Lever, G.; Heess, N.; Degris, T.; Wierstra, D.; Riedmiller, M. Deterministic policy gradient algorithms. In Proceedings of the International Conference on Machine Learning, Pmlr, Beijing, China, 21–26 June 2014; pp. 387–395. [Google Scholar]
- Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, UK, 1998; Volume 1. [Google Scholar]
- Sghaier, W.B.; Gassara, H.; Rouissi, F.; Tlili, F. Adaptive UCB MAB Algorithm forOptimizing Relay Selection in Narrowband Power Line Communication. In Proceedings of the 2023 IEEE Tenth International Conference on Communications and Networking (ComNet), Hammamet, Tunisia, 1–3 November 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 1–6. [Google Scholar]
- Foerster, J.; Farquhar, G.; Afouras, T.; Nardelli, N.; Whiteson, S. Counterfactual Multi-Agent Policy Gradients. In Proceedings of the AAAI Conference on Artificial Intelligence, Orleans, LA, USA, 2–7 February 2018; Volume 32. [Google Scholar] [CrossRef]
- Rashid, T.; Samvelyan, M.; Schroeder, C.; Farquhar, G.; Foerster, J.; Whiteson, S. QMIX: Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning. In Proceedings of the 35th International Conference on Machine Learning, PMLR, Stockholm, Sweden, 10–15 July 2018; Dy, J., Krause, A., Eds.; 2018; Volume 80, pp. 4295–4304. [Google Scholar]
- Lowe, R.; WU, Y.; Tamar, A.; Harb, J.; Pieter Abbeel, O.; Mordatch, I. Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments. In Advances in Neural Information Processing Systems; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
- Sunehag, P.; Lever, G.; Gruslys, A.; Czarnecki, W.M.; Zambaldi, V.; Jaderberg, M.; Lanctot, M.; Sonnerat, N.; Leibo, J.Z.; Tuyls, K.; et al. Value-Decomposition Networks For Cooperative Multi-Agent Learning. arXiv 2017, arXiv:cs.AI/1706.05296. [Google Scholar] [CrossRef]
- Du, W.; Ding, S.; Zhang, C.; Shi, Z. Multiagent reinforcement learning with heterogeneous graph attention network. IEEE Trans. Neural Netw. Learn. Syst. 2022, 34, 6851–6860. [Google Scholar] [CrossRef]
- Liu, Z.; Zhang, J.; Shi, E.; Liu, Z.; Niyato, D.; Ai, B.; Shen, X.S. Graph neural network meets multi-agent reinforcement learning: Fundamentals, applications, and future directions. IEEE Wirel. Commun. 2024, 31, 39–47. [Google Scholar] [CrossRef]
- Dunjko, V.; Friis, N.; Briegel, H.J. Quantum-enhanced deliberation of learning agents using trapped ions. New J. Phys. 2015, 17, 023006. [Google Scholar] [CrossRef]
- Park, S.; Kim, J.P.; Park, C.; Jung, S.; Kim, J. Quantum multi-agent reinforcement learning for autonomous mobility cooperation. IEEE Commun. Mag. 2023, 62, 106–112. [Google Scholar] [CrossRef]
- Abbas, A.; Sutter, D.; Zoufal, C.; Lucchi, A.; Figalli, A.; Woerner, S. The power of quantum neural networks. Nat. Comput. Sci. 2021, 1, 403–409. [Google Scholar] [CrossRef]
- Manasa, M.; Tlajiya, F.; Chavan, S. Optimizing Energy Management and Load Balancing Through AI-Driven Quantum Approximate Optimization. In Proceedings of the 2024 4th International Conference on Mobile Networks and Wireless Communications (ICMNWC), Tumkuru, India, 4–5 December 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–7. [Google Scholar]
- Lu, S.; Duan, L.M.; Deng, D.L. Quantum adversarial machine learning. Phys. Rev. Res. 2020, 2, 033212. [Google Scholar] [CrossRef]
- Hohenfeld, H.; Heimann, D.; Wiebe, F.; Kirchner, F. Quantum deep reinforcement learning for robot navigation tasks. IEEE Access 2024, 12, 87217–87236. [Google Scholar] [CrossRef]
- Nguyen, H.Q.; Nguyen, X.B.; Churchill, H.; Choudhary, A.K.; Sinha, P.; Khan, S.U.; Luu, K. Quantum-brain: Quantum-inspired neural network approach to vision-brain understanding. arXiv 2024, arXiv:2411.13378. [Google Scholar]
- Busemeyer, J.R.; Bruza, P.D. Quantum Models of Cognition and Decision; Cambridge University Press: Cambridge, UK, 2012. [Google Scholar]
- Pothos, E.M.; Busemeyer, J.R. Can quantum probability provide a new direction for cognitive modeling? Behav. Brain Sci. 2013, 36, 255–274. [Google Scholar] [CrossRef]
- Herrman, R.; Lotshaw, P.C.; Ostrowski, J.; Humble, T.S.; Siopsis, G. Multi-angle quantum approximate optimization algorithm. Sci. Rep. 2022, 12, 6781. [Google Scholar] [CrossRef]
- Cadi Tazi, L.; Thom, A.J. Folded spectrum vqe: A quantum computing method for the calculation of molecular excited states. J. Chem. Theory Comput. 2024, 20, 2491–2504. [Google Scholar] [CrossRef] [PubMed]
- Chauhan, J.; Alam, T. Adjustable rotation gate based quantum evolutionary algorithm for energy optimisation in cloud computing systems. Int. J. Comput. Sci. Eng. 2024, 27, 414–433. [Google Scholar] [CrossRef]
- Koczor, B.; Morton, J.J.; Benjamin, S.C. Probabilistic interpolation of quantum rotation angles. Phys. Rev. Lett. 2024, 132, 130602. [Google Scholar] [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).