Next Article in Journal
Investigation on Smoke Flow in Stairwells induced by an Adjacent Compartment Fire in High Rise Buildings
Previous Article in Journal
Centrosymmetric Optical Vortex
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Agent Planning under Uncertainty with Monte Carlo Q-Value Function

1
Space Engineering University, Beijing 101416, China
2
Institute of Natural and Mathematical Sciences, Massey University, Albany, Auckland 0632, New Zealand
3
State Key Laboratory of Complex Electromagnetic Environment Effects on Electronics and Information System, Luoyang 471003, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2019, 9(7), 1430; https://doi.org/10.3390/app9071430
Submission received: 12 February 2019 / Revised: 24 March 2019 / Accepted: 30 March 2019 / Published: 4 April 2019
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Decentralized partially observable Markov decision processes (Dec-POMDPs) are general multi-agent models for planning under uncertainty, but are intractable to solve. Doubly exponential growth of the search space as the horizon increases makes a brute-force search impossible. Heuristic methods can guide the search towards the right direction quickly and have been successful in different domains. In this paper, we propose a new Q-value function representation—Monte Carlo Q-value function Q MC , which is proved to be an upper bound of the optimal Q-value function Q * . We introduce two Monte Carlo tree search enhancements—heavy playout for a simulation policy and adaptive samples—to speed up computation of Q MC . Then, we present a clustering and expansion with Monte-Carlo algorithm (CEMC)—an offline planning algorithm using Q MC as Q-value function, which is based on the generalized multi-agent A* with incremental clustering and expansion (GMAA*-ICE or ICE). CEMC calculates Q-value functions as required, without computing and storing all Q-value functions. An extended policy pruning strategy is used in CEMC. Finally, we present empirical results demonstrating that CEMC outperforms the best heuristic algorithm with a compact Q-value presentation in term of runtime for the same horizon, and has less memory usage for larger problems.

1. Introduction

Cooperative decision-making under uncertainty is a common problem in many scenarios such as sensor networks and airport service robots. Consider a scenario where a robot soccer team cooperates to beat the opponent. A total of 11 robot players dash in a simulated field, kicking, passing, and shooting a ball to score a goal. Each of them could only acquire his local observation and every action taken is uncertain, which means passing may be unsuccessful or shooting may be biased. Another example is congestion control, which is an important task in the Internet. The router’s buffer accommodates transient packets to guarantee a stable network link. The distributed routers need to come up with a strategy to avoid congestion. These kinds of problems can be modeled as a decentralized partially observable Markov decision process (Dec-POMDP), which provides a general multi-agent model for decision-making under uncertainty. However, Dec-POMDPs are provably intractable (NEXP-Complete, nondeterministic exponential time, it is known that P NP PSPACE EXP NEXP) [1] and difficult to solve with the number of joint policies growing explosively [2]. One of the primary reasons is that the search space of Dec-POMDPs grows exponentially with the number of actions and observations, and doubly exponentially with the horizon of the problem [3]. The optimal policy can be extracted from the optimal Q-value function Q * for Dec-POMDPs, but solving Q * is also proved to be NEXP-Complete [4].
In a previous work [5], Monte Carlo tree search (MCTS) [6] is introduced into the MMDP (Multiagent MDP) model, which is a simplified model for Dec-POMDPs, to solve ad hoc agent teams decision-making. MCTS is a sample-based tree search algorithm using Monte Carlo simulations as state evaluations, which has been successful in Computer Go [7]. MCTS will converge to the optimal value if exploration (find new nodes that are not sampled) and exploitation (find promising nodes that are already sampled) are traded off appropriately when given sufficient time and can be stopped at any time to get an approximate result.
To tackle the doubly exponential problem, we propose an offline algorithm called a clustering and expansion with Monte-Carlo algorithm (CEMC) based on incremental clustering and expansion (ICE), a version of generalized multi-agent A* (GMAA*) with incremental clustering and expansion, which is the best heuristic method in this literature. MCTS is integrated into the CEMC framework to fast obtain the approximate Q-value function. MCTS is an anytime algorithm, and its efficiency is up to sample times and the degree of accuracy of the Q-value function evaluation. Thus, we propose an adaptive sample method according to the horizon and the number of joint actions and observations. The best upper bound found in the MCTS algorithm is used to prune the nodes that will not be the optimal ones in any case. The term MCTS in this paper denotes the specific algorithm that uses UCB1 (a version of upper confidence bound) as the tree selection policy proposed by Kocsis and Szepesvári [6,8].
The key contribution is twofold: We present a new Q-value function representation—Monte Carlo Q-value function Q MC and an offline planning algorithm CEMC. Heavy playout for simulation policy and adaptive sample are introduced to speed up the computation of Q MC . The algorithm CEMC calculates the Q-value functions as required, without computing and storing all Q-value functions and it uses an extended pruning strategy for the policy pool. We present empirical results demonstrating that CEMC outperforms the best heuristic algorithm with a compact Q-value presentation in term of runtime for the same horizon, and has less memory usage for larger problems.
The remainder of this paper is structured as follows. We first introduce some related work in Section 2. Then, we outline some background knowledge on Dec-POMDPs, ICE algorithm, approximate Q-value functions, MCTS, and collaborative Bayesian Games in Section 3. Section 4 gives the Q-value function representation and proves that Q MC is an upper bound of the optimal Q-value function. Then, an adaptive sample and heavy rollout technique are proposed. Section 5 presents the core algorithm CEMC and a pruning strategy. We empirically prove the efficiency and effectiveness of CEMC in Section 6 and discuss the results. We conclude and outline the future research in Section 7.

2. Related Work

The Dec-POMDP is an extension of the POMDP model which has been proved to be doubly exponential [1]. The brute force search becomes intractable when the number of agents or the problem size grows. In recent years, a lot of work has focused on fast computing an optimal policy of larger horizon, such as dynamic programming [9,10,11,12], heuristic search [2,3,4,13,14,15], and mixed integer linear programming [16].
GMAA* framework [4] is one of the state-of-the-art exact algorithms in the heuristic search literature, which is based on multi-agent A* (MAA*) [2] and uses an approximate heuristic function to guide the search in the whole search space. GMAA*-ICE (or ICE) [3] is a heuristically guided algorithm based on a GMAA* framework [4] and has demonstrated impressive performance over GMAA*. Two successful components used in ICE are lossless clustering [13] and incremental expansion [14]. Recently, a Bayesian learning method [17] and approximate inferences and heuristics [18] are proposed for solving Dec-POMDP.
Silver and Veness [19] proposed a Monte-Carlo algorithm—POMCP—for online planning in large POMDPs. Unlike our approach, POMCP combines a Monte-Carlo update of the agent’s belief state with a Monte-Carlo tree search from the current belief state. Importantly, POMCP is suitable for single-agent partially-observable and unfactored Markov decision process (MDP). Amato and Oliehoek [20] proposed a scalable approach based on sample-based planning called factored-value partially observable Monte Carlo planning (FV-POMCP), to combat the intractability of an increasing number of agents. However, FV-POMCP requires a decomposition of the value function into a set of overlapping factors.

3. Background

Consider a scenario where a number of agents inhabit in a particular environment modeled with a Dec-POMDP, where the perception of the agents is uncertain and they must work together to achieve their goal. At each discrete time step t = 0 , 1 , , H 1 every agent takes an action and the combination of these actions influences the world, making it transform to the next state. The goal of the agents is to try to come up with a plan that maximizes their expected long-term rewards. The goal or task need to be completed with cooperation, because no single agent is competent enough. Therefore, the plan computed should be decentralized.
In this section, we briefly review some background of the decentralized POMDPs model, ICE, approximate Q-value functions, MCTS, and collaborative Bayesian games. We refer the readers to the previous work [2,3,4,21,22] about the detailed background.

3.1. Decentralized Partially Observable Markov Decision Processes (Dec-POMDPs) Model

Dec-POMDP generalizes POMDP to a multi-agent version where rewards are common and based on joint actions, but observations are individualistic [23,24]. The detail definition is as follows.
Definition 1.
A decentralized partially observable Markov decision process is defined as a 9-tuple <I, S, A, T, 𝒪, O, R, b0, h>, where
  • I = { 1 , , n } is the finite set of agents;
  • S = { s 0 , , s | S | 1 } is the finite set of states, where s 0 is the initial state;
  • A = × i = 1 n A i is the set of joint actions where a = < a 1 , a 2 , , a n > , and A i is the finite set of individual actions of agent i . The component a i in the joint action a belongs to A i ;
  • T is the transition function where P ( s | s , a ) denotes the probability of transiting to the new state s when taking the joint action a in state s ;
  • 𝒪 = × i = 1 n 𝒪 i is the set of joint observations where o = <o1, ..., on> ∈ 𝒪, and 𝒪i is the finite set of individual observations of agent i. The component o i in the joint observation o belongs to 𝒪i;
  • O is the observation function where P( o | a , s ) denotes the probability of observing o when the system state transfer to s taking the joint action a ;
  • R is the reward function, where R ( s , a ) denotes the immediate reward when taking the joint action a in state s ;
  • b 0 is the initial state distribution;
  • h is the finite horizon.
At every step t , the environment is in a particular state s , which emits a joint observation o t according to the observation model. Each agent obtains its individual observation component o i t . Each agent selects its individual action a i t , forming a joint action a t , which leads to a state s according to the transition model. Every agent receives an immediate reward r .
The action-observation history θ i t for agent i is the sequence of actions taken and observations received, which is denoted as a sequence ( o i 0 , a i 0 , o i 1 , , a i t 1 , o i t ) . Thus, the joint action-observation history θ t is < θ 1 t , θ 2 t , , θ n t > . A policy for agent i is a mapping from history to actions π i : θ i A i , and a joint policy is denoted as π = < π 1 , π 2 , , π n > which specifies a policy for each agent. The goal of solving Dec-POMDPs is to calculate an optimal joint policy to maximize the expected long-term accumulated reward, which is defined as
E ( i = 1 h 1 R ( s t , a t ) | π , b 0 )
The value of an optimal joint policy π * can be defined as follows:
V t ( π * ) = θ t P ( θ t | b 0 ) Q * ( θ t , π * ( θ t ) )
where π * ( θ t ) denotes the joint action that π * specifies for θ t , and
Q * ( θ t , a ) = R ( θ t , a ) + o t + 1 O P ( o t + 1 | θ t , a ) Q * ( θ t + 1 , π * ( θ t + 1 ) )
is the optimal Q-value function.
The policy π i of agent i specifies actions for all steps of Dec-POMDPs, and π i can be represented as a sequence of decision rules π i = ( δ i 0 , δ i 1 , , δ i h 1 ) . The decision rule δ i t for agent i at step t is a mapping from the length- t history to action δ i t : θ i t A i . The partial joint policy φ t is defined as a sequence of decision rules specified for steps 0 , , t 1 , which is denoted as φ t = ( δ 0 , δ 1 , , δ t 1 ) .

3.2. Generalized MAA* with Incremental Clustering and Expansion (GMAA*-ICE)

GMAA*-ICE (or ICE) [3] is a heuristically guided algorithm based on GMAA* framework [4], which generalizes the first multi-agent A* search MAA* [2]. It uses V ^ ( φ t ) = V 0 t 1 ( φ t ) + H t h 1 ( φ t ) to guide A* search over partially specified joint policies to construct a tree from top–down, where V 0 t 1 ( φ t ) is the actual expected reward achieved over the first t steps and H t h 1 ( φ t ) is a heuristic value for the remaining h-t steps. ICE uses an admissible heuristic to guarantee that it can find an optimal policy.
The key contributions of ICE are lossless clustering and incremental expansion. The authors introduced a probabilistic equivalence criterion that can be used to decide whether two individual action-observation histories, θ i , a , θ i , b for agent i , can be clustered. Especially in problems that have many histories satisfying probabilistic equivalence, ICE can significantly speed up planning. Incremental expansion defines a method of incrementally constructing a search tree by expanding the most promising child φ t + 1 = ( φ t , β * ) where β * is the optimal solution of collaborative Bayesian games. The author defined a node priority which helps find the tight lower bound early by first expanding deeper nodes [2].
ICE [3] is our baseline algorithm, but our methods are different from it as we integrate Q-value function computing with policy solving, so CEMC calculates the Q-value functions as required, and does not have to store and maintain a large matrix of the Q-value function.

3.3. Approximate Q-Value Functions

As mentioned above, ICE uses V ^ ( φ t ) = V 0 t 1 ( φ t ) + H t h 1 ( φ t ) to guide the search and H t h 1 ( φ t ) is a heuristically estimated value using approximate Q-value functions. There are three well-known approximate Q-value functions, Q MDP , Q POMDP and Q BG , which are all approximations for the optimal Q-value function Q * . Although the optimal policy can be extracted from Q * , it is costly and impractical to compute.
Szer and Charpillet [25] used Q MDP to solve Dec-POMDP that is approximated as an underlying MDP. This underlying MDP is defined by a single agent which takes joint actions but has the same state set, transition function, and reward function as original Dec-POMDPs. It assumes that the state can be fully observed and the single agent gains more information than Dec-POMDPs do.
Q POMDP [25] is another approximation for Q * . In this case, Dec-POMDPs are treated as underlying POMDPs defined by a single agent which takes joint actions and receives joint observations but has the same state set, transition function, observation function, and reward function. It assumes that underlying POMDPs are settings of centralized control, which allow for better coordination. This means that the single agent knows the joint action-observation history while the agents in Dec-POMDPs can only gain local information such as individual actions and observations.
Q BG is proposed by Oliehoek et al. [4], which assumes that the agents know the joint action-observation history θ t 1 and the joint action a t 1 for the previous time step t 1 . This assumption defines a Bayesian games (BGs) for each ( θ t 1 , a ) pair. The actions taken by agents can be solved by constructing BGs.
Q MDP , Q POMDP and Q BG are all proved to be upper bounds of Q * [4]. These approximate Q-value functions form a hierarchy of upper bounds to Q * :
Q * < Q BG < Q POMDP < Q MDP
This means that all these approximate Q-value functions are admissible heuristic for Dec-POMDPs.
Apart from the Q-value functions mentioned above, Oliehoek et al. [18] proposed a factored Q-value function as a heuristic payoff function for factored Dec-POMDP. The precondition of a decomposition is that the Q-value function of every joint policy is factored. Each local component of this kind of Q-value function depends on a subset of state factors and the action-observation histories and actions of a subset of agents.

3.4. Monte Carlo Tree Search (MCTS)

MCTS builds a search tree incrementally and asymmetrically in a given domain by taking a random sample. Each node in the tree denotes a state and links to child nodes which represent actions taken. It works by repeating selection, expansion, simulation, and backpropagation on a loop until some predefined computational budget, usually runtime or memory, is reached [26].
Many extensions have been reviewed in the literature [22]. MCTS is used to find the optimal feature subset through the feature space in [27]. In this paper, we use plain UCT (upper confidence bounds for trees) [6] as the tree policy for computing the Q-value function. Plain UCT is MCTS with UCB1 (a version of upper confidence bound), the tree policy of which is:
x ¯ i + 2 c 2 ln n n i
where x ¯ i is the mean value; n is the number of times that the parent node has been visited; n i is the number of times child i has been visited, and c > 0 is a constant.
The role of a tree policy is to select and create a node, which is carried out in the selection and expansion phase. A default policy is in the simulation phase. Its function is to play out the domain to estimate a state value. In a previous paper [28], four criteria for selecting actions are listed: (i) max child, (ii) robust child, (iii) max-robust child, and (iv) secure child. If the search terminates, one can choose one of the four selection mechanisms above for selecting an action to take. Our tree policy chooses the highest value node in the selection phase and only one node is expanded in the expansion phase. In the default policy (simulation phase), we replaced a random sample with simulation enhancement best-first policy. The advantage of randomly selecting actions is that it is simple, requires no domain knowledge, and can quickly simulate many times. However, it is not likely to be realistic compared to rational players [22]. The enhancement we proposed can use domain knowledge to simulate realistically. The details will be introduced in Section 3.3.

3.5. Collaborative Bayesian Games

A Bayesian game (BG) [29] is an augmented normal form game in which a number of agents interact in a one-shot manner, and each of them holds some private information. This private information defines the type of the agent. A collaborative Bayesian game (CBG) is a BG in which agents receive identical payoffs. A CBG is defined as follows:
Definition 2.
A collaborative Bayesian game (CBG) [3,4] B ( b 0 , φ t ) = < I , A , Θ , Pr ( ) , u > modeling stage t of a Dec-POMDP, given initial state distribution b 0 and past joint policy φ t , consists of:
  • I = { 1 , 2 , , n } is the finite set of agents;
  • A is the set of joint actions;
  • Θ is the set of joint types, each of which specifies a type for each agent θ = < θ 1 , θ 2 , , θ n > ;
  • Pr ( ) is a probability distribution over joint types;
  • u is a heuristic payoff function mapping joint types and joint actions to a real number: u ( θ , a ) .
That is, given b 0 , for each φ t , it is possible to construct a CBG B ( b 0 , φ t ) . Therefore, B ( b 0 , φ t ) represents a decision-making problem for stage t , if φ t was followed for the first t stages starting from b 0 . We simply write B ( φ t ) below.
In a CBG, each agent uses a Bayesian game policy β i that maps individual types to actions β i ( θ i ) = a i . If β denotes the joint policy for the CBG, it corresponds to a joint decision rule: β δ t . Oliehoek et al. [4] model Dec-POMDP with series of CBGs, or in other words, a stage of a Dec-POMDP can be modeled with a CBG. Therefore, we can solve a Dec-POMDP problem stage by stage, which facilitates A*-like search algorithms.

4. Monte Carlo Q-Value Function

In this section, we propose a new Q-value function representation—Monte Carlo Q-value function Q MC —and give proof of an upper bound of the optimal Q-value function Q * . Q MC combines the precision of tree search and the anytime character of sample-based Monte Carlo, and has more flexibility than other Q-value functions do. Then, we present two MCTS enhancements—heavy playout for simulation policy and adaptive sample—to speed up the computation of Q MC .

4.1. Q MC

In reinforcement learning [30], the Q-value function Q ( s , a ) is defined as the expected return starting from state s , taking action a , and then following policy π :
Q π ( s , a ) = E π [ R t + 1 + γ Q π ( S t + 1 , a ) | S t = s , A t = a ] = s , r p ( s , r | s , a ) [ r + γ Q ( s , a ) ] .
Different from the former, the state in Q MC is the joint history θ t at time step t . We define Q MC as:
Q MC ( θ t , a ) = R ( θ t , a ) + o t + 1 O P ( o t + 1 | θ t , a ) max M C ( o t + 1 ) Q MC ( θ t + 1 , M C ( o t + 1 ) )
where MC is the Monte Carlo method, which returns the optimal action M C ( o t + 1 ) when observing o t + 1 . Because of the partially observability, the probability multiplied by Q MC is a conditional probability of all observations at θ t and a .
The solution of Q MC can be illustrated in Figure 1. The black filled nodes are joint histories, and the solid links and dotted links are respectively joint actions taken and joint observations received by agents. Each joint history node contains two statistical values— N ( θ ) , visiting count of node θ and N ( θ , a ) , number of taking a at θ .
At the selection phase of Q MC , UCB1 is used to choose the best action a *
a * = argmax a A { R + R ^ + c 2 ln N ( θ ) N ( θ , a ) }
where R and R ^ are respectively the immediate and future reward at .., and the term added to it is called a UCT bonus, which is used to balance exploitation and exploration. R + R ^ is the mean value of reward. The c parameter is a constant ( c > 0 ), which is used to tune how much influence the bonus has on the action selection. It is set bigger for problems with more branches and smaller for deeper ones. We did not consider adaptive setting of parameter c in this paper.
At the expansion phase of Q MC , it expands only one node. At the simulation phase, Q MC replaces the default random policy with heavy playout—an enhancement of Monte Carlo, which is presented in Section 3.3. At the backpropagation phase, Q MC updates the values or statistics of nodes that are visited in this iteration with the result of the finished play.
Lemma 1.
Q MC is an upper bound of the optimal Q-value function Q*.
Proof 
See appendix. ☐
From the representation of Q MC we can see that it treats Dec-POMDPs as models in which a single agent takes joint actions and receives joint observations, but has the same state set S , transition function T , observation function O and reward function R . The assumption makes more information available than the underlying Dec-POMDPs. Q MC is an overestimation of the optimal Q-value function and makes A*-like search find the optimal policy. The pseudocode of computing Q MC is shown in Algorithms 1 and 2.
Applsci 09 01430 i001
Applsci 09 01430 i002

4.2. Adaptive Sample

Due to the anytime character of MCTS, sufficient sample times ( K ) can make Q MC converge to Q * , but it is costly to calculate a converged Q MC . However, a small K leads to a loose Q-value function or even a wrong policy. Thus, in this article, we present a method for an adaptive sample, which is a function of joint actions and time steps. That is:
K = m | A | ( t + 1 ) 2
where m is a constant, determined by the specific problem, | A | is the number of joint actions and t is a time step.
K is in proportion to joint actions and inverse proportion to the square of time steps. This means that problems with more joint actions need larger K to make Q MC converge. With the tree of Q MC constructed from top–down, K shrinks gradually. We use this technique to balance the accuracy of convergence and computing time.

4.3. Best-First Simulation Policy

Simulation policy uses a default policy to play a game quickly. After that, MCTS propagates back the simulation results. The default policy is typically domain dependent. As there is no value like Q ( θ t , a ) for guiding the action selection, uniformly choosing available actions is a common scheme. However, the heuristic knowledge of the problem can be involved into the default policy to strengthen simulation. Silver [31] discusses some factors that influence simulation performance. In this paper, we extend the rollout phase by incorporating domain knowledge into playouts, which are called heavy playouts [32]. If it is not the last time step, the algorithm recursively calls for the Rollout function. Then it returns the best immediate reward (immR) and backpropagates the value obtained and updates N ( θ ) and N ( θ , a ) . This Best-First Simulation Policy (BFSP) returns the promising node first to ensure that the heuristic value is admissible or a guaranteed overestimation. The detail is shown in Algorithm 3.
Especially we would like to note that the selectBestA() function selects the best action combining an idea of reinforcement learning and the history heuristic. Reinforcement learning (RL) reinforces the actions that make success. The history heuristic assumes that an action good in one position may be good in another. Joint actions, which acquire the highest simulating value, will be given more chances in the later simulation. Concretely, we keep an average accumulated value v ( a ) of each joint action in a playout. Apart from v ( a ) , a separate visit of each action n i and total visits of all actions, n in the simulation phase is also stored. An action is selected by a softmax function with inputs z = v ( a ) n i n . This implies that the larger z will correspond to a larger probability. If it is at an initial state or the chance of nodes are equal, a random legal action will be selected. For example, if a problem has nine joint actions, in the start of a simulation phase, an action is selected through this softmax function. After the simulation ends, its accumulated value to the terminal is recorded and the visit number increases by one. In another simulation phase, the statistics are reused, which we call the history heuristic.
Applsci 09 01430 i003

5. Multi Agent Planning Using Q MC

In this section, we present our offline planning algorithm—CEMC—which is based on an ICE framework. Then, we introduce a new policy pruning strategy to speed up planning. CEMC integrates Q-value function computing and plan solving, and calculates the Q-value functions as required, which avoids the need to compute and store all Q-value functions. Thus, CEMC does not calculate and maintain a large matrix for Q-value, which achieves improvement in runtime and memory.

5.1. Offline Planning Algorithm—CEMC

Offline algorithms generate a complete policy before execution begins. The policy specifies what actions to take in any possible runtime situation. Before presenting CEMC, we first introduce two successful techniques used in ICE, which is also adopted by CEMC.
The first one is lossless clustering. Two individual action-observation histories θ i , a , θ i , b for agent i can be losslessly clustered when they satisfy probabilistic equivalence criterion:
θ i , s , P ( s , θ i | θ i , a ) = P ( s , θ i | θ i , b )
where θ i is the joint action-observation histories of all agents except agent i . This means that the probability distribution over the other agent’s joint action-observation histories are identical for both θ i , a , θ i , b .
The second one is incremental expansion which defines a method of incrementally constructing a search tree by expanding the most promising child φ t + 1 = ( φ t , β * ) where β * is the optimal solution of CBGs. A child is selected by node priority. Node q in the search tree is a 3-tuple < φ t , v ^ , h > . Node priority q < q if:
{ v ^ < v ^ h < h , , if   v ^ v ^ otherwise   if   h h .
That is, if the heuristic values are equal, the greater depth will be chosen first. This helps find the tight lower bound early by first expanding deeper nodes and converge quickly.
The CEMC algorithm (Algorithm 4) starts planning with an unspecified partial joint policy φ 0 in an initialized policy pool. It greedily selects the best partial joint policy in the pool according to the node priority(Select(L)). Then, it proceeds by constructing a CBG using lossless clustering (ClusterBG()). After that, the CBG is solved, which returns a 2-tupe β t , V ^ ( β t ) where β t is a Bayesian game policy, and V ^ ( β t ) is given by the function Solver():
V ^ ( β t ) = θ t P ( θ t | b 0 , φ t ) Q MC ( θ t , β ( θ t ) )
where P ( θ t | b 0 , φ t ) is the probability distribution over joint histories and Q MC ( θ t , β ( θ t ) ) is the Q-value function computed by Algorithm 1.
Applsci 09 01430 i004
The set of child nodes N IE is expanded using the incremental expansion technique and only the promising child node β t , * will be put into N IE (IncExp()). The heuristic value v ^ of the node is given by the function ComputePolicyAndValue():
V ^ ( φ t + 1 ) = V 0 t 1 ( φ t ) + V ^ ( β t )
where V 0 t 1 ( φ t ) is the actual expected reward of φ t and V ^ ( β t ) is given by Equation (12).
We use two pruning strategies during planning. If time step t < h 1 , nodes with values less than the best Q-value function Q MC * will be pruned in advance. If t = h 1 , CEMC computes fully expanded policies. Nodes with values less than the best lower bound v _ can be pruned. CEMC ends when the policy pool becomes empty or the highest value in the policy pool is less than that of the best full policy.
The main idea of CEMC is that it integrates the Q-value function computing with plan solving, which benefits from the anytime character of Q MC . The original ICE computes and stores the Q-value functions in a N j a o h × N j a matrix, where N j a o h is the number of joint action-observation histories and N j a is the number of joint actions. It is costly to calculate complete Q-value functions in advance. For example, for the benchmark problem, Mars Rovers (Appendix Table A1), there are about 2.82 × 1014 joint action-observation histories and 36 joint actions. It needs to compute about 1.02 × 1016 ( θ , a ) value pairs. The algorithm CEMC calculates the Q-value functions only when required, without computing and storing all Q-value functions.

5.2. Policy Pruning Strategy

Pruning is split into hard pruning, which removes branches of the tree permanently, and soft pruning, which temporarily removes branches, but then adds them back into the search at some later point [33]. In this paper, we use hard pruning as our pruning strategy.
The ICE algorithm maintains a policy pool. A node in the pool is denoted as a tuple < φ , v ^ , h > , where φ is a partial policy, v ^ is its value, and h is its depth. We know the value of a node is an actual value plus an estimated value: φ = V 0 t 1 ( φ t ) + V ^ ( β t ) . The estimated value V ^ ( β t ) is an overestimation of the actual value from this node to a terminal. Actually, the value of a node φ is the value of a path or a policy across this node. Therefore, with the search descending deep down to the terminal or the finite horizon, the value of a node is also descending (note that the actual value becomes larger and larger, and the estimated value becomes close to the actual value). When the search reaches the terminal, the value of the terminal node can be seen as the best lower bound ever found. If the value of a node in the policy pool is smaller than the best lower bound, it would never be the optimal. Therefore, it uses the best lower bound v _ to prune the policy pool. However, for some problems in the literature, the policy pool shrinks slowly, which wastes a lot of time in computing useless nodes.
We use an example to illustrate it. Consider, for example, Figure 2 is a search tree. The policy pool is a priority list, which has several nodes, such as [ φ 2 , 10.125 , 2 , φ 1 , 10.125 , 1 , φ 2 , 9.721 , 2 , φ 2 , 6.182 , 2 ,…]. The superscript of φ is its time step or its depth. At each loop the ICE algorithm chooses the first node in the pool and expands it. If the algorithm reaches the last step, it returns the fully expanded node ( φ 3 , 5.363 , 3 ) as the best lower bound to prune the policy pool. However, node φ 2 , 6.182 , 2 > node φ 3 , 5.363 , 3 according to the node priority, and it will not be pruned. Assuming that the child of the node φ 2 is φ 3 , 4.182 , 3 , this fully expanded policy is obviously not the optimal one and should have been pruned in advance, but node φ 2 , 6.182 , 2 will be still expanded in these iterations. Not pruning in advance makes the search algorithm go the wrong direction and this is to be solved in our policy pruning strategy.
Thus, these useless nodes like φ 2 , 6.182 , 2 should be removed as early as possible. We improve the pruning function using the max Q-value function at time step t Q MC t , * , combined with the best lower bound v _ , that is:
{ { q L | v ^ > v * } { q L | v ^ > v _ } , , if   t < h 1 if   t = h 1
where t is the deepest level in the policy pool, v * is computed through Q MC t , * :
v * = V 0 t 1 ( φ t ) + max a Q MC ( θ t , a )
The Equation (14) illustrates that if t < h 1 , all nodes with values less than v * can be pruned and if CEMC reaches the last step, any node with values less than v _ can be pruned. Note that pruning using v * is done during the planning, it does not have to wait for full expansion. This technique can reduce the number of nodes in the policy pool and make the algorithm converge quickly, which is tested experimentally in Section 5.

6. Experiments

This section experimentally demonstrates and validates the effectiveness of all techniques proposed: the new Q-value function representation Q MC , the adaptive sample, best-first simulation policy (BFSP), policy pruning strategy, and the core algorithm CEMC. We give several benchmark problems in the literature, followed by experimental data. Finally, we analyze and discuss the results.

6.1. Experiments Setup

We give some benchmark problems, including Dec-Tiger, Grid-Small, Recycling Robots, Box Pushing, Mars Rovers, Broadcast Channel. All these problems have two agents, with different numbers of states, actions and observations, as shown in Table A1 (see Appendix A), which lists the number of joint policies for different horizons. These domains can be downloaded at MASPlan (http://masplan.org/).
The problem, Broadcast Channel, has two agents, each of them must choose whether or not to send a message in a Broadcast Channel. If both agents send, a collision occurs. Therefore, they must cooperate to send messages. The problem, Recycling Robots, also has two agents in the environment, in which both of them move around using their motors, pick up cans, and place them in a specified place using their mechanical arms. They must complete the task cooperatively. The benchmark Mars Rovers involve two rovers performing certain scientific experiments by choosing to check, drill, or sample. They can move in four directions. Some locations need only one agent and some locations require their cooperation.
Experiments were run on Ubuntu 16.04 with Intel Core i5 CPU and 6 GB memory. CEMC was implemented in C++ based on the MADP Toolbox [34]. Every average result was run over 100 times. We compared CEMC with the state-of-the-art exact heuristic algorithm ICE with a compact Q-value function implementation Q BG . Five statistics are taken into consideration: time of computing Q-value function t Q , time of calculating the optimal joint policy t p , overall time t o , value of optimal policy V * , and value of random policy V r . We would like to note that ICE may not get the same performance contrasted with the original paper due to the computing resources, but all results including ICE and CEMC were tested in the same environment, so the time and scalability results are directly comparable.

6.2. Experiment Results

6.2.1. Comparing CEMC and ICE

Table 1 shows performance results for ICE and CEMC. “-” indicates unknown values, caused by getting the wrong results or a time limit violation (less than 3600 s is valid). “*” indicates out of memory. Table 1 clearly shows that CEMC outperforms the state-of-the-art heuristic algorithm with Q BG . Especially in solving scalable problems such as Grid-Small, Box Pushing, and Mars Rovers, CEMC shows significant improvement without computing and storing all Q-values. CEMC allows for getting the needed Q-value to guide A* search without pre-computation of all Q-value function.
From Table 1 we can see that in some problems like Grid-Small, Box Pushing, and Mars Rovers, CEMC is significantly better than ICE, while in some problems like Dec-Tiger, Recycling Robots, and Broadcast Channel the difference between both may not be as obvious as the three previous ones. Actually, the number of joint decision rules for time step t in the search tree is
O ( | A | n ( | 𝒪 | t ) )
which means the complexity is exponential with the number of actions and observations, and doubly exponential with the time step. The search tree contains the action-observation histories, the actions taken, and the observations received. Therefore, the number of branches is related to the action and observation space. Although the domain Mars Rovers has a larger state space, it has little influence on the scale of the search tree. What it influences is the efficiency of state transition and the computing time of the probability like P ( o | a , s ) .

6.2.2. Analysis of Policy Pruning Strategy

To illustrate the effectiveness of the extended policy pruning strategy, we list the max number of policies in the policy pool for different horizons. Figure 3 shows a comparison of the original ICE and CEMC. Note that the original ICE maintains a larger policy pool than CEMC. For problems with approximatively V ^ ( φ t ) , ICE expands nodes back and forth to construct a tree carefully. This will significantly increase the search time. We use Equation (14) to prune the policy pool, and the promising nodes are left.

6.2.3. Convergence of Q MC

More sample times can make the approximate Q-value function Q MC converge and be close to Q * , but it needs more runtime. To illustrate the convergence of Q MC , we gathered the statistics of the max Q MC of Dec-Tiger with the sample time K increasing, as shown in Figure 4. The max values of Q MDP , Q POMDP and Q BG for Dec-Tiger with horizon h = 3 and h = 4 are also listed. We can see from the figure that the max Q MC declines and converges as K increases. Note that Q MC is as compact as Q POMDP with horizon h = 3. In horizon h = 4, Q MC is more compact than Q POMDP but looser than Q BG . The goal of Figure 4 is to show that Q MC is an upper bound of Q * and we can use a less compact Q-value function to reduce time of computing, but it still can guide A* to search the optimal policy.

6.3. Analysis and Discussion

CEMC outperforms the state-of-the-art exact heuristic algorithm ICE with compact Q-value function Q BG in terms of runtime and memory. There are three reasons for CEMC’s performance.
Firstly, we propose a new Q-value function representation Q MC , which is an upper bound for the optimal Q-value function Q * . QMC is an anytime Q-value function, which makes it possible to compute as required. Adaptive sample and heavy rollout speed up the computation of Q MC .
Secondly, CEMC calculates the Q-value function as required and does not maintain a large matrix to store all Q ( θ t , a ) . For problems with a large joint action-observation space and horizon, maintaining the matrix will run out of memory. In addition, it can be seen from problems, Box Pushing and Mars Rovers, in Table 1 that it is costly to calculate the Q-value function (time of computing Q-value for ICE is more than 1100 s). If not every ( θ t , a ) pair is computed, it will save much time. A tighter upper bound can speed up planning, as it allows the algorithm to prune the policy pool more [4], but a tighter value function increases time of computing Q-value function, which accordingly increases the overall time (total time for computing Q-value and planning).
Thirdly, an extended policy pruning strategy is used in CEMC. The pruning strategy uses the best lower bound ever found v _ and the best accumulated Q-value function at time step t Q MC t , * to prune the policy pool. This technique avoids expanding those nodes that will not be the optimal ones in any case.

7. Conclusions

In this paper, we combine the advantages of the heuristic algorithm and MCTS, and propose a new Q-value function representation and a novel offline algorithm CEMC, generalizing the prior method. In particular, we demonstrate two important results: (i) Computation of the Q-value function as required is an important reason for the CEMC’s performance. This permits us no longer to maintain a large matrix to store all Q-value function pairs, which saves much memory; (ii) There is no need to compute Q-value functions as compact as possible. Experimental results demonstrate that it is costly to compute compact Q-value functions. Other techniques such as an adaptive sample, heavy rollout, and policy pruning strategy are introduced to speed up the algorithm. In the future work, we plan to analyze the relationship between the loose extent and sample times in order to find the smallest K to optimally solving Dec-POMDPs.

Author Contributions

J.Z. proposed the method and wrote the main part of this paper; Y.P. designed the experiment of this paper; R.W. and Y.F. checked the grammar; H.Y. supported this work financially and checked the paper.

Funding

This research was funded by the research foundation of Space Engineering University, grant number: zx10356.

Acknowledgments

The authors acknowledge the financial support received from the research foundation of Space Engineering University, grant number: zx10356, for their support and encouragement in carrying out this research.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Domain parameters and number of π for different h .
Table A1. Domain parameters and number of π for different h .
ProblemDomain Parameters num .   of   π   for   h
n|S||A||O|h = 2h = 5h = 10
Dec-Tiger223265613.43 × 10301.39 × 10977
Recycling Robots243265613.43 × 10301.39 × 10977
Grid-Small216523906255.42 × 10443.09 × 101431
Box Pushing2100453.34 × 1075.23 × 109401.25 × 102,939,746
Mars Rovers2256681.69 × 10141.88 × 1072852.57 × 10238,723,869
Broadcast Channel24222561.84 × 10193.23 × 10616
Lemma 1. Proofs
We use a Dec-POMDP with k steps delayed communication model to prove Lemma 1. This model can be denoted as an augmented MDP, which is a reformulation of the prior work [35,36] and extended to the Dec-POMDP setting [4].
The augmented MDP is defined as M ^ = < S ^ , A ^ , T ^ , R ^ > , where an augmented state s ^ S ^ is composed of a joint action-observation history θ t and a joint policy tree q t :
s ^ = { θ t , q t θ t , q τ = h t , , 0 t h k 1 h k t h 1
The contained q t is a joint depth-k policy tree q t = q 1 t , , q n t starting from stage t and q τ = h t is a τ -steps-to-go sub-tree policy. For example, if k = 2, q t represents a joint depth-2 policy tree for 0 t h 3 . For h 2 t h 1 , state s ^ contains q τ = 2 and q τ = 1 .
An augmented action a ^ A ^ is a joint policy a ^ = < δ 1 t + k , , δ n t + k > mapping length-k observation histories to joint actions at stage t + k. The transition model T ^ is a probability P ( s ^ t + 1 | s ^ t , a t ) = P ( θ t + 1 , q t + 1 | θ t , q t , δ t + k ) = P ( o t + 1 | θ t , q t ( o ) ) with an empty observation o at stage t, where θ t + 1 = ( θ t , a t , o t + 1 ) , q t ( o ) is an initial joint action at stage t and q t + 1 = q t δ t + k meaning appending a policy δ t + k to q t . The reward model R ^ is defined as R ^ ( s ^ ) = R ^ ( θ t , q t ) = R ( θ t , q t ( o ) ) .
Therefore, the optimal Q-value function Q k for a k-steps delayed communication model for 0 t h k 1 is as follows.
Q k ( s ^ , δ t + k ) = R ( θ t , q t ( o ) ) + o t + 1 O P ( o t + 1 | θ t , q t ( o ) ) max δ t + k Q k ( s ^ t + 1 , δ t + k )
For h k t h 1 , there are τ = h t stages to go and the Q-value function is the same as the original one.
If k = 0, q t is a depth-0 policy and s ^ reduces to θ t , or θ t , δ t + k reduces to a t . Equation (18) reduces to Q 0 ( θ t , a t ) = R ( θ t , a t ) + o t + 1 O P ( o t + 1 | θ t , a t ) max a t + 1 Q 0 ( θ t + 1 , a t + 1 ) . Contrasting with Equation (7), we can get that Q MC corresponds to the Q-value function Q 0 of a 0-delayed communication system. According to the theory of [4], a Dec-POMDP is identical to an h-steps delayed communication system. The optimal Q-value function Q k of a Dec-POMDP with k-steps delayed communication is an upper bound to Q k + 1 , that of a k + 1-steps delayed communication system. This means that Q MC , the Q-value function of a 0-delayed communication system, is an upper bound of the optimal Q-value function Q * of a Dec-POMDP (h-steps delayed communication system), that is:
Q * < Q MC

References

  1. Bernstein, D.S.; Givan, R.; Immerman, N.; Zilberstein, S. The complexity of decentralized control of Markov decision processes. Math. Oper. Res. 2002, 27, 819–840. [Google Scholar] [CrossRef]
  2. Szer, D.; Charpillet, F.; Zilberstein, S. MAA*: A heuristic search algorithm for solving decentralized POMDPs. In Proceedings of the 21st Conference on Uncertainty in Artificial Intelligence, Edinburgh, UK, 4 June 2005; pp. 576–583. [Google Scholar]
  3. Oliehoek, F.A.; Spaan, M.T.; Amato, C.; Whiteson, S. Incremental clustering and expansion for faster optimal planning in Dec-POMDPs. J. Artif. Intell. Res. 2013, 46, 449–509. [Google Scholar] [CrossRef]
  4. Oliehoek, F.A.; Spaan, M.T.; Vlassis, N. Optimal and approximate Q-value functions for decentralized POMDPs. J. Artif. Intell. Res. 2008, 32, 289–353. [Google Scholar] [CrossRef]
  5. Wu, F.; Zilberstein, S.; Chen, X. Online planning for ad hoc autonomous agent teams. In Proceedings of the International Joint Conference on Artificial Intelligence, Barcelona, Catalonia, Spain, 28 June 2011; pp. 439–445. [Google Scholar]
  6. Kocsis, L.; Szepesvári, C. Bandit based Monte-Carlo planning. In Proceedings of the 17th European Conference on Machine Learning, Berlin, Germany, 18 September 2006; pp. 282–293. [Google Scholar]
  7. Silver, D.; Schrittwieser, J.; Simonyan, K.; Antonoglou, I.; Huang, A.; Guez, A.; Hubert, T.; Baker, L.; Lai, M.; Bolton, A.; et al. Mastering the game of Go without human knowledge. Nature 2017, 550, 354–359. [Google Scholar] [CrossRef] [PubMed]
  8. Kocsis, L.; Szepesvári, C.; Willemson, J. Improved Monte-Carlo Search; University of Tartu: Tartu, Estonia, 2006. [Google Scholar]
  9. Amato, C.; Dibangoye, J.S.; Zilberstein, S. Incremental policy generation for finite-horizon DEC-POMDPs. In Proceedings of the Proceedings of the Nineteenth International Conference on Automated Planning and Scheduling, Thessaloniki, Greece, 16 October 2009; pp. 569–576. [Google Scholar]
  10. Lam, R.; Willcox, K.; Wolpert, D.H. Bayesian Optimization with a Finite Budget: An Approximate Dynamic Programming Approach. In Advances in Neural Information Processing Systems; Lee, D.D., Sugiyama, M., Luxburg, U.V., Guyon, I., Garnett, R., Eds.; Curran Associates, Inc.: Barcelona, Spain, 2016; pp. 883–891. [Google Scholar]
  11. Rahmanian, H.; Warmuth, M.K. Online Dynamic Programming. In Advances in Neural Information Processing Systems; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Long Bench/Los Angeles, CA, USA, 2017; pp. 2827–2837. [Google Scholar]
  12. Boularias, A.; Chaib-Draa, B. Exact dynamic programming for decentralized POMDPs with lossless policy compression. In Proceedings of the Eighteenth International Conference on Automated Planning and Scheduling, Sydney, Australia, 14 September 2008; pp. 20–27. [Google Scholar]
  13. Oliehoek, F.A.; Whiteson, S.; Spaan, M.T.J. Lossless clustering of histories in decentralized POMDPs. In Proceedings of the International Conference on Autonomous Agents and Multiagent Systems, Budapest, Hungary, 10 May 2009; pp. 577–584. [Google Scholar]
  14. Spaan, M.T.J.; Oliehoek, F.A.; Amato, C. Scaling up optimal heuristic search in Dec-POMDPs via incremental expansion. In Proceedings of the International Joint Conference on Artificial Intelligence, Barcelona, Spain, 28 June 2011; pp. 2027–2032. [Google Scholar]
  15. Lanctot, M.; Winands, M.H.M.; Pepels, T.; Sturtevant, N.R. Monte Carlo Tree Search with heuristic evaluations using implicit minimax backups. In Proceedings of the 2014 IEEE Conference on Computational Intelligence and Games, Dortmund, Germany, 26 August 2014; pp. 1–8. [Google Scholar]
  16. Aras, R.; Dutech, A. An investigation into mathematical programming for finite horizon decentralized POMDPs. J. Artif. Intell. Res. 2010, 37, 329–396. [Google Scholar] [CrossRef]
  17. Liu, M.; Amato, C.; Liao, X.; Carin, L.; How, J.P. Stick-Breaking Policy Learning in Dec-POMDPs. In Proceedings of the 24th International Conference on Artificial Intelligence, Buenos Aires, Argentina, 25–31 July 2015; pp. 2011–2017. [Google Scholar]
  18. Oliehoek, F.; Whiteson, S.; Spaan, M.T.J. Approximate Solutions for Factored Dec-POMDPs with Many Agents. In Proceedings of the 12th International Conference on Autonomous Agents and Multiagent Systems (AAMAS), St. Paul, MN, USA, 6–10 May 2013; Volume 1, pp. 563–570. [Google Scholar]
  19. Silver, D.; Veness, J. Monte-Carlo Planning in Large POMDPs. In Advances in Neural Information Processing Systems 23; Lafferty, J.D., Williams, C.K.I., Shawe-Taylor, J., Zemel, R.S., Culotta, A., Eds.; Curran Associates, Inc.: Barcelona, Spain, 2010; pp. 2164–2172. [Google Scholar]
  20. Amato, C.; Oliehoek, F.A. Scalable Planning and Learning for Multiagent POMDPs. In Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, Austin, TX, USA, 25–30 January 2015. [Google Scholar]
  21. Oliehoek, F.A.; Amato, C. A Concise Introduction to Decentralized POMDPs; Springer: Berlin, Germany, 2016. [Google Scholar]
  22. Browne, C.B.; Powley, E.; Whitehouse, D.; Lucas, S.M.; Cowling, P.I.; Rohlfshagen, P.; Tavener, S.; Perez, D.; Samothrakis, S.; Colton, S. A Survey of Monte Carlo Tree Search Methods. IEEE Trans. Comput. Intell. Ai Games 2012, 4, 1–43. [Google Scholar] [CrossRef]
  23. Amato, C.; Chowdhary, G.; Geramifard, A.; Üre, N.K. Decentralized control of partially observable Markov decision processes. In Proceedings of the IEEE Conference on Decision and Control, Florence, Italy, 10–13 December 2013; pp. 2398–2405. [Google Scholar]
  24. Omidshafiei, S.; Aghamohammadi, A.; Amato, C.; How, J.P. Decentralized Control of Partially Observable Markov Decision Processes using Belief Space Macro-actions. In Proceedings of the 2015 IEEE International Conference on Robotics and Automation (ICRA), Seattle, WA, USA, 26–30 May 2015; pp. 5962–5969. [Google Scholar]
  25. Szer, D.; Charpillet, F. An Optimal Best-First Search Algorithm for Solving Infinite Horizon Dec-POMDPs. In Machine Learning: ECML 2005; Springer: Berlin/Heidelberg, Germany; Porto, Portugal, 2005; Volume 3720, pp. 389–399. [Google Scholar]
  26. Chaslot, G.; Bakkes, S.; Szita, I.; Spronck, P. Monte-Carlo Tree Search: A New Framework for Game AI. In Proceedings of the Fourth Artificial Intelligence and Interactive Digital Entertainment Conference, Stanford, CA, USA, 22 October 2008; pp. 216–217. [Google Scholar]
  27. Chaudhry, M.U.; Lee, J.-H. MOTiFS: Monte Carlo Tree Search Based Feature Selection. Entropy 2018, 20, 385. [Google Scholar] [CrossRef]
  28. Schadd, F. Monte-Carlo Search Techniques in the Modern Board Game Thurn and Taxis; Maastricht University: Maastricht, The Netherlands, 2009. [Google Scholar]
  29. Osborne, M.J.; Rubinstein, A. A Course in Game Theory; MIT Press: Cambridge, MA, USA, 1994. [Google Scholar]
  30. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; The MIT Press: Cambridge, MA, USA, 2018. [Google Scholar]
  31. Silver, D.; Tesauro, G. Monte-Carlo simulation balancing. In Proceedings of the International Conference on Machine Learning, Montreal, QC, Canada, 14 June 2009; pp. 945–952. [Google Scholar]
  32. Drake, P.; Uurtamo, S. Move Ordering vs Heavy Playouts: Where Should Heuristics Be Applied in Monte Carlo Go. In Proceedings of the 3rd North American Game-On Conference, Shiga, Japan, 15 September 2007; pp. 171–175. [Google Scholar]
  33. Sephton, N.; Cowling, P.I.; Powley, E.; Slaven, N.H. Heuristic move pruning in Monte Carlo Tree Search for the strategic card game Lords of War. In Proceedings of the 2014 IEEE Conference on Computational Intelligence and Games, Dortmund, Germany, 26–29 August 2014; pp. 1–7. [Google Scholar]
  34. Oliehoek, F.A.; Spaan, M.T.J.; Terwijn, B.; Robbel, P.; Messias, J.V. The MADP toolbox: An open source library for planning and learning in (multi-)agent systems. J. Mach. Learn. Res. 2017, 18, 1–5. [Google Scholar]
  35. Aicardi, M.; Davoli, F.; Minciardi, R. Decentralized optimal control of Markov chains with a common past information set. IEEE Trans. Autom. Control 1987, 32, 1028–1031. [Google Scholar] [CrossRef]
  36. Ooi, J.M.; Wornell, G.W. Decentralized control of a multiple access broadcast channel: Performance bounds. In Proceedings of the 35th IEEE Conference on Decision and Control, Kobe, Japan, 11 December 1996; Volume 1, pp. 293–298. [Google Scholar]
Figure 1. Tree of Monte Carlo Q-value function.
Figure 1. Tree of Monte Carlo Q-value function.
Applsci 09 01430 g001
Figure 2. Incremental clustering and expansion (ICE) search tree. Due to inappropriate policy pruning, some nodes like φ 2 will be left in the policy pool and they may be still expanded in later iterations. The dashed circle is an unvisited node.
Figure 2. Incremental clustering and expansion (ICE) search tree. Due to inappropriate policy pruning, some nodes like φ 2 will be left in the policy pool and they may be still expanded in later iterations. The dashed circle is an unvisited node.
Applsci 09 01430 g002
Figure 3. Max number of policies in the policy pool for different horizons. This indicator illustrates that ICE maintains a larger policy pool than CEMC, which may make the former expand useless nodes.
Figure 3. Max number of policies in the policy pool for different horizons. This indicator illustrates that ICE maintains a larger policy pool than CEMC, which may make the former expand useless nodes.
Applsci 09 01430 g003
Figure 4. Convergence of Q MC . as the sample time K increases. Q MC converges to Q POMDP , which means Q MC is an upper bound for Q * , but is less compact than Q BG .
Figure 4. Convergence of Q MC . as the sample time K increases. Q MC converges to Q POMDP , which means Q MC is an upper bound for Q * , but is less compact than Q BG .
Applsci 09 01430 g004
Table 1. Comparable results of some benchmarks.
(a)
(a)
hAlgo t Q ( s ) t p ( s ) t o ( s ) V * V r
Dec-Tiger
2ICE<0.01<0.010.27−4.00−93.05
CEMC0<0.010.27−4.00−93.05
3ICE0.060.020.435.19−140.81
CEMC00.060.285.19−140.81
4ICE-----
CEMC065.0865.364.80−186.67
“-” indicates unknown values, caused by getting the wrong results or a time limit violation (less than 3600 s is valid).
(b)
(b)
Recycling Robots
2ICE<0.01<0.010.276.82.6
CEMC0<0.010.276.82.6
3ICE0.050.020.409.763.17
CEMC00.060.339.763.17
4ICE1.550.011.8511.733.24
CEMC00.490.8011.733.24
5ICE-----
CEMC00.831.2113.763.78
10ICE*****
CEMC01.071.5121.2-
20ICE*****
CEMC02.532.9228.19-
“*” indicates out of memory.
(c)
(c)
GridSmall
2ICE0.010.010.950.860.20
CEMC00.040.950.860.20
3ICE3.880.034.841.370.38
CEMC00.911.811.370.38
4ICE158.440.03159.231.880.56
CEMC011.7612.661.880.56
5ICE*****
CEMC050.2351.332.53-
(d)
(d)
Box Pushing
2ICE30.730.0132.3917.60−0.01
CEMC00.541.8117.60−0.01
3ICE492.5821.76540.0066.08−0.61
CEMC00.822.3666.08−0.61
4ICE2974.12371.833253.2298.59-
CEMC03.245.9898.59-
5ICE*****
CEMC021.4930.79107.72-
(e)
(e)
Mars Rovers
2ICE1171.760.171179.055.80−3.17
CEMC03.8110.935.80−3.17
3ICE-----
CEMC07.5014.749.38-
4ICE-----
CEMC0102.74122.9710.18-
(f)
(f)
Broadcast Channel
2ICE<0.01<0.010.2820.85
CEMC0<0.010.2720.85
3ICE<0.010.010.282.991.13
CEMC00.030.292.991.13
4ICE0.13<0.010.423.891.47
CEMC00.210.303.891.47
5ICE2.770.013.094.791.76
CEMC00.230.324.791.76
10ICE*****
CEMC00.380.469.29-
20ICE*****
CEMC02.062.6318.31-
30ICE*****
CEMC0219.2222027.42-

Share and Cite

MDPI and ACS Style

Zhang, J.; Pan, Y.; Wang, R.; Fang, Y.; Yang, H. Multi-Agent Planning under Uncertainty with Monte Carlo Q-Value Function. Appl. Sci. 2019, 9, 1430. https://doi.org/10.3390/app9071430

AMA Style

Zhang J, Pan Y, Wang R, Fang Y, Yang H. Multi-Agent Planning under Uncertainty with Monte Carlo Q-Value Function. Applied Sciences. 2019; 9(7):1430. https://doi.org/10.3390/app9071430

Chicago/Turabian Style

Zhang, Jian, Yaozong Pan, Ruili Wang, Yuqiang Fang, and Haitao Yang. 2019. "Multi-Agent Planning under Uncertainty with Monte Carlo Q-Value Function" Applied Sciences 9, no. 7: 1430. https://doi.org/10.3390/app9071430

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop