Next Article in Journal
Accelerating Bayesian Estimation of Solar Cell Equivalent Circuit Parameters Using JAX-Based Sampling
Next Article in Special Issue
A Software-Defined Distributed Architecture for Controlling Unmanned Swarm Systems
Previous Article in Journal
Risk-Management Framework and Information-Security Systems for Small and Medium Enterprises (SMEs): A Meta-Analysis Approach
Previous Article in Special Issue
Multi-USV Dynamic Navigation and Target Capture: A Guided Multi-Agent Reinforcement Learning Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cooperative Decisions of a Multi-Agent System for the Target-Pursuit Problem in Manned–Unmanned Environment

1
China North Artificial Intelligence & Innovation Research Institute, Beijing 100072, China
2
Collective Intelligence & Collaboration Laboratory (CIC Lab), Beijing 100072, China
3
College of Artificial Intelligence, Nankai University, Tianjin 300071, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(17), 3630; https://doi.org/10.3390/electronics12173630
Submission received: 7 July 2023 / Revised: 13 August 2023 / Accepted: 23 August 2023 / Published: 28 August 2023
(This article belongs to the Special Issue New Technologies and Applications of Human-Robot Intelligence)

Abstract

:
With the development of intelligent technology, multi-agent systems have been widely applied in military and civilian fields. Compared to a single platform, multi-agent systems can complete more dangerous, difficult, and heavy tasks. However, due to the limited autonomy of unmanned platforms and the regulatory needs of personnel, multi-agent systems cooperating with manned platforms to perform tasks have been more widely promoted at this stage of development. This paper addresses a differential game method for cooperative decision-making of a multi-agent system cooperating with the manned platform for the target-pursuit problem. The manned platform pursues the target according to a certain trajectory, and its state can be obtained by the multi-agent system. Firstly, for the case that the target moves with a fixed trajectory, the target-pursuit problem in a manned–unmanned environment is viewed in the form of game based on a communication graph among agents. Secondly, strategies of all agents are proposed while maintaining their group cohesion. A set of coupled differential equations is solved to implement strategy calculation. Compared to purely unmanned systems, the strategies combine the advantages of the manned platform and add a reference item, which can achieve team cohesion relatively quickly. Furthermore, a brief analysis is made on the scenarios where the target is in another case or adopts other strategies. Finally, comparative simulations have verified the effectiveness and synergy of the strategy.

1. Introduction

The multi-agent system is widely concerned with the development of the unmanned system and intelligent technology [1,2]. Due to its advantages such as autonomy, loose coupling, high fault tolerance, and scalability, multi-agent systems are widely used in fields such as formation control [3], intelligent logistics [4], and collaborative search [5], among others. The cooperative decision-making of multi-agent systems focuses on the interactive behaviors among agents as well as the information exchanging between agents and the environment. Agents ultimately cooperate or compete with each other to accomplish tasks. In recent years, game theory has become a hot topic of research because of its ability to establish models for strategic interactions among agents, which can involve both cooperation and competition [6,7,8].
Due to the limitations in autonomy and intelligence of current unmanned systems, in many complex real-world scenarios, tasks are often executed through the collaboration of manned platforms and multiple unmanned platforms [9,10]. This collaborative approach allows for the full utilization of the strengths of both manned and unmanned platforms, addressing challenges that unmanned systems alone may struggle to tackle. Moreover, the participation of manned platforms can enhance task execution efficiency, success rate, and safety, thereby reducing people’s concerns about responsibility, decision-making, safety, and other regulatory aspects in multi-agent systems to some extent [11]. Therefore, as an emerging area of study, manned–unmanned collaboration is rapidly developing and being applied.
As a typical topic of the collaborative field, target-pursuit problems have received significant attention in recent decades due to their widespread application, such as drone planning [12], territorial defense [13], and search and rescue [14]. In [15], the pursuit problem was proposed first, where the agents move in a grid to capture randomly moving targets. A cooperative algorithm for the multi-agent system to pursue static targets was proposed in [16]. In [17], a real-time approach for the target-pursuit problem of the multi-agent system was studied, wherein two coordination strategies, blocking escape directions (BES), and using alternative proposals (UAL) were adopted. For the moving targets, an opportunistic framework was proposed for multi-robot pursuit of targets in a dynamic environment with partial observability [18]. The consensus pursuit problem was discussed in [19] for multi-agent systems; the authors designed a distributed multi-flocking approach based on the local information. Agents could choose the target adaptively by combining information with the circle formation control law proposed. Inspired by the division of roles in wolf hunting, new strategies for multi-robot collaborative tracking of targets was proposed in [20], which improved the convergence performance of the algorithm and shortened the hunting time significantly.
Furthermore, dynamic game theory is often used for target-pursuit problems because of its advantage in dynamic interaction and decision-making, and because the decision-making process of the game players is the process of cooperation or competition within the team [21]. The game theory approach in [22] attempted to address the target-tracking problem by using cost functions in an obstacle-free environment, and a PD-like fuzzy controller was used to tune the cost function weights. In [23], the pursuit of invisible targets was discussed within the framework of a game, and some results on how the minor closed properties are related to pathwidth are concluded. In the case of multiple pursuers, three kinds of pursuit strategies, cooperative, non-cooperative, and greedy, are proposed in [24], and the sufficient conditions for achieving uniform exponential stability are given. Furthermore, the multiple pursuers strategy in the presence of the parameter of uncertainty was given in [25], and the acquisition probability was determined using the Monte Carlo method. The distributed strategy designed by combining differential game theory with collaborative control theory can handle target-pursuit situations in system communication topology switching [26]. In [27], evolutionary game theory was introduced to solve the pursuit problem of multiple targets, an innovative three-level decomposition method was used to decompose the previously complex multiplayer game into multiple small-scale games, and a multi-agent Q-learning method based on evolutionary game models was proposed.
The studies mentioned above only consider the situations of target-pursuit problems by unmanned platforms or multi-agent systems. However, in practice, it is almost impossible that unmanned platforms perform complex tasks independently until now. The multi-agent system often needs to cooperate with manned platforms or accept commands from manned platforms. There are some results considering manned–unmanned cooperation, mostly qualitative decision-making on manned platforms and quantitative decision-making on unmanned platforms [28,29]. Cooperative operations of Manned/Unmanned Aerial Vehicle hybrid formation in antisubmarine warfare were analyzed in [30]. In order to compensate for the lack of intelligence in unmanned systems, the composition and key challenges of the MAV-UAV collaborative combat system were explored in [31]. These studies mostly focused on top-level planning from a global perspective, without considering coordination issues from the perspective of unmanned platforms. And note that none of the above literature focuses on target-pursuit problems.
In response to the above discussion, this paper proposes a cooperative decision-making method for the multi-agent system cooperating with a manned platform using differential game theory. The key idea of this method is to model the target-pursuit problem in the form of a game, where each agent is seen as a player, and each player can interact with their neighbors for state and strategies. The agents make decisions in the process of the dynamic game continuously, then cooperate with a manned platform to complete the target-pursuit task. More specifically, the contributions of this paper include the following:
  • A novel formulation of the differential game as a target-pursuit strategy in a manned–unmanned environment that can be applied in various manned–unmanned cooperative systems.
  • We propose a hierarchical decision-making approach for target-pursuit scenarios. The proposed method can enable the multi-agent system to cooperate with the manned platforms in a reasonable manner, ultimately achieving team cohesion while capturing the target. Furthermore, we discuss scenarios where the target is in different forms of motion. The method proposed in this paper can be applied to the static target, the target moving along fixed trajectories, and it can be extended to situations where the target adopts escape strategies.
  • Simulations prove that the proposed method can pursue the target in a manned–unmanned environment successfully. And the effects of each parameter are analyzed through comparative simulations.
The rest of this paper is organized as follows: Section 2 presents the required preliminaries and the formulation of the target-pursuit problem. In Section 3, the strategies for agents are designed using linear quadratic differential game theory, and simulations are presented to verify the effectiveness of the proposed approach. The discussion and outlook on the results are presented in Section 4. Section 5 concludes the paper.
Notation 1.
The notation used in the paper is fairly standard except where otherwise stated. R denotes the field of real numbers, R n denotes the set of n-dimensional real-valued column vectors and R n × m denotes the set of all n × m real matrices. I denotes the identity matrix of the compatible dimension and 0 m × n denotes the n-by-m dimensional zero matrix. A T represents the transpose of the matrix A; A 1 represents the inverse of the matrix A. For a matrix Q R n , the notation Q 0 means that Q is positive definite and Q 0 means positive semidefinite. In symmetric block matrices, ★ is used to denote a term induced by symmetry. The notation blkdiag { A 1 , , A n } represents a block diagonal matrix whose main diagonal consists of square matrices A 1 , , A n . The norm x denotes the Euclidean norm of vector x, x = ( x T x ) 1 / 2 , x R n . A B is the Kronecker product of matrix A and matrix B.

2. Problem Formulation

We consider the target-pursuit problem of the multi-agent system in this paper. The multi-agent system cooperates with an manned platform to pursue a target while maintaining its cohesive state. We describe the target-pursuit problem in the form of a game, which terminates when the target is successfully captured. Without loss of generality, agents are represented as nodes in the network, and the communication relationships among agents are represented by the edges of the network. Then, we introduce some basic knowledge of graph theory to describe the multi-agent system and the internal relationships.

2.1. Communication Graph

The internal relationships of the multi-agent system and communication among agents can be represented by a directed graph G = ( V , E ) , where V = { 1 , 2 , . . . , l } is the finite nonempty vertices set and E V × V is the set of edges. The set of vertices corresponds to the agents and the edges represent the interconnection between two agents. The edge of the graph ( i , j ) denotes that agent i can obtain information from agent j with weight ω i j 0 . The set of all neighbors of agent i in the graph G is represented by N ( i ) . If there is a path from vertex i to vertex j in graph G , then i and j are said to be connected. If every pair of vertices is connected, then the graph G is said to be connected.
For the directed graph G , the incidence matrix D describes the relationship between vertices and edges. D is the 0 , ± 1 matrix, where its u v th element is equal to 1 if node u is the head of the edge v, 1 if node u is the tail of the edge v, and 0 if u is not the vertex of v. For the weighted graph, each edge has its own weight coefficient. The weight matrix is represented by W = d i a g ( ω i j ) , which is a diagonal matrix. For agent i, its weight matrix W i is related to its neighbors. The Laplacian matrix L of the directed graph G can be defined as L = D W D T , which contains important properties of the graph.
Assumption 1.
The graph G does not have self-loops and only contains ordered pairs of distinct vertices. Graph G is a connected graph.
Assumption 2.
The manned platform can observe the state information of the target. All nodes of the multi-agent system can observe the state information of the manned platform, and some nodes can observe the state information of the target.

2.2. Game Model of Target-Pursuit Problem

Consider a scenario where a manned platform and a group of l agents cooperate to pursue a target. The manned platform pursues the target along a certain trajectory, and the multi-agent system makes decisions based on observing the states of the manned platform. The cooperative decision of the multi-agent system can be modeled as a class of a linear quadratic differential game with l players.
The cooperative decision-making process between the manned platform and multi-agent system can be seen as the form of hierarchical game. For a target, the manned platform takes the initiative in making decisions due to its superior capabilities, the multi-agent system observes the state of the manned platform and makes decisions accordingly to collaborate with the manned platform in completing the task.
Each agent acts as a player in the game, and its state equation is as follows:
x ˙ i = A i x i + b i u i , i = 1 , 2 , . . . , l ,
where i represents the number of the players, x i ( t ) R 2 n is the state vector, and u i ( t ) R n is the control vector for player i. A i = 0 n × n I n 0 n × n 0 n × n , b i = 0 n × n I n .
Consider the vector z = [ x 1 T , x 2 T , . . . , x l T ] T R 2 n l , according to (1), the system dynamics is given as follows:
z ˙ = A z + i = 1 l B i u i ,
where A = b l k d i a g A 1 , A 2 , . . . , A l , B i = [ 0 1 × ( i 1 ) , 1 , 0 1 × ( l i ) ] T b i R 2 n l × n .
Suppose that the target moves along a fixed trajectory, and its state equation is
x ˙ t = A t x t ,
where x t ( t ) R 2 n is the state vector of the target. Suppose that the manned platform moves along a certain trajectory to pursue the target, and its state equation is
x ˙ m = A m x m ,
where x m ( t ) R 2 n is the state vector of the manned platform.
Consider the vector z ¯ = [ x 1 T , x 2 T , . . . , x l T , x m T , x t T ] T R 2 n ( l + 2 ) , the augmented system dynamics is given as follows:
z ¯ ˙ = A ¯ z ¯ + i = 1 l B ¯ i u i ,
where A ¯ = A 0 2 n l × 2 n 0 2 n l × 2 n 0 2 n × 2 n l A m 0 2 n × 2 n 0 2 n × 2 n l 0 2 n × 2 n A t , B ¯ i = B i 0 2 n × n 0 2 n × n .
Then, we are intended to design individual cost functions for each agent according to its role in the system. Define the set of agents that can obtain the target’s state information as P , and the other agents as set F . For agent i P , the goal is to maintain a cohesive state with their neighbors N ( i ) while cooperating with the manned platform to pursue the target as efficiently as possible. Hence, the weighted distance that agent i needs to optimize is
p N ( i ) ω i p x i x p 2 + x i x m 2 + x i x t 2 ,
where the first term is the weighted sum of the distance between agent i and its neighbors N ( i ) , which is the term that denotes system cohesion; the second term is the distance between between agent i and the manned platform; the third term is the distance between agent i and the target.
The weighted distance (6) can be transformed into the following form:
z ¯ T L ¯ i z ¯ + z ¯ T K ¯ i 2 z ¯ + z ¯ T K ¯ i 3 z ¯ ,
where L ¯ i = L i I 2 n 0 2 n l × 2 n 0 2 n l × 2 n 0 2 n × 2 n l 0 2 n × 2 n 0 2 n × 2 n 0 2 n × 2 n l 0 2 n × 2 n 0 2 n × 2 n , L i = D W i D T , W i is the weight matrix of agent i; K ¯ i 2 = K ˜ i 2 I 2 n , K ˜ i 2 R ( l + 2 ) × ( l + 2 ) , K ˜ i 2 ( i , i ) = 1 , K ˜ i 2 ( i , l + 1 ) = 1 , K ˜ i 2 ( l + 1 , i ) = 1 , K ˜ i 2 ( l + 1 , l + 1 ) = 1 , and the remaining elements are 0; K ¯ i 3 = K ˜ i 3 I 2 n , K ˜ i 3 R ( l + 2 ) × ( l + 2 ) , K ˜ i 3 ( i , i ) = 1 , K ˜ i 3 ( i , l + 2 ) = 1 , K ˜ i 3 ( l + 2 , i ) = 1 , K ˜ i 3 ( l + 2 , l + 2 ) = 1 , and the remaining elements are 0.
The agent minimizes their own cost function during the game process. Using u i represents the weighted sum of strategies of all neighbors of agent i, which attempts to minimize the cost function of i; its form is as follows:
u i = α i j N ( i ) u j .
Design B ¯ i = α i j N ( i ) B ¯ j , Q ¯ i = τ ¯ i 1 L ¯ i + τ ¯ i 2 K ¯ i 2 + τ ¯ i 3 K ¯ i 3 , where τ ¯ i 1 , τ ¯ i 2 , and τ ¯ i 3 are weight coefficients. τ ¯ i 1 represents the importance that the agent i places on achieving cohesive states, τ ¯ i 2 indicates the importance that the agent i places on collaborating with human platforms to accomplish the task, and τ ¯ i 3 indicates the importance that the agent i places on pursuing the target.
Remark 1.
For agents that cannot observe the target’s information (the set F ), the weighted distance does not include the third term, hence, we set τ ¯ i 3 = 0 . The strategy of this type of agent is to maintain the cohesive state with its neighbors and the manned platform as much as possible. In practical scenarios, this type of agent is often a lower-level platform.
The cost function for agent i to be minimized can be expressed as follows:
J ¯ i = 0 t f ( z ¯ T Q ¯ i z ¯ + u i T R ¯ i u i + u i T R ¯ i u i ) d t + z ¯ T ( t f ) Q ¯ i f z ¯ ( t f ) , i = 1 , 2 , . . . , l ,
where t f is is the terminal moment, Q ¯ i = Q ¯ i T , Q ¯ i f = Q ¯ i f T , R i 0 and R i 0 . The purpose of this article is to solve strategy u i * so that it satisfies the following condition:
J ¯ i ( u i * , u i * ) J ¯ i ( u i , u i * ) ,
where i V = { 1 , 2 , . . . , l } .

3. Results

3.1. Solution to the Target-Pursuit Problem

In this subsection, we give strategies for the multi-agent system which cooperate with a manned platform to execute the target-pursuit task when the target moves along a fixed trajectory. We address a linear quadratic differential game method for the target-pursuit problem. According to the game model of Equations (5) and (9), the following theorem provides the pursuit strategies of agents while maintaining a cohesive state with the manned platform and neighbors.
Theorem 1.
Consider the system (5), the state Equations (1) of agents, and (9) of cost functions, then, the strategy of agent i
u i * = R i 1 B i T P ¯ i 11 z R i 1 B i T b i R i 1 B i T c i ,
satisfies the inequality (10).
In Equation (11), P ¯ i 11 = P ¯ i 11 T 0 and b c i , c i are solutions of the following coupled differential equation in [ 0 , t f ] :
P ¯ ˙ i 11   =   Q i + P ¯ i 11 A p = 1 , p i l S p P ¯ p 11 + A p = 1 , p i l S p P ¯ p 11 T P ¯ i 11
P ¯ i 11 S i P ¯ i 11 P ¯ i 11 S i P ¯ i ,
P ¯ i 11 ( t f )   =   Q i f ,
b ˙ i   =   A T + P ¯ i 11 S i + P ¯ i 11 S i + p = 1 , p i l P ¯ p 11 S p b i Q ¯ i 12 x m + P ¯ i 11 p = 1 , p i l S p b p ,
b i ( t f )   =   Q ¯ i 12 f x m ( t f ) ,
c ˙ i   =   A T + P ¯ i 11 S i + P ¯ i 11 S i + p = 1 , p i l P ¯ p 11 S p c i Q ¯ i 13 x t + P ¯ i 11 p = 1 , p i l S p c p ,
c i ( t f )   =   Q ¯ i 13 f x t ( t f ) ,
where S i = B i R i 1 B i T , S i = B i R i 1 B i T .
Proof of Theorem 1.
Define the function V i ( t , z ¯ ) = J ¯ i * ; according to [32], we can obtain
V i ( t , z ¯ ) t = min u i u i V i ( t , z ¯ ) z ¯ T ( z ¯ ˙ + B ¯ i u i ) + z ¯ T Q ¯ i z ¯ + u i T R ¯ i u i u i T R ¯ i u i ,
V i ( t f , z ¯ ) = z ¯ T ( t f ) Q ¯ i f z ¯ ( t f ) .
Minimizing Equation (18), we have
V i ( t , z ¯ ) z ¯ T B ¯ i + 2 R ¯ i u i * = 0 ,
V i ( t , z ¯ ) z ¯ T B ¯ i + 2 R ¯ i u i * = 0 ,
then, the following equations hold
u i * = 1 2 R ¯ i 1 B ¯ i T V i ( t , z ¯ ) z ¯ ;
u i * = 1 2 R ¯ i 1 B ¯ i T V i ( t , z ¯ ) z ¯ .
Define S ¯ i = B ¯ i R ¯ i 1 B ¯ i T , S ¯ i = B ¯ i R ¯ i 1 B ¯ i T , by applying Equations (22) and (23) to (18), so we easily obtain
V i ( t , z ¯ ) t = V i ( t , z ¯ ) z ¯ T A ¯ z ¯ 1 2 S ¯ i V i ( t , z ¯ ) z ¯ 1 2 S ¯ i V i ( t , z ¯ ) z ¯ + p = 1 , p i l B ¯ p u p + z ¯ T Q ¯ i z ¯ + 1 4 V i ( t , z ¯ ) z ¯ T S ¯ i V i ( t , z ¯ ) z ¯ + 1 4 V i ( t , z ¯ ) z ¯ T S ¯ i V i ( t , z ¯ ) z ¯ .
In order to solve the above partial differential equation, define V i ( t , z ¯ ) = z ¯ T P ¯ i z ¯ , then we have
V i ( t , z ¯ ) z ¯ = 2 P ¯ i z ¯ ,
V i ( t , z ¯ ) t = z ¯ T P ¯ ˙ i z ¯ .
By applying Equation (25) to (20), we obtain u i * = R ¯ i 1 B ¯ i T P ¯ i z ¯ . Similarly, it can be obtained that u i * = R i 1 B ¯ i T P ¯ i z ¯ . u i * , u i * and Equations (24)–(26) lead to the conclusion that
z ¯ T P ¯ ˙ i z ¯ = 2 z ¯ T P ¯ i A ¯ z ¯ S ¯ i P ¯ i z ¯ S ¯ i P ¯ i z ¯ p = 1 , p i l S ¯ p P ¯ p z ¯ + z ¯ T Q ¯ d i z ¯ + z ¯ T P ¯ i S ¯ i P ¯ i z ¯ + z ¯ T P ¯ i S ¯ i P ¯ i z ¯ .
Then, it is straightforward to see that
0 = z ¯ T [ P ¯ ˙ i + Q ¯ i P ¯ i S ¯ i P ¯ i P ¯ i S ¯ i P ¯ i + P ¯ i A ¯ p = 1 , p i l S ¯ p P ¯ p + A ¯ p = 1 , p i l S ¯ p P ¯ p T P ¯ i ] z ¯ ,
hence, the following equation holds
P ¯ ˙ i   =   Q ¯ i + P ¯ i A ¯ p = 1 , p i l S ¯ p P ¯ p + A ¯ p = 1 , p i l S ¯ p P ¯ p T P ¯ i
  P ¯ i S ¯ i P ¯ i P ¯ i S ¯ i P ¯ i ,
P ¯ i ( t f )   =   Q ¯ i f .
The time derivative of V i ( t , z ¯ ) is given by
V ˙ i = z ¯ T Q ¯ i z ¯ u i T R ¯ i u i + ( u i u i * ) T R ¯ i ( u i u i * ) u i T R ¯ i u i + ( u i u i * ) T R ¯ i ( u i u i * ) .
We integrate both sides of Equation (27) in [ 0 , t f ] and then have
J ¯ i = 0 t f ( ( u i u i * ) T R ¯ i ( u i u i * ) + ( u i u i * ) T R ¯ i ( u i u i * ) ) d t + V i ( z ¯ ( 0 ) ) .
The above equation shows that the control strategy u i * can minimize the cost function of the agent i.
Partitioning the matrix P ¯ i R 2 n ( l + 2 ) × 2 n ( l + 2 ) yields
P ¯ i = P ¯ i 11 P ¯ i 12 P ¯ i 13 P ¯ i 22 P ¯ i 23 P ¯ i 33 ,
where P ¯ i 11 R 2 n l × 2 n ) , P ¯ i 12 R 2 n l × 2 n , P ¯ i 22 R 2 n l × 2 n .
Similarly, matrices Q ¯ i , S ¯ i , S ¯ i , and A ¯ can also be partitioned according to this rule. Let Q i = τ i 1 L ^ i + τ i 2 K ^ i 2 + τ i 3 K ^ i 3 , where τ i 1 , τ i 2 , and τ i 3 are weight coefficients, L ^ i = L i I 2 n , K ^ i 2 = K i 2 I 2 n and K ^ i 3 = K i 3 I 2 n , K i 2 and K i 3 are l × l matrices, K i 2 ( i , i ) = 1 , K i 3 ( i , i ) = 1 , and the remaining elements are 0. Therefore,
Q ¯ i 11 = Q i , A ¯ 11 = A , A ¯ 22 = A m , A ¯ 33 = A t , S ¯ i 11 = S i = B i R i 1 B i T , S ¯ i 11 = S i = B i R i 1 B i T Q ¯ i 12 = { 0 1 × ( i 1 ) , 1 , 0 1 × ( l i ) } T I 2 n , Q ¯ i 13 = { 0 1 × ( i 1 ) , 1 , 0 1 × ( l i ) } T I 2 n .
According to the above definition of block matrices, Riccati differential Equations (29) and (30) can be decomposed; specifically, we obtain
P ¯ ˙ i 11   =   Q i + P ¯ i 11 A p = 1 , p i l S p P ¯ p 11 + A p = 1 , p i l S p P ¯ p 11 T P ¯ i 11
  P ¯ i 11 S i P ¯ i 11 P ¯ i 11 S i P ¯ i ,
P ¯ i 11 ( t f )   =   Q i f ,
P ¯ ˙ i 12   =   P ¯ i 12 A m P ¯ i 11 p = 1 , p i l S p P ¯ p 12 p = 1 , p i l S p P ¯ p 11 T P ¯ i 12
+   A T P ¯ i 12 P ¯ i 11 S i P ¯ i 12 P ¯ i 11 S i P ¯ i 12 + Q ¯ i 12 ,
P ¯ i 12 ( t f )   =   Q ¯ i 12 f ,
  P ¯ ˙ i 13   =   P ¯ i 13 A t P ¯ i 11 p = 1 , p i l S p P ¯ p 13 p = 1 , p i l S p P ¯ p 11 T P ¯ i 13
+   A T P ¯ i 13 P ¯ i 11 S i P ¯ i 13 P ¯ i 11 S i P ¯ i 13 + Q ¯ i 13 ,
P ¯ i 13 ( t f )   =   Q ¯ i 13 f .
Obviously, Equation (34) and (35) are coupled differential Equations (12) and (13) of Theorem 1.
Due to z ¯ = [ z T , x m T , x t T ] T , control strategy u i * = R ¯ i 1 B ¯ i T P ¯ i z ¯ can be expressed as
u i *   =   R i 1 B i T P ¯ i 11 z R i 1 B i T P ¯ i 12 x m R i 1 B i T P ¯ i 13 x t ,
where the first term R i 1 B i T P ¯ i 11 z on the right-hand side is dependent on the state of neighboring agents, the second term R i 1 B i T P ¯ i 12 x m is related to the state of the manned platform, and the third term R i 1 B i T P ¯ i 13 x t is related to the state of the target.
Let b i = P ¯ i 12 x m , i = 1 , 2 , . . . m , taking the derivative of b i , we have
b ˙ i   =   P ¯ ˙ i 12 x m + P ¯ i 12 x ˙ m   =   P ¯ i 12 A m x m ( Q ¯ i 12 P ¯ i 11 S i P ¯ i 12 P ¯ i 11 S i P ¯ i 12 + P ¯ i 12 A m + A T P ¯ i 12 ) x m +   p = 1 , p i l S p P ¯ p 11 T P ¯ i 12 x m + P ¯ i 11 p = 1 , p i l S p P ¯ p 12 x m   =   A T + P ¯ i 11 S i + P ¯ i 11 S i + p = 1 , p i l P ¯ p 11 S p b i Q ¯ i 12 x m + P ¯ i 11 p = 1 , p i l S p b p ,
and b i ( t f ) = P ¯ i 12 ( t f ) x m ( t f ) = Q ¯ i 12 f x m ( t f ) , namely, coupled differential Equations (14) and (15) are valid.
Let c i = P ¯ i 13 x t , i = 1 , 2 , . . . m , taking the derivative of c i , we have
c ˙ i =   P ¯ ˙ i 13 x t + P ¯ i 13 x ˙ t =   P ¯ i 13 A t x t ( Q ¯ i 13 P ¯ i 11 S i P ¯ i 13 P ¯ i 11 S i P ¯ i 13 + P ¯ i 13 A m + A T P ¯ i 13 ) x m +   p = 1 , p i l S p P ¯ p 11 T P ¯ i 13 x t + P ¯ i 11 p = 1 , p i l S p P ¯ p 13 x m =   A T + P ¯ i 11 S i + P ¯ i 11 S i + p = 1 , p i l P ¯ p 11 S p c i Q ¯ i 13 x t + P ¯ i 11 p = 1 , p i l S p c p ,
and c i ( t f ) = P ¯ i 13 ( t f ) x t ( t f ) = Q ¯ i 13 f x t ( t f ) , therefore, coupled differential Equations (16) and (17) hold. □
Remark 2.
Once given the initial state of each agent, communication topology, and weight coefficients, we can calculate P ¯ i 11 , b i , c i at different times by performing backward iterations on the terminal values of the coupled differential equations. The strategy u i * can be obtained by applying the equations’ solutions to (11). Furthermore, we can solve the trajectory of each agent through the state Equation (1).
Remark 3.
The theorem in this paper can also be extended to the case where the target is stationary. Without loss of generality, supposed that the target x t is at the origin, namely, x t = [ 0 , . . . , 0 ] T R 2 n , then, the last term R i 1 B i T P ¯ i 13 x t in the control input (11) is equal to 0. Therefore, strategy u i * only includes the the first two items. On the other hand, x t is a constant value if the target x t is not at the origin, so we can solve strategically similarly.
Remark 4.
It can be seen from Theorem 1 that solving the strategy of each agent does not need the specific form of state equation of the manned platform and the target. We provided A m and A t in the previous section only for the convenience of the system’s expression. In fact, models of the manned platform and the target often cannot be obtained in practical scenarios, we can only observe its states through sensors. Therefore, the strategic solving method only relies on the states x m and x t in this paper. This assumption also corresponds to the situation where the multi-agent system cooperates with the manned platform to perform tasks in the actual manned–unmanned environment.

3.2. Simulation

3.2.1. Simulations under Different Target Trajectories

This subsection shows the results of the simulation examples. Assume there is a team of three agents ( l = 3 ) and the dimension of the state vectors for all agents is 2 (e.g., n = 2 ). Each agent adopts a double integral dynamic model. In order to facilitate computation, R i , R i in cost functions are both taken as identity matrices corresponding to their respective dimensions. Consider the target x t and the manned platform x m , assume that agent 1 and agent 2 can obtain the target’s information while agent 3 cannot observe the target. The communication graph among agents is shown in Figure 1. The corresponding incidence matrix is
D = 1 1 0 1 1 1 1 0 0 0 1 1 .
(1)
A Target Moving along a Fixed Trajectory
Firstly, assume that the target moves in a straight line, and the state of the target is x t = [ 0.5 t , 5 , 1 , 0 ] T . Assume that the manned platform pursues the target with the certain strategy by itself. The agents make decisions based on the state of the target and the manned platform. Let the capture radius to be 0.2 m, sample time to be 0.05 s, and the terminal time t f = 10 s. The weight coefficients are
α i = 1 , τ ¯ 11 = 1 , τ ¯ 12 = 10 , τ ¯ 13 = 5 , τ ¯ 11 f = 1 , τ ¯ 12 f = 10 , τ ¯ 13 f = 5 , τ ¯ 21 = 2 , τ ¯ 22 = 9 , τ ¯ 23 = 6 , τ ¯ 21 f = 2 , τ ¯ 22 f = 9 , τ ¯ 23 f = 6 , τ ¯ 31 = 3 , τ ¯ 32 = 12 , τ ¯ 33 = 0 , τ ¯ 31 f = 3 , τ ¯ 32 f = 12 , τ ¯ 33 f = 0 .
Notice that agent 3 cannot observe the target, therefore τ ¯ 33 = 0 and τ ¯ 33 f = 0 . Set the initial state of each agent x 1 ( 0 ) = [ 2 , 0.2 , 0 , 0 ] T , x 2 ( 0 ) = [ 4.7 , 1 , 0 , 0 ] T , x 3 ( 0 ) = [ 0.2 , 5.5 , 0 , 0 ] T . According to the strategies u i * in Theorem 1, the motion trajectories of each agent can be obtained (as shown in the Figure 2a). It can be seen that the manned platform and agents have captured the target successfully. Figure 2b shows the relative distance between the target and each agent. At the terminal time t f , the distances between each agent and the target are 0.0995 m, 0.1077 m, and 0.1012 m, which are all smaller than the capture radius.
Secondly, assume that the target moves along a sine trajectory, and the state of the target is x t = [ t , s i n ( 0.8 t ) , 1 , 0.8 c o s ( 0.8 t ) ] T . Assume that the manned platform pursues the target with the certain strategy. The weight coefficients are
α i = 1 , τ ¯ 11 = 1 , τ ¯ 12 = 10 , τ ¯ 13 = 5 , τ ¯ 11 f = 1 , τ ¯ 12 f = 10 , τ ¯ 13 f = 5 , τ ¯ 21 = 0.1 , τ ¯ 22 = 9 , τ ¯ 23 = 6 , τ ¯ 21 f = 1 , τ ¯ 22 f = 9 , τ ¯ 23 f = 6 , τ ¯ 31 = 3 , τ ¯ 32 = 12 , τ ¯ 33 = 0 , τ ¯ 31 f = 3 , τ ¯ 32 f = 12 , τ ¯ 33 f = 0 .
Set the initial state of each agent x 1 ( 0 ) = [ 1 , 2 , 0 , 0 ] T , x 2 ( 0 ) = [ 1 , 2 , 0 , 0 ] T , x 3 ( 0 ) = [ 0.2 , 5.5 , 0 , 0 ] T . The simulation results are shown in Figure 3a,b. The distances between each agent and the target are 0.1174 m, 0.1058 m, and 0.1556 m, and the target was captured successfully.
Thirdly, assume that the target moves along a straight line, and the state of the target is x t = [ 0.5 t , 5 , 0 , 0 ] T . Assume that the manned platform pursues the target at a faster speed on the same straight line. The weight coefficients are
α 1 = 0.99 , α 2 = 1.11 , α 3 = 0.005 τ ¯ 11 = 1 , τ ¯ 12 = 0.99 , τ ¯ 13 = 21 , τ ¯ 11 f = 1 , τ ¯ 12 f = 0.99 , τ ¯ 13 f = 21 , τ ¯ 21 = 0.5 , τ ¯ 22 = 0 , τ ¯ 23 = 22 , τ ¯ 21 f = 0.5 , τ ¯ 22 f = 0 , τ ¯ 23 f = 22 , τ ¯ 31 = 10 , τ ¯ 32 = 7.35 , τ ¯ 33 = 0 , τ ¯ 31 f = 10 , τ ¯ 32 f = 7.35 , τ ¯ 33 f = 0 .
Set the initial state of each agent x 1 ( 0 ) = [ 1.5 , 6.5 , 0 , 0 ] T , x 2 ( 0 ) = [ 0.5 , 2 , 0 , 0 ] T , x 3 ( 0 ) = [ 1.2 , 1 , 0 , 0 ] T . The simulation results are shown in Figure 4a,b. The distances between each agent and the target are 0.1993 m, 0.0251 m, and 0.0965 m, and the target was captured successfully.
(2)
Stationary Target
Assume that the target is stationary ( x t = [ 0 , 0 , 0 , 0 ] T ). The manned platform pursues the target with the certain strategy. The weight coefficients are
α i = 1 , τ ¯ 11 = 1 , τ ¯ 12 = 10 , τ ¯ 13 = 5 , τ ¯ 11 f = 1 , τ ¯ 12 f = 10 , τ ¯ 13 f = 5 , τ ¯ 21 = 0.1 , τ ¯ 22 = 9 , τ ¯ 23 = 6 , τ ¯ 21 f = 1 , τ ¯ 22 f = 9 , τ ¯ 23 f = 6 , τ ¯ 31 = 3 , τ ¯ 32 = 12 , τ ¯ 33 = 0 , τ ¯ 31 f = 3 , τ ¯ 32 f = 12 , τ ¯ 33 f = 0 .
Set the initial state of each agent x 1 ( 0 ) = [ 2 , 3.2 , 0 , 0 ] T , x 2 ( 0 ) = [ 4.7 , 2 , 0 , 0 ] T , x 3 ( 0 ) = [ 3 , 5.5 , 0 , 0 ] T . The simulation results are shown in Figure 5a,b. The distances between each agent and the target are 0.0682 × 10 3 m, 0.1479 × 10 3 m, and 0.2865 × 10 3 m, and the target was captured successfully.

3.2.2. The Influence of Different Weight Coefficients on Strategies

Furthermore, consider the case of 5 agents ( l = 5 ), where the corresponding incidence matrix is as follows, assuming that agent 3 and agent 5 cannot observe the target ( τ ¯ 33 = 0 , τ ¯ 53 = 0 ). In the initial state, set the weights τ ¯ i 1 = 1 , τ ¯ i 2 = 1 , τ ¯ i 3 = 1 ( i 3 , 5 ) . We adjust the weight coefficients τ ¯ i 1 , τ ¯ i 2 , τ ¯ i 13 separately to study the impact of different weights on simulation results, then obtain the trajectory (Figure 6a–f) of each agent and distance (Figure 7a–f) between each agent and the target.
D = [ 1 1 0 0 1 0 0 0 0 1 1 0 1 1 0 0 0 0 0 0 0 1 1 0 1 1 1 1 0 0 0 0 0 1 0 1 0 1 1 0 0 0 0 0 0 0 1 0 1 1 ] .
From the Figure 6a–c and Figure 7a–c, it can be seen that the larger the weight coefficient τ ¯ i 1 , the higher the degree of cohesiveness of agents, and agents tend to shorten their distance from their neighbors. From the Figure 6a,d,e and Figure 7a,d,e, it can be seen that increasing the weight coefficient τ ¯ i 2 means that each agent places more emphasis on cooperating with the manned platform, leading to a tendency for agents to narrow their distance from the manned platform in the decision-making process. The entire manned–unmanned system reaches the cohesive state in a shorter time. From the Figure 6a,f and Figure 7a,f, it can be seen that increasing the weight coefficient τ ¯ i 3 makes each agent more inclined to pursue the target and reduces the degree of cohesiveness of the team. Due to the inability of agents 3 and 5 to observe the target, their trajectories’ changes are relatively small, while agents 1, 2, and 4 clearly maintained a close distance from the target at an earlier time.

3.2.3. Comparison with the Target-Pursuit Problem in the Multi-Agent System

In the known studies on target-pursuit problems, most of them only consider the case of multi-agent systems, without considering the coordination of the manned platform. Therefore, we compare such a simulation with this paper; the simulation results of the multi-agent system pursuing the target are shown in the Figure 8. By comparing with the Figure 6 and Figure 7, it can be seen that each agent tends to maintain team cohesion while pursuing the target without the trajectory of a manned platform. In manned–unmanned scenarios, the trajectory of the manned platform is equivalent to the reference trajectory of various agents. Agents that cannot observe the target also have motion references, so the manned–unmanned team tends to converge faster. Relatively speaking, in scenarios without a manned platform, it takes longer for the team to reach the cohesive state. Similarly, agents also need to spend more time reaching the target’s capture distance.

4. Discussion

In practical scenarios, the manned platform makes driving decisions using drivers. The strategies of drivers are complex and variable, and the numerical examples in this paper adopt different simulated trajectories. However, the multi-agent system makes decisions by observing the state of the manned platform x t , so the simulation is effective in verifying Theorem 1.
The theorem in this paper can also be extended to the case where the target adopts an escape strategy. The actual scenario corresponding to this case is that the target can also observe our action states, as the strategy such as u t = R t i 1 B t T P ¯ t z is taken away from our platform. Therefore, the problem transforms into a standard pursuit–evasion game. The strategies are u i * = R i 1 B i T P ˜ i 11 z ˜ R i 1 B i T b ˜ i according to [24] ( z ˜ = [ x 1 T , x 2 T , . . . , x l T , x t T ] T ). At this point, iterative calculation of P ˜ i 11 requires information from A t . If we cannot obtain the target’s model, we can use the reinforcement learning method to approximate the solution [33]. For these situations where adversarial strategies are adopted towards the target, it is more feasible to use reinforcement learning or other approximate approaches since it is difficult to obtain the target’s model. This is one of our future research directions. For future work, we will consider the following in more detail:
  • The environment in this paper is relatively ideal; we will consider the target-pursuit problem when obstacles exist. At the same time, we will consider the situation of multiple targets, so that the strategy can adapt to more practical scenarios.
  • For the situation where the target adopts an escape strategy, we will conduct further analysis, introduce reinforcement learning methods in cases where the model is difficult to obtain, and balance the relationship between computational cost and algorithm performance.
  • The cooperation between multi-agent systems and the manned platform is passive in this paper. In the future, we will introduce trajectory prediction methods for the manned platform, so that agents can make active decisions based on a certain degree of predictive information, thereby improving the efficiency of team task execution.

5. Conclusions

In this paper, we consider a manned–unmanned cooperative decision-making problem in a target-pursuit case. For the target, the manned platform and the multi-agent adopt a hierarchical decision-making approach. The manned platform makes decisions first, followed by agents observing the state of the manned platform and making decisions with interacting information with neighbors. The target-pursuit problem is formulated as a linear quadratic differential game through the directed communication graph. The strategies that be given by solving a system of coupled differential equations iteratively. To make decisions in different pursuit scenarios, we have explored the feasibility of strategies that can adapt to multiple types of targets. We have demonstrated the feasibility of the proposed method through simulation experiments. From the simulation results, the following can be seen: (1) The strategies can enable agents to cooperate with the manned platform to capture a target successfully. (2) The strategies can achieve the pursuit of both the static target and the dynamic target. (3) Different weight coefficients have impacts on algorithm convergence speed, team cohesiveness degree, and target-pursuit time. (4) Compared to the pure unmanned system performing target-pursuit tasks, the manned platform trajectory provides a certain reference for agents, enabling the manned–unmanned collaborative system to complete tasks more efficiently.
However, this approach is proposed in a relatively ideal environment. Additionally, the simulations did not consider more complex platform kinematics models and dynamics models. Therefore, the proposed approach needs to be further experimentally validated in different manned–unmanned collaborative systems, while gradually increasing the complexity of the scenario to ensure the robustness of these strategies.

Author Contributions

Conceptualization, L.H. and W.S.; methodology, L.H. and W.S.; software, L.H. and X.Y.; validation, T.Y., Z.T. and X.A.; formal analysis, L.H. and W.S.; investigation, T.Y. and Z.T.; resources, X.Y.; writing—original draft preparation, L.H.; visualization, L.H.; project administration, L.H. and W.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, X.; Su, H.; Wang, X.; Chen, G. An Overview of Coordinated Control for Multi-agent Systems Subject to Input Saturation. Perspect. Sci. 2016, 7, 133–139. [Google Scholar] [CrossRef]
  2. Ge, S.; Liu, X.; Goh, C.; Xu, L. Formation Tracking Control of Multiagents in Constrained Space. IEEE Trans. Control Syst. Technol. 2016, 24, 992–1003. [Google Scholar] [CrossRef]
  3. Wu, H.; Karkoub, M.; Hwang, C. Mixed Fuzzy Slidingmode Tracking with Backstepping Formation control for Multinonholonomic Mobile Robots subject to Uncertainties. J. Intell. Robot. Syst. 2015, 79, 73–86. [Google Scholar] [CrossRef]
  4. Kawa, A.; Pawlewski, P.; Golinska, P.; Hajdul, M. Cooperative Purchasing of Logistics Services among Manufacturing Companies Based on Semantic Web and Multi-agent System. In Proceedings of the International Conference on Practical Applications of Agents and Multiagent Systems, Salamanca, Spain, 26–28 April 2010. [Google Scholar]
  5. Zuiani, F.; Vasile, M. Multi-agent Collaborative Search with Tchebycheff Decomposition and Monotonic basin Hopping Steps. In Bioinspired Optimization Methods & Their Applications; BIOMA: Bohinj, Slovenia, 2012. [Google Scholar]
  6. Marden, J.; Shamma, J. Game Theory and Distributed Control. In Handbook of Game Theory with Economic Applications; Elsevier: Amsterdam, The Netherlands, 2015; Volume 4, pp. 861–899. [Google Scholar]
  7. Barreiro-Gomez, J. Distributed Formation Control using Population Games. In The Role of Population Games in the Design of Optimization-Based Controllers; Springer: Berlin/Heidelberg, Germany, 2019; pp. 97–110. [Google Scholar]
  8. Qin, J.; Ban, X.; Li, X. Evolutionary Dynamics of Multiagent Formation. In Proceedings of the Chinese Control and Decision Conference, Guilin, China, 17–19 June 2009; pp. 3557–3561. [Google Scholar]
  9. Mcgrew, T.M. Army Aviation Addressing Battlefield Anomalies in Real Time with the Teaming and Collaboration of Manned and Unmanned Aircraft. Geochem. Geophys. Geosyst. 2009. [Google Scholar] [CrossRef]
  10. Dubois, T.; Host, C.; Butt, J.; Clemens, J.A.; Blanton, B. Manned/unmanned Collaborative Mission Concepts with Tiltrotor Aircraft in Support of Maritime Patrol and Search-and-rescue Operations. In Proceedings of the Australian International Aerospace Congress, Melbourne, Australia, 25–28 February 2013. [Google Scholar]
  11. Das, A.; Kol, P.; Lundberg, C.; Doelling, K.; Sevil, H.; Lewis, F. A Rapid Situational Awareness Development Framework for Heterogeneous Manned-Unmanned Teams. In Proceedings of the National Aerospace and Electronics Conference, Dayton, OH, USA, 23–26 July 2018; pp. 417–424. [Google Scholar]
  12. Zhou, H.; Wang, X.; Shan, Y.-Z.; Zhao, Y.; Cui, N. Synergistic Path Planning for Multiple Vehicles based on an Iimproved Particle Swarm Optimization Method. Acta Autom. Sin. 2020, 46, 2670–2676. [Google Scholar]
  13. Garcia, E.; Casbeer, D.; Moll, A.; Pachter, M. Multiple Pursuer Multiple Evader Differential Games. IEEE Trans. Autom. Control 2020, 66, 2345–2350. [Google Scholar] [CrossRef]
  14. Nourbakhsh, I.; Sycara, K.; Koes, M.; Yong, M.; Lewis, M.; Burion, S. Human-Robot Teaming for Search and Rescue. Pervasive Comput. 2005, 4, 72–79. [Google Scholar] [CrossRef]
  15. Benda, M.; Jagannathan, V.; Dodhiawalla, R. On Optimal Cooperation of Knowledge Sources: An Empirical Investigation; Technical Report BCS-G2010-28; The Boeing Company: Seattle, WA, USA, 1986. [Google Scholar]
  16. Kitamura, Y.; Teranishi, K.; Tatsumi, S. Organizational Strategies for Multiagent Real-time Search. In Proceedings of the International Conference on Multi-agent Systems, Kyoto, Japan, 9–13 December 1996; pp. 150–156. [Google Scholar]
  17. Undeger, C.; Polat, F. Multi-agent Real-time Pursuit. Auton. Agents Multi-Agent Syst. 2010, 21, 69–107. [Google Scholar] [CrossRef]
  18. Keshmiri, S.; Payandeh, S. Toward Opportunistic Collaboration in Target Pursuit Problems. In Proceedings of the Autonomous and Intelligent Systems-Second International Conference, Burnaby, BC, Canada, 22–24 June 2011. [Google Scholar]
  19. Pei, H.; Chen, S.; Lai, Q. Multi-target Consensus Circle Pursuit for Multi-agent Systems via a Distributed Multi-flocking Method. Int. J. Syst. Sci. 2015, 47, 3741–3748. [Google Scholar] [CrossRef]
  20. Hamed, O.; Hamlich, M. Improvised Multi-robot Cooperation Strategy for Hunting a Dynamic Target. In Proceedings of the 2020 International Symposium on Advanced Electrical and Communication Technologies (ISAECT), Marrakech, Morocco, 25–27 November 2020; p. 168691. [Google Scholar]
  21. Liu, K.; Zheng, X.; Lin, Y.; Han, L.; Xia, Y. Design of Optimal Strategies for the Pursuit-evasion Problem based on Differential Game. Acta Autom. Sin. 2021, 47, 15. [Google Scholar]
  22. István, H.; Skrzypczyk, K. Robot Team Coordination for Target Tracking using Fuzzy Logic Controller in Game Theoretic Framework. Robot. Auton. Syst. 2009, 57, 75–86. [Google Scholar]
  23. Bhattacharyya, A. Forbidden Minors for a Pursuit Game on Graphs. Sr. Proj. Spring 2013, 408. [Google Scholar]
  24. Talebi, S.; Simaan, M.; Qu, Z. Cooperative, Non-cooperative and Greedy pursuers Strategies in Multiplayer Pursuit-evasion Games. In Proceedings of the IEEE Conference on Control Technology and Applications, Mauna Lani, HI, USA, 27–30 August 2017; pp. 2049–2056. [Google Scholar]
  25. Talebi, S.; Simaan, M. Multi-pursuer Ppursuit-evasion Games under Parameters Uncertainty: A Monte Carlo approach. In Proceedings of the System of Systems Engineering Conference, Waikoloa, HI, USA, 18–21 June 2017; pp. 1–6. [Google Scholar]
  26. Qu, Z.; Simaan, M. A Design of Distributed Game Strategies for Networked Agents. IFAC Proc. Vol. 2009, 42, 270–275. [Google Scholar] [CrossRef]
  27. Liu, R.; Cai, Z. A Novel Approach based on Evolutionary Game Theoretic Model for Multi-player Pursuit Evasion. In Proceedings of the International Conference on Computer, Mechatronics, Control and Electronic Engineering, Changchun, China, 24–26 August 2010; pp. 107–110. [Google Scholar]
  28. Fu, L.; Liu, J.; Meng, G.; Wang, D. Survey Of Manned/Unmanned Air Combat Decision Technology. In Proceedings of the Chinese Control and Decision Conference, Qingdao, China, 23–25 May 2015. [Google Scholar]
  29. Liu, Y.; Zhang, A. Cooperative Task Assignment Method of Manned/unmanned Aerial Vehicle Formation. Xi Tong Gong Cheng Yu Dian Zi Ji Shu/Syst. Eng. Electron. 2010, 32, 584–588. [Google Scholar]
  30. Xu, L.; Pan, X.; Wu, M. Analysis on Mmanned/unmanned Aerial Vehicle Cooperative Operation in Antisubmarine Warfare. Chin. J. Ship Res. 2018, 13, 154–159. [Google Scholar]
  31. Liu, J.; Yuan, S.; Qi, Y.; Ye, W. Research on the Key Technology of the Cooperative Combat System of Manned Vehicle and Unmanned Aerial Vehicle. Ship Electron. Eng. 2012, 32, 1–3. [Google Scholar]
  32. Engwerda, J. LQ Dynamic Optimization and Differential Games; John Wiley & Sons: Hoboken, NJ, USA, 2005. [Google Scholar]
  33. Li, M.; Qin, J.; Ma, Q.; Zheng, W.; Kang, Y. Hierarchical Optimal Synchronization for Linear Systems via Reinforcement Learning: A Stackelberg-Nash Game Perspective. IEEE Trans. Neural Networks Learn. Syst. 2020, 32, 1600–1611. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The communication graph.
Figure 1. The communication graph.
Electronics 12 03630 g001
Figure 2. The simulation of the target moving in a straight line and being pursued by a manned platform with the certain strategy. (a) Trajectories of agents and the manned platform. (b) Distances between the target and each agent.
Figure 2. The simulation of the target moving in a straight line and being pursued by a manned platform with the certain strategy. (a) Trajectories of agents and the manned platform. (b) Distances between the target and each agent.
Electronics 12 03630 g002
Figure 3. Simulation of the target moving along a sine trajectory and being pursued by a manned platform with the certain strategy. (a) Trajectories of agents and the manned platform. (b) Distances between the target and each agent.
Figure 3. Simulation of the target moving along a sine trajectory and being pursued by a manned platform with the certain strategy. (a) Trajectories of agents and the manned platform. (b) Distances between the target and each agent.
Electronics 12 03630 g003
Figure 4. Simulation of the target moving along a straight line and being pursued by a manned platform at a faster speed on the same straight line. (a) Trajectories of agents and the manned platform. (b) Distances between the target and each agent.
Figure 4. Simulation of the target moving along a straight line and being pursued by a manned platform at a faster speed on the same straight line. (a) Trajectories of agents and the manned platform. (b) Distances between the target and each agent.
Electronics 12 03630 g004
Figure 5. Simulation of the target is stationary and being pursued by a manned platform with the certain strategy. (a) Trajectories of agents and the manned platform. (b) Distances between the target and each agent.
Figure 5. Simulation of the target is stationary and being pursued by a manned platform with the certain strategy. (a) Trajectories of agents and the manned platform. (b) Distances between the target and each agent.
Electronics 12 03630 g005
Figure 6. The trajectories of agents under different weight coefficients.
Figure 6. The trajectories of agents under different weight coefficients.
Electronics 12 03630 g006
Figure 7. The distances between the target and each agent under different weight coefficients.
Figure 7. The distances between the target and each agent under different weight coefficients.
Electronics 12 03630 g007
Figure 8. Simulation of the multi-agent system pursuing the target. (a) Trajectories of agents. (b) Distances between the target and each agent.
Figure 8. Simulation of the multi-agent system pursuing the target. (a) Trajectories of agents. (b) Distances between the target and each agent.
Electronics 12 03630 g008
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Han, L.; Song, W.; Yang, T.; Tian, Z.; Yu, X.; An, X. Cooperative Decisions of a Multi-Agent System for the Target-Pursuit Problem in Manned–Unmanned Environment. Electronics 2023, 12, 3630. https://doi.org/10.3390/electronics12173630

AMA Style

Han L, Song W, Yang T, Tian Z, Yu X, An X. Cooperative Decisions of a Multi-Agent System for the Target-Pursuit Problem in Manned–Unmanned Environment. Electronics. 2023; 12(17):3630. https://doi.org/10.3390/electronics12173630

Chicago/Turabian Style

Han, Le, Weilong Song, Tingting Yang, Zeyu Tian, Xuewei Yu, and Xuyang An. 2023. "Cooperative Decisions of a Multi-Agent System for the Target-Pursuit Problem in Manned–Unmanned Environment" Electronics 12, no. 17: 3630. https://doi.org/10.3390/electronics12173630

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop