Next Article in Journal
MDB-YOLO: A Lightweight, Multi-Dimensional Bionic YOLO for Real-Time Detection of Incomplete Taro Peeling
Next Article in Special Issue
Optimal Control for Networked Control Systems with Stochastic Transmission Delay and Packet Dropouts
Previous Article in Journal
Multi-Scale Convolutional Progressive Network for Fine-Grained Vehicle Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Decentralized Control for Interrelated Systems with Asymmetric Information Architecture

1
School of Automation and Electrical Engineering, Linyi University, Linyi 276000, China
2
College of Shandong Traffic Technician, Linyi 276000, China
*
Author to whom correspondence should be addressed.
Electronics 2026, 15(1), 96; https://doi.org/10.3390/electronics15010096
Submission received: 21 November 2025 / Revised: 18 December 2025 / Accepted: 22 December 2025 / Published: 24 December 2025

Abstract

This paper focuses on finite-horizon optimum state feedback control problems for interconnected systems of two players involved with asymmetric one-step delay information. For the finite horizon optimum decentralized control problem, a crucial and adequate condition is derived by using Pontryagin’s maximum principle. Under this framework, player 1 transmits its state and control input data with a one-step delay to the controller of player 2, while player 1’s controller does not have access to the real-time or delayed states and control inputs of player 2, resulting in an asymmetric information structure characterized by a one-step delay Then, the solutions to the forward and backward stochastic difference equations are derived. A target tracking system is given in numerical examples to verify the proposed algorithm.

1. Introduction

The traditional stochastic optimal control problem was first formulated by [1] in the 1960s, with subsequent in-depth investigations conducted by [2,3,4,5,6,7,8,9,10,11,12,13,14]. Specifically, Ref. [7] proposed an outlined Riccati equation to establish the solvability of the stochastic optimal control problem. For delay-free stochastic systems, Ref. [11] addressed the infinite-horizon optimal control problem where the status and control weighting matrices are indefinite. Regarding stochastic linear systems with time delays, Ref. [14] solved the optimal control and equalization problem by proposing a novel Riccati-ZXL difference equation. In summary, the traditional stochastic optimal control problem has been well resolved.
With the growing demand for practical employments—such as networked control systems, formation flight, and target tracking—decentralized control has attracted considerable attention [15,16,17,18,19,20]. A key distinction between traditional stochastic optimal control and decentralized control lies in the feedback information (i.e., state or observation information) available to different controllers: while the feedback information is uniform for all controllers in traditional stochastic optimal control, it varies across controllers in decentralized control, leading to asymmetric information. Generally, for traditional stochastic optimal control, linear controllers are globally optimal, with control gains computable via Riccati equations and the feedback form expressible as a linear function of the conditional expectation of the status [10]. In contrast, for decentralized control, nonlinear strategies may achieve better performance than linear ones [17], and the optimal decentralized control problem may not be convex [18]. Consequently, the methodologies developed for traditional stochastic optimal control cannot be directly applied to decentralized control scenarios.
While the optimal decentralized control problem is intractable, many studies have been devoted to it; see [18,21,22,23,24,25,26,27] and the references therein. Ref. [22] showed that the optimal control strategies were linear by proposing the partial nestedness structure. Ref. [23] introduced a general decentralized model named partial historical sharing structure. Under this structure, assuming that controllers satisfied some linear forms, Ref. [24] presented the optimal controllers in virtue of the common information. Refs. [25,26,27,28] extended this work to networked control systems with special structure. By using dynamic programming and common information, Refs. [26,28] solved the optimal control and equalization problem for systems with a special unreliable communication channel. Under the same special structure, Ref. [27] presented the optimal controllers and stabilization condition with Pontryagin’s maximum principle and algebraic Riccati equations. Recently, Ref. [29] extended the above special structure to the general case of a standard observation equation and obtained linear optimal controllers.
However, the decentralized control problem with time delay is not considered in the above references. Meanwhile, in the context of reliable control under component failures, Ref. [30] proposed a guaranteed-cost LQ control based on algebraic Riccati equations for discrete-time systems with actuator failures, ensuring the system’s asymptotic stability and bounded quadratic cost. Ref. [31] developed a reliable LQG control method that addresses sensor failures via coupled Riccati equations and LMIs, guaranteeing robust stability for systems with partial degradation or outage. These works inspire our research to resolve information asymmetry induced by asymmetric delays while enhancing tolerance to model uncertainties. With the identical combination of delay and sparsity constraints, Ref. [32] presented solutions to the decentralized state feedback control case. Ref. [33] studied applications of heavy-duty vehicle platooning and solved the optimal control problem for chain structures with delayed information. For the finite-horizon case, Ref. [34] presented an optimal state feedback controller for large-scale delayed systems. By requiring plant dynamics to exhibit lower block triangular characteristics, Ref. [35] put forward a decentralized system model with two players that has an asymmetric one-step delayed information pattern. Employing the dynamic programming approach, Ref. [35] obtained the solution of optimal control strategies for the finite horizon. Ref. [5] studied the asymmetric information control involved in different delayed state information, and optimal controllers were derived for the finite-horizon case. Ref. [36] addressed control problem formulations for discrete-time stochastic systems equipped with multiple input paths and input time delays. However, merely one state is involved in [36]. To the best of our knowledge, for general plant dynamics, the optimal control of the asymmetric delayed decentralized problem has not been solved yet.
In this paper, we investigate the finite-horizon optimal state feedback control problem for two-player decentralized interconnected systems with asymmetric one-step delay information—a critical gap in existing research that has not been fully addressed for general coupled dynamics. Unlike previous studies that impose special structural constraints on plant dynamics or assume decoupled control inputs, our work completely removes structural restrictions on system dynamics, enabling the handling of fully coupled state and control input relationships between the two players, which is more aligned with practical interconnected system scenarios such as multi-agent coordination. The main contributions are summarized as follows:
(1) Applying Pontryagin’s maximum principle, we propose an innovative framework specifically designed to derive coupled costate–state backward stochastic difference equations and inter-player equilibrium equations. This framework can explicitly capture the impact of asymmetric delay on information asymmetry, overcoming the limitation of traditional methods that struggle to quantify the correlation between delay and information coupling, and laying a solid theoretical foundation for the subsequent derivation of optimal control strategies.
(2) Building on these equations, we further derive the analytical solution to the forward and backward stochastic difference equations (FBSDEs) and rigorously establish finite-horizon optimal decentralized control strategies through two sets of Riccati equations. We clarify the mild condition for the unique solvability of the strategy, namely the invertibility of relevant matrices, which ensures the theoretical rigor of the method and its feasibility in engineering applications, thereby addressing the issue regarding whether optimal control solutions exist uniquely for coupled systems with asymmetric delays.
(3) A practical unmanned aerial vehicle (UAV) target tracking system is adopted to serve as a case study for confirming the suggested method, fully demonstrating its effectiveness in real-world applications. Experimental results show that the method exhibits stable performance in terms of tracking accuracy and energy efficiency, capable of meeting the dual requirements of practical engineering for control effects and resource consumption, and providing an implementable solution for the optimal control of interrelated systems with asymmetric delays.
Subsequent parts of the paper are arranged in the manner described below. Section 2 presents a study on the finite-horizon optimum decentralized control case. In Section 3, numerical examples are shown on the target tracking system. Relevant proofs are shown in Appendix A in detail.

2. Optimum Control

2.1. Problem Formulation

We study the linear system with a coupled two-player structure:
ψ 1 ( ϱ + 1 ) ψ 2 ( ϱ + 1 )   = D 11 D 12 D 21 D 22 ψ 1 ( ϱ ) ψ 2 ( ϱ ) + G 11 G 12 G 21 G 22 u 1 ( ϱ ) u 2 ( ϱ ) +   λ 1 ( ϱ ) λ 2 ( ϱ ) ,
where ψ 1 ( ϱ ) , u 1 ( ϱ ) and λ 1 ( ϱ ) are the status, control input and system noise of player 1, respectively. ψ 2 ( ϱ ) , u 2 ( ϱ ) and λ 2 ( ϱ ) are the status, control input and system noise of player 2, respectively. ψ 1 ( 0 ) , ψ 2 ( 0 ) , λ 1 ( ϱ ) and λ 2 ( ϱ ) are Gaussian white noises and are featured to not influence each other, with a mean of ( ψ ¯ 1 ( 0 ) , ψ ¯ 2 ( 0 ) , 0 , 0 ) and a covariance of ( σ 1 , σ 2 , Q λ 1 , Q λ 2 ), respectively. The objective for both players is to minimize the following quadratic cost function related to the system (1):
Ω ( ϰ )   = E { ϱ = 0 ϰ i = 1 2 [ ψ i ( ϱ ) Q i t ψ i ( ϱ ) + u i ( ϱ ) R i t u i ( ϱ ) +   ψ i ( ϰ + 1 ) H i ( ϰ + 1 ) ψ i ( ϰ + 1 ) ] } ,
where Q i t , R i t and H i ( ϰ + 1 ) are positive semi-definite matrices for i = 1 , 2 .
At each time instant ϱ , the control actions for the two players obey the system structure information, i.e., the available information for u 1 ( ϱ ) are { ψ 1 ( ϱ ) , ψ 1 ( ϱ 1 ) , , ψ 1 ( 0 ) , u 1 ( ϱ 1 ) , , u 1 ( 0 ) } . Accordingly, u 1 ( ϱ ) is F 1 ( ϱ ) -measurable. The accessible information for u 2 ( ϱ ) is { ψ 1 ( ϱ 1 ) , , ψ 1 ( 0 ) , u 1 ( ϱ 1 ) , , u 1 ( 0 ) , ψ 2 ( ϱ ) , ψ 2 ( ϱ 1 ) , , ψ 2 ( 0 ) , u 2 ( ϱ 1 ) , , u 2 ( 0 ) } . Accordingly, u 2 ( ϱ ) is F 2 ( ϱ ) -measurable. Obviously, the common information for the two players are F c ( ϱ ) = { ψ 1 ( ϱ 1 ) , , ψ 1 ( 0 ) , u 1 ( ϱ 1 ) , , u 1 ( 0 ) } . The whole information for the system is denoted by F ( ϱ ) = { F 1 ( ϱ ) , F 2 ( ϱ ) } .
To make the context much clearer, (1) and (2) are rewritten as
ψ ( ϱ + 1 )   = D ψ ( ϱ ) + G ¯ 1 u 1 ( ϱ ) + G ¯ 2 u 2 ( ϱ ) + e ( ϱ ) ,
Ω ( ϰ )   = E { ϱ = 0 ϰ i = 1 2 [ ψ ( ϱ ) Q t ψ ( ϱ ) + u i ( ϱ ) R i t u i ( ϱ ) +   ψ ( ϰ + 1 ) H ( ϰ + 1 ) ψ ( ϰ + 1 ) ] } ,
where ψ ( ϱ ) = ψ 1 ( ϱ ) ψ 2 ( ϱ ) , D = D 11 D 12 D 21 D 22 , G ¯ 1 = G 11 G 21 , G ¯ 2 = G 12 G 22 , e ( ϱ ) = λ 1 ( ϱ ) λ 2 ( ϱ ) , Q t = Q 1 t 0 0 Q 2 t .
The issue to be dealt with in this section is presented as follows.
Problem 1.
Find the F 1 ( ϱ ) -measurable u 1 ( ϱ ) and F 2 ( ϱ ) -measurable u 2 ( ϱ ) to minimize (4) subject to (3).

2.2. Strategy to Solve Problem 1

Similarly to the procedure of [14], Pontryagin’s maximum principle is applied to (3) and (4) to derive the equations listed below:
η ( ϱ 1 ) = E [ D η ( ϱ ) | F ( ϱ ) ] + Q t ψ ( ϱ ) ,
0 = E [ G ¯ 1 η ( ϱ ) | F 1 ( ϱ ) ] + R 1 t u 1 ( ϱ ) ,
0 = E [ G ¯ 2 η ( ϱ ) | F 2 ( ϱ ) ] + R 2 t u 2 ( ϱ ) ,
η ( ϰ ) = H ( ϰ + 1 ) ψ ( ϰ + 1 ) ,
where η ( ϱ ) represents the costate variable.
Noting the available information for the controllers u 1 ( ϱ ) and u 2 ( ϱ ) , it can be found that u 1 ( ϱ ) is not F 2 ( ϱ ) -measurable and u 2 ( ϱ ) is not F 1 ( ϱ ) -measurable. With the above costate Equations (5)–(7), this leads to an unsolvable mathematical problem where five unknown variables exist in four Equations, (3), (5)–(7). To this end, in view of the common information of u 1 ( ϱ ) and u 2 ( ϱ ) , we define
u 1 c ( ϱ ) = E [ u 1 ( ϱ ) | F c ( ϱ ) ] , u 1 p ( ϱ ) = u 1 ( ϱ ) u 1 c ( ϱ ) ,
u 2 c ( ϱ ) = E [ u 2 ( ϱ ) | F c ( ϱ ) ] , u 2 p ( ϱ ) = u 2 ( ϱ ) u 2 c ( ϱ ) .
The following relationship can be obtained:
E [ u 1 p ( ϱ ) | F c ( ϱ ) ] = 0 , E [ u 2 p ( ϱ ) | F c ( ϱ ) ] = 0 .
In virtue of (9) and (10), we rewrite (3) and (4) as
ψ ( ϱ + 1 ) = D ψ ( ϱ ) + G ¯ u c ( ϱ ) + G ¯ 1 u 1 p ( ϱ ) + G ¯ 2 u 2 p ( ϱ ) + e ( ϱ ) ,
Ω ( ϰ ) = E { ϱ = 0 ϰ i = 1 2 [ ψ ( ϱ ) Q t ψ ( ϱ ) + u c ( ϱ ) R t u c ( ϱ ) +   u i p ( ϱ ) R i t u i p ( ϱ ) + ψ ( ϰ + 1 ) H ( ϰ + 1 ) ψ ( ϰ + 1 ) ] } ,
where G ¯ = G ¯ 1 G ¯ 2 , u c ( ϱ ) = u 1 c ( ϱ ) u 2 c ( ϱ ) , R t = R 1 t 0 0 R 2 t .
As a result of the preceding, we give the next lemma.
Lemma 1.
The costate Equations (5)–(8) can be expressed as
η ( ϱ 1 ) = E [ D η ( ϱ ) | F ( ϱ ) ] + Q t ψ ( ϱ ) ,
0 = E [ G ¯ η ( ϱ ) | F c ( ϱ ) ] + R t u c ( ϱ ) ,
0 = E [ G ¯ 1 η ( ϱ ) | F 1 ( ϱ ) ] E [ G ¯ 1 η ( ϱ ) | F c ( ϱ ) ] + R 1 t u 1 p ( ϱ ) ,
0 = E [ G ¯ 2 η ( ϱ ) | F 2 ( ϱ ) ] E [ G ¯ 2 η ( ϱ ) | F c ( ϱ ) ] + R 2 t u 2 p ( ϱ ) ,
η ( ϰ ) = H ( ϰ + 1 ) ψ ( ϰ + 1 ) .
Proof
By virtue of (9) and (10), applying expectation to both sides of (6) and (7) with F c ( ϱ ) , respectively, we can obtain the following:
0 = E [ G ¯ 1 η ( ϱ ) | F c ( ϱ ) ] + R 1 t u 1 c ( ϱ ) ,
0 = E [ G ¯ 2 η ( ϱ ) | F c ( ϱ ) ] + R 2 t u 2 c ( ϱ ) .
Noting (19) and (20) and observing (12) and (13), (15) can be readily obtained.
Subtracting (19) from (6) and using (9), we have
0 = E [ G ¯ 1 η ( ϱ ) | F 1 ( ϱ ) ] E [ G ¯ 1 η ( ϱ ) | F c ( ϱ ) ] + R 1 t u 1 ( ϱ ) R 1 t u 1 c ( ϱ ) , 0 = E [ G ¯ 1 η ( ϱ ) | F 1 ( ϱ ) ] E [ G ¯ 1 η ( ϱ ) | F c ( ϱ ) ] + R 1 t u 1 p ( ϱ ) ,
which implies that (16) holds. Similarly, subtracting (20) from (7) and applying (10), (17) is valid. This ends the proof. □
It is noted that through the transformation in lemma 1 and the definitions of (9) and (10), the FBSDEs (12) and (14) can be successfully solved as 4 unknown variables exist in 5 Equations: (12), (14)–(17).
Now the following Riccati equations are introduced:
H c ( ϱ ) = D H c ( ϱ + 1 ) D M c ( ϱ ) Υ c ( ϱ ) 1 M c ( ϱ ) + Q t ,
H 1 ( ϱ ) = D H c ( ϱ + 1 ) D M 1 ( ϱ ) Υ 1 ( ϱ ) 1 M 1 ( ϱ ) + Q t ,
H 2 ( ϱ ) = D H 2 ( ϱ + 1 ) D M 2 ( ϱ ) Υ 2 ( ϱ ) 1 M 2 ( ϱ ) + Q t ,
where
Υ c ( ϱ ) = G ¯ H c ( ϱ + 1 ) G ¯ + R , M c ( ϱ ) = G ¯ H c ( ϱ + 1 ) D ,
Υ 1 ( ϱ ) = G ¯ 1 H c ( ϱ + 1 ) G ¯ 1 + R 1 t , M 1 ( ϱ ) = G ¯ 1 H c ( ϱ + 1 ) D ,
Υ 2 ( ϱ ) = G ¯ 2 H 2 ( ϱ + 1 ) G 2 ¯ + R 2 t , M 2 ( ϱ ) = G ¯ 2 H 2 ( ϱ + 1 ) D .
with terminal values H c ( ϰ + 1 ) = H 1 ( ϰ + 1 ) = H 2 ( ϰ + 1 ) = H ( ϰ + 1 ) .
The main results are presented now.
Theorem 1.
A unique solution to Problem 1 exists precisely when Υ c ( ϱ ) , Υ 1 ( ϱ ) and Υ 2 ( ϱ ) are invertible for ϱ = 0 , , ϰ . And the optimal controllers are
u c ( ϱ ) = Υ c ( ϱ ) 1 M c ( ϱ ) ψ ^ c ( ϱ | ϱ ) ,
u 1 p ( ϱ ) = Υ 1 ( ϱ ) 1 M 1 ( ϱ ) ψ ˜ 1 ( ϱ | ϱ ) 0 ,
u 2 p ( ϱ ) = Υ 2 ( ϱ ) 1 M 2 ( ϱ ) 0 ψ ˜ 2 ( ϱ | ϱ ) ,
where
ψ ^ c ( ϱ | ϱ ) = E [ ψ ( ϱ ) | F c ( ϱ ) ] ,
ψ ˜ 1 ( ϱ | ϱ ) = ψ 1 ( ϱ ) E [ ψ 1 ( ϱ ) | F c ( ϱ ) ] = ψ 1 ( ϱ ) ψ ^ 1 c ( ϱ | ϱ ) ,
ψ ˜ 2 ( ϱ | ϱ ) = ψ 2 ( ϱ ) E [ ψ 2 ( ϱ ) | F c ( ϱ ) ] = ψ 2 ( ϱ ) ψ ^ 2 c ( ϱ | ϱ ) .
Then, the optimum cost is
Ω * ( ϰ ) =   E ψ ( 0 ) H c ( 0 ) ψ ¯ ( 0 ) + ψ ( 0 ) H 2 ( 0 ) 0 ψ ˜ 2 ( 0 | 0 ) +   ϱ = 0 ϰ { t r [ ( H 1 ( ϱ ) + H 1 ( ϱ + 1 ) ) Q λ 1 0 0 0 +   H 2 ( ϱ + 1 ) 0 0 0 Q λ 2 ] } ,
where ψ ¯ ( 0 ) = ψ ¯ 1 ( 0 ) ψ ¯ 2 ( 0 ) . Furthermore, the result of the FBSDEs (12) and (14) is
η ( ϱ 1 )   = H c ( ϱ ) ψ ^ c ( ϱ | ϱ ) + H 1 ( ϱ ) ψ ˜ 1 ( ϱ | ϱ ) 0 +   H 2 ( ϱ ) 0 ψ ˜ 2 ( ϱ | ϱ ) .
Proof
Please view Appendix A. □

3. Numerical Examples

Target tracking (TT) systems have drawn significant attention due to their vast applications, such as navigation, collision avoidance, monitoring missions and so on [37].
Consider a simple TT system of two unmanned aerial vehicles (UAVs), which is depicted as in Figure 1. The two unmanned aerial vehicles are named UAV1 and UAV2. In the TTS, UAV1 is the target aerial vehicle with inferior equipments; UAV2 with superior equipments tracks UAV1 within a safe distance and monitors UAV1’s information. But UAV1 cannot observe the information of UAV2 due to its poor equipments when the distances between UAV1 and UAV2 are far enough. To avoid being detected, UAV2 should keep a safe distance from UAV1. Thus, when UAV2 monitors the information of UAV1, delay is unavoidable. The location and velocity for UAV1 and UVA2 are defined as ϵ 1 ( k ) , λ 1 ( k ) and ϵ 2 ( k ) , λ 2 ( k ) , respectively (for simplicity, we suppose that UAVs fly in the straight route; the variables are scalars and the delay is a unit).
Correspondingly, the dynamic equations for UAV1 and UAV2 can be obtained as follows:
ϵ 1 ( ϱ + 1 ) =   d 11 ϵ 1 ( ϱ ) + g 11 v 1 ( ϱ ) + λ 1 ( ϱ ) , ϵ 2 ( ϱ + 1 ) =   ϵ 2 ( ϱ ) + d 21 ϵ 1 ( ϱ ) + g 22 v 2 ( ϱ ) + g 21 v 1 ( ϱ ) + λ 2 ( ϱ ) ,
where λ 1 ( ϱ ) and λ 2 ( ϱ ) present the uncertain flight condition, and d 11 , d 21 , g 11 , g 21 , g 22 are scalar constants. The initial locations ϵ 1 ( 0 ) , ϵ 2 ( 0 ) , λ 1 ( 0 ) and λ 2 ( 0 ) are Gaussian white noises and non-interacting with one another, with a mean of ( μ 1 , μ 2 , 0 , 0 ) and a covariance of ( σ 1 , σ 2 , Q λ 1 , Q λ 2 ) , respectively.
From Figure 1, at time ϱ , UAV2 can observe the locations { ϵ 1 ( ϱ 1 ) , , ϵ 1 ( 0 ) } , the velocities { v 1 ( ϱ 1 ) , , v 1 ( 0 ) } of UAV1, and its own locations { ϵ 2 ( ϱ ) , ϵ 2 ( ϱ 1 ) , , ϵ 1 ( 0 ) } and velocities { v 2 ( ϱ ) , v 2 ( ϱ 1 ) , , v 2 ( 0 ) } . Obviously, the communication channel from UAV1 to UAV2 presents a unit delay. UVA1 can only obtain its own locations { ϵ 1 ( ϱ ) , ϵ 1 ( ϱ 1 ) , , ϵ 1 ( 0 ) } and velocities { v 1 ( ϱ ) , v 1 ( ϱ 1 ) , , v 1 ( 0 ) } . In this scenario, the aim of the TT systems is to keep the distance of UAV1 and UAV2 to the safe distance l and keep the energy cost of UAV1 and UAV2 to the relative minimum. Accordingly, the cost function is of the form
Ω ( ϰ ) = ϱ = 0 ϰ E { [ ϵ 2 ( ϱ ) ϵ 1 ( ϱ ) l ] 2 + R 1 t v 1 ( ϱ ) 2 + R 2 t v 2 ( ϱ ) 2 } ,
where R 1 t > 0 and R 2 t > 0 are the weighting matrices of the energy cost for UAV1 and UAV2, respectively.
This problem can be resolved by using the results obtained in this paper. Firstly, we define ψ 1 ( ϱ ) = ϵ 1 ( ϱ ) , ψ 2 ( ϱ ) = ϵ 2 ( ϱ ) l , u 1 ( ϱ ) = v 1 ( ϱ ) and u 2 ( ϱ ) = v 2 ( ϱ ) . Then, the corresponding linear system and the finite-horizon cost function for the can be defined as
ψ ( ϱ + 1 ) = D ψ ( ϱ ) + G ¯ 1 u 1 ( ϱ ) + G ¯ 2 u 2 ( ϱ ) + λ ( ϱ ) ,
Ω ( ϰ ) =   E { ϱ = 0 ϰ i = 1 2 [ ψ ( ϱ ) Q t ψ ( ϱ ) + R i t u i 2 ( ϱ ) ] } ,
where ψ ( ϱ ) = ψ 1 ( ϱ ) ψ 2 ( ϱ ) , D = d 11 0 d 21 1 , G ¯ 1 = g 11 g 21 , G ¯ 2 = 0 g 22 , e ( ϱ ) = λ 1 ( ϱ ) λ 2 ( ϱ ) , Q t = 1 1 1 1 .
By virtue of Theorem 1, we give the following corollary.
Corollary 1.
The optimal controllers u 1 * ( ϱ ) and u 2 * ( ϱ ) are of the forms
u 1 * ( ϱ ) = 1 0 Υ c ( ϱ ) 1 M c ( ϱ ) ψ ^ c ( ϱ | ϱ )   Υ 1 ( ϱ ) 1 M 1 ( ϱ ) ψ ˜ 1 ( ϱ | ϱ ) 0 ,
u 2 * ( ϱ ) = 0 1 Υ c ( ϱ ) 1 M c ( ϱ ) ψ ^ c ( ϱ | ϱ )   Υ 2 ( ϱ ) 1 M 2 ( ϱ ) 0 ψ ˜ 2 ( ϱ | ϱ ) ,
where ψ ^ c ( ϱ | ϱ ) is the estimation of ψ ( ϱ ) based on { ψ 1 ( ϱ 1 ) , , ψ 1 ( 0 ) } , ψ ˜ 1 ( ϱ | ϱ ) = ψ 1 ( ϱ ) ψ ^ 1 c ( ϱ | ϱ ) is the estimation error of ψ 1 ( ϱ ) , ψ ˜ 2 ( ϱ | ϱ ) = ψ 2 ( ϱ ) ψ ^ 2 c ( ϱ | ϱ ) is the estimation error of ψ 2 ( ϱ ) , and Υ c ( ϱ ) , M c ( ϱ ) , Υ 1 ( ϱ ) , M 1 ( ϱ ) , Υ 2 ( ϱ ) , M 2 ( ϱ ) are as in (24)–(26).
For similarity, we set the safe distance l = 3 , d 11 = d 21 = g 11 = g 21 = g 22 = 1 ,   μ 1 = μ 2 = 0 , σ 1 = σ 2 = Q λ 1 = Q λ 2 = 1 , R 1 t = 1.5 , R 2 t = 2 and N = 100 .
Figure 2 shows the distances between UAV1 and UAV2. It can be observed that the distances between the two UAVs are around the safe distance l = 3 . Figure 3 indicates the velocities of the two UAVs. It can be seen that the velocities of UAV2 are close to those of UAV1.

4. Conclusions

This paper handles the problems of optimal control and stability for decentralized system with one-step asymmetric delay information. Employing Pontryagin’s maximum principle, costate equation and equilibrium equations are given. By virtue of these equations, we obtain the solution to the FBSDEs. Based on this solution, the finite-horizon optimization problem is solved. In view of the finite-horizon optimal cost, the Lyapunov function is defined and the mean-square stabilization conditions for the system are derived in terms of two algebraic Riccati equations. We generalize the proposed algorithm to the examples of a target tracking system to verify the effectiveness of the given results. Future research will focus on extending the proposed framework to multi-player interconnected systems, relaxing the one-step delay constraint to multi-step delays, investigating infinite-horizon scenarios, weakening the matrix invertibility assumption, and accommodating more general non-Gaussian noise distributions to enhance the method’s generality and applicability.

Author Contributions

Conceptualization, Y.W. (Yixing Wang) and X.L.; methodology, Y.W. (Yixing Wang); software, Y.W. (Yirun Wang); validation, Y.W. (Yirun Wang) and B.T.; formal analysis, X.L.; writing—review and editing, X.L. and B.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by Youth Innovation Team Program of Shandong Higher Education Institution under grant 2023KJN049.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Proof of Theorem 1

Necessity: Assuming that Problem 1 admits the unique solutions, we shall show that Υ c ( ϱ ) , Υ 1 ( ϱ ) and Υ 2 ( ϱ ) are invertible for ϱ = 0 , , ϰ and that the optimal controllers are as (27)–(29) by mathematical induction. Noting (18) and (30)–(32), (34) holds for ϱ = ϰ + 1 . Define
Ω ( j ) = E { ϱ = j ϰ i = 1 2 [ ψ ( ϱ ) Q t ψ ( ϱ ) + u c ( ϱ ) R t u c ( ϱ ) +   u i p ( ϱ ) R i t u i p ( ϱ ) + ψ ( ϰ + 1 ) H ( ϰ + 1 ) ψ ( ϰ + 1 ) ] } .
For ϱ = ϰ , since Υ c ( ϰ ) , Υ 1 ( ϰ ) and Υ 2 ( ϰ ) are uncorrelated with ψ ( ϰ ) , Q e , using (12) and letting ψ ( ϰ ) = 0 , Q e = 0 , with (24)–(26), we have the following:
Ω ( ϰ ) = E { i = 1 2 [ ψ ( ϰ ) Q t ψ ( ϰ ) + u c ( ϰ ) R t u c ( ϰ ) +   u i p ( ϰ ) R i t u i p ( ϰ ) + ψ ( ϰ + 1 ) H ( ϰ + 1 ) ψ ( ϰ + 1 ) ] } = E { i = 1 2 [ u c ( ϰ ) ( G ¯ H ( ϰ + 1 ) G ¯ + R ) u c ( ϰ ) +   u i p ( ϰ ) ( R i t + G ¯ i H ( ϰ + 1 ) G ¯ i ) u i p ( ϰ ) ] } = E { i = 1 2 [ u c ( ϰ ) Υ c ( ϰ ) u c ( ϰ ) + u i p ( ϰ ) Υ i ( ϰ ) u i p ( ϰ ) ] } .
The uniqueness of u c ( ϰ ) , u 1 p ( ϰ ) and u 2 p ( ϰ ) implies that the quadratic terms of the controllers are positive for any nonzero controllers. It follows that Υ c ( ϰ ) > 0 , Υ 1 ( ϰ ) > 0 and Υ 2 ( ϰ ) > 0 .
Using (11), (12), (15) and (18), we have the following:
0 = E [ G ¯ H ( ϰ + 1 ) ψ ( ϰ + 1 ) | F c ( ϰ ) ] + R t u c ( ϰ ) = G ¯ H ( ϰ + 1 ) [ D ψ ^ c ( ϰ | ϰ ) + G ¯ u c ( ϰ ) ] + R t u c ( ϰ ) .
Therefore, the optimal u c ( ϰ ) satisfies
u c ( ϰ ) = Υ c ( ϰ ) 1 M c ( ϰ ) ψ ^ c ( ϰ | ϰ ) .
By virtue of (11), (12), (16), (18) and (31), it yields
0   = E [ G ¯ 1 H ( ϰ + 1 ) ψ ( ϰ + 1 ) | F 1 ( ϰ ) ]   E [ G ¯ 1 H ( ϰ + 1 ) ψ ( ϰ + 1 ) | F c ( ϰ ) ] + R 1 t u 1 p ( ϰ ) = G ¯ 1 H ( ϰ + 1 ) ×   D 11 ψ 1 ( ϰ ) + G 11 u 1 ( ϰ ) D 21 ψ 1 ( ϰ ) + D 22 ψ ^ 2 c ( ϰ | ϰ ) + G 21 u 1 ( ϰ ) + G 22 u 2 c ( ϰ )   G ¯ 1 H ( ϰ + 1 ) ×   D 11 ψ ^ 1 c ( ϰ | ϰ ) + G 11 u 1 c ( ϰ ) D 21 ψ ^ 1 c ( ϰ | ϰ ) + D 22 ψ ^ 2 c ( ϰ | ϰ ) + G 21 u 1 c ( ϰ ) + G 22 u 2 c ( ϰ ) + R 1 t u 1 p ( ϰ ) = G ¯ 1 H ( ϰ + 1 ) D 11 ψ ˜ 1 ( ϰ | ϰ ) + G 11 u 1 p ( ϰ ) D 21 ψ ˜ 1 ( ϰ | ϰ ) + G 21 u 1 p ( ϰ ) + R 1 t u 1 p ( ϰ ) = G ¯ 1 H ( ϰ + 1 ) D ψ ˜ 1 ( ϰ | ϰ ) 0 +   ( G ¯ 1 H ( ϰ + 1 ) G ¯ 1 + R 1 t ) u 1 p ( ϰ ) .
Thus, the optimum u 1 p ( ϰ ) is as follows
u 1 p ( ϰ ) = Υ 1 ( ϰ ) 1 M 1 ( ϰ ) ψ ˜ 1 ( ϰ | ϰ ) 0 .
By applying (12), (17), (18) and (32), we obtain
0   = E [ G ¯ 2 H ( ϰ + 1 ) ψ ( ϰ + 1 ) | F 2 ( ϰ ) ]   E [ G ¯ 2 H ( ϰ + 1 ) ψ ( ϰ + 1 ) | F c ( ϰ ) ] + R 2 t u 2 p ( ϰ ) = G ¯ 2 H ( ϰ + 1 ) ×   D 11 ψ ^ 1 c ( ϰ | ϰ ) + G 11 u 1 c ( ϰ ) D 21 ψ ^ 1 c ( ϰ | ϰ ) + D 22 ψ ^ 2 c ( ϰ | ϰ ) + G 21 u 1 c ( ϰ ) + G 22 u 2 c ( ϰ ) G ¯ 2 H ( ϰ + 1 ) ×   D 11 ψ ^ 1 c ( ϰ | ϰ ) + G 11 u 1 c ( ϰ ) D 21 ψ ^ 1 c ( ϰ | ϰ ) + D 22 ψ ^ 2 c ( ϰ | ϰ ) + G 21 u 1 c ( ϰ ) + G 22 u 2 c ( ϰ ) +   R 2 t u 2 p ( ϰ ) = G ¯ 2 H ( ϰ + 1 ) 0 D 22 ψ ˜ 2 ( ϰ | ϰ ) + G 22 u 2 p ( ϰ ) + R 2 t u 2 p ( ϰ ) = G ¯ 2 H ( ϰ + 1 ) D 0 ψ ˜ 2 ( ϰ | ϰ ) +   ( G ¯ 2 H ( ϰ + 1 ) G ¯ 2 + R 2 t ) u 2 p ( ϰ ) .
Thus, we have the optimum u 2 p ( ϰ ) , which is as follows:
u 2 p ( ϰ ) = Υ 2 ( ϰ ) 1 M 2 ( ϰ ) 0 ψ ˜ 2 ( ϰ | ϰ ) .
Using (12), (15), (18), (A1)–(A3), we obtain
η ( ϰ 1 ) = E [ D H ( ϰ + 1 ) ψ ( ϰ + 1 ) | F ( ϰ ) ] + Q t ψ ( ϰ ) = D H ( ϰ + 1 ) [ D ψ ( ϰ ) + G ¯ u c ( ϰ ) + G ¯ 1 u 1 p ( ϰ ) + G ¯ 2 u 2 p ( ϰ ) ] +   Q t ψ ( ϰ ) = D H ( ϰ + 1 ) D ψ ( ϰ ) M c ( ϰ ) Υ c ( ϰ ) 1 M c ( ϰ ) ψ ^ 1 c ( ϰ | ϰ ) ψ ^ 2 c ( ϰ | ϰ )   M 1 ( ϰ ) Υ 1 ( ϰ ) 1 M 1 ( ϰ ) ψ ˜ 1 ( ϰ | ϰ ) 0   M 2 ( ϰ ) Υ 2 ( ϰ ) 1 M 2 ( ϰ ) 0 ψ ˜ 2 ( ϰ | ϰ ) + Q t ψ ( ϰ ) = [ D H ( ϰ + 1 ) D M c ( ϰ ) Υ c ( ϰ ) 1 M c ( ϰ ) + Q t ] ×   ψ ^ 1 c ( ϰ | ϰ ) ψ ^ 2 c ( ϰ | ϰ ) +   [ D H ( ϰ + 1 ) D M 1 ( ϰ ) Υ 1 ( ϰ ) 1 M 1 ( ϰ ) + Q t ] ×   ψ ˜ 1 ( ϰ | ϰ ) 0
+   [ D H ( ϰ + 1 ) D M 2 ( ϰ ) Υ 2 ( ϰ ) 1 M 2 ( ϰ ) + Q t ] ×   0 ψ ˜ 2 ( ϰ | ϰ ) = H c ( ϰ ) ψ ^ 1 c ( ϰ | ϰ ) ψ ^ 2 c ( ϰ | ϰ ) + H 1 ( ϰ ) ψ ˜ 1 ( ϰ | ϰ ) 0 +   H 2 ( ϰ ) 0 ψ ˜ 2 ( ϰ | ϰ ) ,
which means that (34) holds.
To accomplish the mathematical induction, letting random l satisfying 0 l N , suppose that Υ c ( ϱ ) , Υ 1 ( ϱ ) , Υ c ( ϱ ) are invertible and that the optimal u c ( ϱ ) , u 1 p ( ϱ ) , u 2 p ( ϱ ) and η ( ϱ 1 ) are (27)–(29) and (34), respectively, for all ϱ l + 1 .
By virtue of (12), (14)–(17), we have
E [ ψ ( ϱ ) η ( ϱ 1 ) ψ ( ϱ + 1 ) η ( ϱ ) ] = E { ψ ( ϱ ) E [ D η ( ϱ ) | F ( ϱ ) ] + ψ ( ϱ ) Q t ψ ( ϱ )   [ ψ ( ϱ ) D u c ( ϱ ) G ¯ u 1 p ( ϱ ) G ¯ 1 u 2 p ( ϱ ) G ¯ 2 ] η ( ϱ ) } = E { ψ ( ϱ ) Q t ψ ( ϱ ) u c ( ϱ ) E [ G ¯ η ( ϱ ) | F c ( ϱ ) ]   u 1 p ( ϱ ) E [ G ¯ 1 η ( ϱ ) | F 1 ( ϱ ) ] u 2 p ( ϱ ) E [ G ¯ 2 η ( ϱ ) | F 2 ( ϱ ) ] } = E [ ψ ( ϱ ) Q t ψ ( ϱ ) + u c ( ϱ ) R t u c ( ϱ ) +   u 1 p ( ϱ ) R 1 t u 1 p ( ϱ ) + u 2 p ( ϱ ) R 2 t u 2 p ( ϱ ) ] .
Aggregating data from ϱ = κ + 1 to ϱ = ϰ on both sides of the above equation and using (18), we obtain
E [ ψ ( κ + 1 ) η ( κ ) ψ ( ϰ + 1 ) H ( ϰ + 1 ) ψ ( ϰ + 1 ) ] = ϱ = κ + 1 ϰ E [ ψ ( ϱ ) Q t ψ ( ϱ ) + u c ( ϱ ) R t u c ( ϱ ) +   u 1 p ( ϱ ) R 1 t u 1 p ( ϱ ) + u 2 p ( ϱ ) R 2 t u 2 p ( ϱ ) ] .
It follows that
Ω ( κ ) = E [ ψ ( κ ) Q t ψ ( κ ) + u c ( κ ) R t u c ( κ ) + u 1 p ( κ ) R 1 t u 1 p ( κ ) +   u 2 p ( κ ) R 2 t u 2 p ( κ ) ] + ϱ = κ + 1 ϰ E [ ψ ( ϱ ) Q t ψ ( ϱ ) +   u c ( ϱ ) R t u c ( ϱ ) + u 1 p ( ϱ ) R 1 t u 1 p ( ϱ ) + u 2 p ( ϱ ) R 2 t u 2 p ( ϱ ) +   ψ ( ϰ + 1 ) H ( ϰ + 1 ) ψ ( ϰ + 1 ) ] = E [ ψ ( κ ) Q t ψ ( κ ) + u c ( κ ) R t u c ( κ ) + u 1 p ( κ ) R 1 t u 1 p ( κ ) +   u 2 p ( κ ) R 2 t u 2 p ( κ ) ] + E [ ψ ( κ + 1 ) η ( κ ) ] .
From (34), for ϱ = κ + 1 , setting ψ ( κ ) = 0 , e ( κ ) = 0 , we have
η ( κ ) = H c ( κ + 1 ) D 11 ψ 1 ( κ ) + G 11 u 1 ( κ ) D 21 ψ 1 ( κ ) + D 22 ψ ^ 2 c ( κ | κ ) + G 21 u 1 ( κ ) + G 22 u 2 c ( κ ) +   H 1 ( κ + 1 ) λ 1 ( κ ) 0 + H 2 ( κ + 1 ) ×   0 D 22 ψ ˜ 2 ( κ | κ ) + G 22 u 2 p ( κ ) + λ 2 ( κ ) = H c ( κ + 1 ) G ¯ u c ( κ ) + H c ( κ + 1 ) G ¯ 1 u 1 p ( κ ) + H 2 ( κ + 1 ) G ¯ 2 u 2 p ( κ ) .
Substituting (A5) into (A4) and setting ψ ( κ ) = 0 , e ( κ ) = 0 , we obtain
Ω ( N ) = E { u c ( κ ) [ R + G ¯ H c ( κ + 1 ) G ¯ ] u c ( κ ) +   u 1 p ( κ ) [ R 1 t + G ¯ 1 H c ( κ + 1 ) G ¯ 1 ] u 1 p ( κ ) +   u 2 p ( κ ) [ R 2 t + G ¯ 2 H 2 ( κ + 1 ) G ¯ 2 ] u 2 p ( κ ) } = u c ( κ ) Υ c ( κ ) u c ( κ ) + u 1 p ( κ ) Υ 1 ( κ ) u 1 p ( κ ) + u 2 p ( κ ) Υ 2 ( κ ) u 2 p ( κ )
It can be seen that Υ c ( κ ) > 0 , Υ 1 ( κ ) > 0 and Υ 2 ( κ ) > 0 based on the uniqueness of the optimal controllers. The proofs of the optimal u c ( κ ) , u 1 p ( κ ) , u 2 p ( κ ) and η ( κ 1 ) are analogous to the above procedures of κ = N . Thus, we omit them here. The essential proof is completed.
Sufficiency: Assuming that Υ c ( ϱ ) , Υ 1 ( ϱ ) and Υ 2 ( ϱ ) are invertible for ϱ = 0 , , N , we will demonstrate that Problem 1 can be solved unparalleled. Define
V ( ϱ , ψ ( ϱ ) ) = E { ψ ( ϱ ) [ H c ( ϱ ) ψ ^ c ( ϱ | ϱ ) + H 1 ( ϱ ) ψ ˜ 1 ( ϱ | ϱ ) 0 +   H 2 ( ϱ ) 0 ψ ˜ 2 ( ϱ | ϱ ) ] } .
By applying (12), (21)–(26) and (30)–(32), it yields
V ( ϱ , ψ ( ϱ ) ) V ( ϱ + 1 , ψ ( ϱ + 1 ) ) = E   { ψ ( ϱ ) [ H c ( ϱ ) D H c ( ϱ + 1 ) D + M c ( ϱ ) Υ c ( ϱ ) 1 M c ( ϱ ) ] ×   ψ ( ϱ ) ψ ( ϱ ) M c ( ϱ ) Υ c ( ϱ ) 1 M c ( ϱ ) ψ ( ϱ )   2 ψ ^ c ( ϱ | ϱ ) D H c ( ϱ + 1 ) G ¯ u c ( ϱ ) u c ( ϱ ) [ Υ c ( ϱ ) R ] u c ( ϱ )   2 ψ ˜ 1 ( ϱ | ϱ ) 0 D H c ( ϱ + 1 ) G ¯ 1 u 1 p ( ϱ )   u 1 p ( ϱ ) [ Υ 1 ( ϱ ) R 1 t ] u 1 p ( ϱ )   2 0 ψ ˜ 2 ( ϱ | ϱ ) D H 2 ( ϱ + 1 ) G ¯ 2 u 2 p ( ϱ )   u 2 p ( ϱ ) [ Υ 2 ( ϱ ) R 2 t ] u 2 p ( ϱ )   0 ψ ˜ 2 ( ϱ | ϱ ) M 2 ( ϱ ) Υ 2 ( ϱ ) 1 M 2 ( ϱ ) 0 ψ ˜ 2 ( ϱ | ϱ ) +   ψ ˜ 1 ( ϱ | ϱ ) ψ ˜ 2 ( ϱ | ϱ ) M c ( ϱ ) Υ c ( ϱ ) 1 M c ( ϱ ) ψ ˜ 1 ( ϱ | ϱ ) ψ ˜ 2 ( ϱ | ϱ )   ψ ˜ 1 ( ϱ | ϱ ) 0 M 1 ( ϱ ) Υ 1 ( ϱ ) 1 M 1 ( ϱ ) ψ ˜ 1 ( ϱ | ϱ ) 0 }
=   E [ ψ ( ϱ ) Q t ψ ( ϱ ) + u ( c ϱ ) R t u c ( ϱ ) + u 1 p ( ϱ ) R 1 t u 1 p ( ϱ ) +   u 2 p ( ϱ ) R 2 t u 2 p ( ϱ ) ] E { [ u c ( ϱ ) + Υ c ( ϱ ) 1 M c ( ϱ ) ψ ^ c ( ϱ | ϱ ) ] ×   Υ c ( ϱ ) [ u c ( ϱ ) + Υ c ( ϱ ) 1 M c ( ϱ ) ψ ^ c ( ϱ | ϱ ) ] +   [ u 1 p ( ϱ ) + Υ 1 ( ϱ ) 1 M 1 ( ϱ ) ψ ˜ 1 ( ϱ | ϱ ) 0 ] ×   Υ 1 ( ϱ ) [ u 1 p ( ϱ ) + Υ 1 ( ϱ ) 1 M 1 ( ϱ ) ψ ˜ 1 ( ϱ | ϱ ) 0 ] +   [ u 2 p ( ϱ ) + Υ 2 ( ϱ ) 1 M 2 ( ϱ ) 0 ψ ˜ 2 ( ϱ | ϱ ) ] ×   Υ 2 ( ϱ ) [ u 2 p ( ϱ ) + Υ 2 ( ϱ ) 1 M 2 ( ϱ ) 0 ψ ˜ 2 ( ϱ | ϱ ) ] }   E { λ 1 ( ϱ ) 0 [ H 1 ( ϱ ) + H 1 ( ϱ + 1 ) ] λ 1 ( ϱ ) 0 +   0 λ 2 ( ϱ ) H 2 ( ϱ + 1 ) 0 λ 2 ( ϱ ) } .
Adding ϱ = 0 to ϱ = ϰ on two sides of the above equation, we have
Ω ( ϰ ) =   E ψ ( 0 ) H c ( 0 ) ψ ¯ ( 0 ) + ψ ( 0 ) H 2 ( 0 ) 0 ψ ˜ 2 ( 0 | 0 ) +   ϱ = 0 ϰ { t r [ ( H 1 ( ϱ ) + H 1 ( ϱ + 1 ) ) Q λ 1 0 0 0 +   H 2 ( ϱ + 1 ) 0 0 0 Q λ 2 ] } +   ϱ = 0 ϰ E { [ u c ( ϱ ) + Υ c ( ϱ ) 1 M c ( ϱ ) ψ ^ c ( ϱ | ϱ ) ] ×   Υ c ( ϱ ) [ u c ( ϱ ) + Υ c ( ϱ ) 1 M c ( ϱ ) ψ ^ c ( ϱ | ϱ ) ] +   [ u 1 p ( ϱ ) + Υ 1 ( ϱ ) 1 M 1 ( ϱ ) ψ ˜ 1 ( ϱ | ϱ ) 0 ] ×   Υ 1 ( ϱ ) [ u 1 p ( ϱ ) + Υ 1 ( ϱ ) 1 M 1 ( ϱ ) ψ ˜ 1 ( ϱ | ϱ ) 0 ] +   [ u 2 p ( ϱ ) + Υ 2 ( ϱ ) 1 M 2 ( ϱ ) 0 ψ ˜ 2 ( ϱ | ϱ ) ] ×   Υ 2 ( ϱ ) [ u 2 p ( ϱ ) + Υ 2 ( ϱ ) 1 M 2 ( ϱ ) 0 ψ ˜ 2 ( ϱ | ϱ ) ] } .
Since Υ c ( ϱ ) , Υ 1 ( ϱ ) and Υ 2 ( ϱ ) are definitely positive, optimal controllers satisfy (27)–(29) and the optimal performance index is (33). This completes the sufficiency part of the proof.

References

  1. Wonham, W.M. On a matrix Riccati equation of stochastic control. SIAM J. Control 1968, 6, 681–697. [Google Scholar] [CrossRef]
  2. Lu, M.; Liu, S.; Luo, Y.; Qi, Q.; Sun, Y.; Li, J. Modeling and Output Feedback Tracking Control of Self-Developed Autonomous Underwater Vehicle. Unmanned Systems 2025, 13, 1639–1660. [Google Scholar] [CrossRef]
  3. Wang, X.; Tan, C.P.; Wang, Y.; Qi, Q.; Wang, X. Adaptive Interval Observer-Based Fault-Tolerant Control for a 3-DOF Helicopter Without Angular Velocity Measurement. IEEE Trans. Control Syst. Technol 2025, 33, 2476–2482. [Google Scholar] [CrossRef]
  4. Qi, Q.; Xie, L.; Zhang, H.; Liang, X. Jointly Optimal Local and Remote Controls for Networked Multiple Systems with Multiplicative Noises and Unreliable Uplink Channels. IEEE Trans. Autom. Control 2025, 70, 1054–1067. [Google Scholar] [CrossRef]
  5. Qi, Q.; Xie, L.; Zhang, H. Optimal Control for Stochastic Systems with Multiple Controllers of Different Information Structures. IEEE Trans. Autom. Control 2021, 66, 4160–4175. [Google Scholar] [CrossRef]
  6. Kleinman, D. Optimal stationary control of linear systems with control-dependent noise. IEEE Trans. Autom. Control 1969, 14, 673–677. [Google Scholar] [CrossRef]
  7. Bismut, J.M. Linear quadratic optimal stochastic control with random coefficients. SIAM J. Control 1976, 14, 419–444. [Google Scholar] [CrossRef]
  8. Zou, L.; Wang, Z.; Hu, J.; Dong, H. Ultimately Bounded Filtering Subject to Impulsive Measurement Outliers. IEEE Trans. Autom. Control 2022, 67, 304–319. [Google Scholar] [CrossRef]
  9. Zou, L.; Wang, Z.; Hu, J.; Liu, Y.; Liu, X. Communication-protocol-based analysis and synthesis of networked systems: Progress, prospects and challenges. Int. J. Syst. Sci. 2021, 52, 3013–3034. [Google Scholar] [CrossRef]
  10. Zhang, H.; Xu, J. Control for Itô stochastic systems with input delay. IEEE Trans. Autom. Control 2016, 62, 350–365. [Google Scholar] [CrossRef]
  11. Rami, M.A.; Zhou, X.Y. Linear matrix inequalities, Riccati equations, and indefinite stochastic linear quadratic controls. IEEE Trans. Autom. Control 2000, 45, 1131–1143. [Google Scholar] [CrossRef]
  12. Yong, J.; Zhou, X.Y. Stochastic Controls: Hamiltonian Systems and HJB Equations; Springer: New York, NY, USA, 1999; Volume 43. [Google Scholar]
  13. Sinopoli, B.; Schenato, L.; Franceschetti, M.; Poolla, K.; Jordan, M.I.; Sastry, S.S. Kalman filtering with intermittent observations. IEEE Trans. Autom. Control 2004, 49, 1453–1464. [Google Scholar] [CrossRef]
  14. Zhang, H.; Li, L.; Xu, J.; Fu, M. Linear quadratic regulation and stabilization of discrete-time systems with delay and multiplicative noise. IEEE Trans. Autom. Control 2015, 60, 2599–2613. [Google Scholar] [CrossRef]
  15. Lessard, L.; Lall, S. An algebraic approach to the control of decentralized systems. IEEE Trans. Control Netw. Syst. 2014, 1, 308–317. [Google Scholar] [CrossRef]
  16. Fanti, M.P.; Mangini, A.M.; Pedroncelli, G.; Ukovich, W. A decentralized control strategy for the coordination of AGV systems. Control Eng. Pract. 2018, 70, 86–97. [Google Scholar] [CrossRef]
  17. Lipsa, G.M.; Martins, N.C. Optimal memoryless control in Gaussian noise: A simple counterexample. Automatica 2011, 47, 552–558. [Google Scholar] [CrossRef]
  18. Bamieh, B.; Voulgaris, P.G. A convex characterization of distributed control problems in spatially invariant systems with communication constraints. Syst. Control Lett. 2005, 54, 575–583. [Google Scholar] [CrossRef]
  19. Torabi, A.; Zarei, J.; Razavi-Far, R.; Saif, M. Decentralized Resilient Output-Feedback Control Design for Networked Control Systems Under Denial-of-Service. IEEE Syst. J. 2022, 16, 5620–5629. [Google Scholar] [CrossRef]
  20. He, W.; Li, S.; Ahn, C.K.; Guo, J.; Xiang, Z. Global Decentralized Control of p-Normal Large-Scale Nonlinear Systems Based on Sampled-Data Output Feedback. IEEE Syst. J. 2021, 15, 3540–3548. [Google Scholar] [CrossRef]
  21. Tsitsiklis, J.; Athans, M. On the complexity of decentralized decision making and detection problems. IEEE Trans. Autom. Control 1985, 30, 440–446. [Google Scholar] [CrossRef]
  22. Ho, Y.C.; Chu, K.C. Team decision theory and information structures in optimal control problems–Part I. IEEE Trans. Autom. Control 1972, 17, 15–22. [Google Scholar] [CrossRef]
  23. Nayyar, A.; Mahajan, A.; Teneketzis, D. Decentralized stochastic control with partial history sharing: A common information approach. IEEE Trans. Autom. Control 2013, 58, 1644–1658. [Google Scholar] [CrossRef]
  24. Mahajan, A.; Nayyar, A. Sufficient statistics for linear control strategies in decentralized systems with partial history sharing. IEEE Trans. Autom. Control 2015, 60, 2046–2056. [Google Scholar] [CrossRef]
  25. Liang, X.; Xu, J.; Zhang, H. Optimal control and stabilization for networked control systems with asymmetric information. IEEE Trans. Control Netw. Syst. 2020, 7, 1355–1365. [Google Scholar] [CrossRef]
  26. Ouyang, Y.; Asghari, S.M.; Nayyar, A. Optimal infinite horizon decentralized networked controllers with unreliable communication. IEEE Trans. Autom. Control 2020, 66, 1778–1785. [Google Scholar] [CrossRef]
  27. Liang, X.; Xu, J. Control for networked control systems with remote and local controllers over unreliable communication channel. Automatica 2018, 98, 86–94. [Google Scholar] [CrossRef]
  28. Asghari, S.M.; Ouyang, Y.; Nayyar, A. Optimal local and remote controllers with unreliable uplink channels. IEEE Trans. Autom. Control 2018, 64, 1816–1831. [Google Scholar] [CrossRef]
  29. Liang, X.; Qi, Q.; Zhang, H.; Xie, L. Decentralized control for networked control systems with asymmetric information. IEEE Trans. Autom. Control 2022, 67, 2076–2083. [Google Scholar] [CrossRef]
  30. Yang, G.H.; Wang, J.L.; Soh, Y.C. Reliable LQG control with sensor failures. IEE Proc. Control Theory Appl. 2000, 147, 433–439. [Google Scholar] [CrossRef]
  31. Yang, Y.; Yang, G.H.; Soh, Y.C. Reliable control of discrete-time systems with actuator failure. IEE Proc. Control Theory Appl. 2000, 147, 428–432. [Google Scholar] [CrossRef]
  32. Lamperski, A.; Lessard, L. Optimal decentralized state-feedback control with sparsity and delays. Automatica 2015, 58, 143–151. [Google Scholar] [CrossRef]
  33. Feyzmahdavian, H.R.; Alam, A.; Gattami, A. Optimal distributed controller design with communication delays: Application to vehicle formations. In Proceedings of the 2012 IEEE 51st IEEE Conference on Decision and Control (CDC), Maui, HI, USA, 10–13 December 2012; pp. 2232–2237. [Google Scholar]
  34. Wang, Y.; Xiong, J.; Ho, D.W. Globally optimal state-feedback LQG control for large-scale systems with communication delays and correlated subsystem process noises. IEEE Trans. Autom. Control 2019, 64, 4196–4201. [Google Scholar] [CrossRef]
  35. Nayyar, N.; Kalathil, D.; Jain, R. Optimal decentralized control with asymmetric one-step delayed information sharing. IEEE Trans. Control Netw. Syst. 2016, 5, 653–663. [Google Scholar] [CrossRef]
  36. Wang, H.; Zhang, H.; Li, L.; Fu, M. LQR and stabilization for discrete-time systems with multiplicative noises and input delays. IEEE Trans. Autom. Control 2023, 69, 3515–3530. [Google Scholar] [CrossRef]
  37. Mazor, E.; Averbuch, A.; Bar-Shalom, Y.; Dayan, J. Interacting multiple model methods in target tracking: A survey. IEEE Trans. Aerosp. Electron. Syst. 1998, 34, 103–123. [Google Scholar] [CrossRef]
Figure 1. The TT system of UAV1 and UAV2.
Figure 1. The TT system of UAV1 and UAV2.
Electronics 15 00096 g001
Figure 2. Distance between UAV1 and UAV2.
Figure 2. Distance between UAV1 and UAV2.
Electronics 15 00096 g002
Figure 3. Velocities of UAV1 and UAV2.
Figure 3. Velocities of UAV1 and UAV2.
Electronics 15 00096 g003
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Y.; Wang, Y.; Tan, B.; Li, X.; Liang, X. Decentralized Control for Interrelated Systems with Asymmetric Information Architecture. Electronics 2026, 15, 96. https://doi.org/10.3390/electronics15010096

AMA Style

Wang Y, Wang Y, Tan B, Li X, Liang X. Decentralized Control for Interrelated Systems with Asymmetric Information Architecture. Electronics. 2026; 15(1):96. https://doi.org/10.3390/electronics15010096

Chicago/Turabian Style

Wang, Yixing, Yirun Wang, Boqun Tan, Xinghua Li, and Xiao Liang. 2026. "Decentralized Control for Interrelated Systems with Asymmetric Information Architecture" Electronics 15, no. 1: 96. https://doi.org/10.3390/electronics15010096

APA Style

Wang, Y., Wang, Y., Tan, B., Li, X., & Liang, X. (2026). Decentralized Control for Interrelated Systems with Asymmetric Information Architecture. Electronics, 15(1), 96. https://doi.org/10.3390/electronics15010096

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop