Next Article in Journal
Comparative Evaluation of Symmetrical Titanium and Polyetheretherketone (PEEK) Hollow Structures for Mandibular Reconstruction: Strength, Geometry, and Biomechanical Performance
Previous Article in Journal
Formulas for the Number of Weak Homomorphisms from Paths to Rectangular Grid Graphs
Previous Article in Special Issue
A New Adaptive Control Design of Permanent Magnet Synchronous Motor Systems with Uncertainties
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimal Control of Stochastic Dynamic Systems with Semi-Markov Parameters

by
Taras Lukashiv
1,2,*,
Igor V. Malyk
2,
Ahmed Abdelmonem Hemedan
1 and
Venkata P. Satagopam
1
1
Luxembourg Centre for Systems Biomedicine, University of Luxembourg, L-4370 Belvaux, Luxembourg
2
Department of Mathematical Problems of Control and Cybernetics, Yuriy Fedkovych Chernivtsi National University, 58000 Chernivtsi, Ukraine
*
Author to whom correspondence should be addressed.
Symmetry 2025, 17(4), 498; https://doi.org/10.3390/sym17040498
Submission received: 17 February 2025 / Revised: 10 March 2025 / Accepted: 21 March 2025 / Published: 26 March 2025
(This article belongs to the Special Issue Symmetry in Optimal Control and Applications)

Abstract

:
This paper extends classical Markov switching models. We introduce a generalized semi-Markov switching framework in which the system dynamics are governed by an Itô stochastic differential equation. Of note, the optimal control synthesis problem is formulated for stochastic dynamic systems with semi-Markov parameters. Further, a system of ordinary differential equations is derived to characterize the Bellman functional and the corresponding optimal control. We investigate the case of linear dynamics in detail, and propose a closed-form solution for the optimal control law. A numerical example is presented to illustrate the theoretical results.
MSC:
60K15; 93E03; 93E20

1. Introduction

Optimal control synthesis plays a critical role in managing the dynamics of controlled objects or processes. This is particularly relevant when the dynamic behavior is described by Ito stochastic differential equations (SDEs). The choice of Ito SDE is due to their prevalence as applied models and their widespread use in biology, chemistry, telecommunications, etc. The main type of random perturbations in stochastic differential equations is the symmetric Wiener process, the trajectories of which have the reflection property, which very well characterizes the symmetry of this process. The symmetry of the solution of the stochastic differential equation is also well reflected in the example of this work, where the symmetry of the solution trajectories is traced relative to the so-called averaged trajectory, and optimal control does not violate this property. The paper [1] considers the optimal control of a linear SDE of general type perturbed by a random process with independent increments and a quadratic quality functional. It is shown that under certain conditions the optimal control is linear and can be determined by solving an additional vector quadratic problem. The optimal solution of the deterministic problem with feedback is obtained. In paper [2], the theory of optimal control for stochastic systems whose performance is measured by the exponent of an integral form is developed. In paper [3], the problem of the existence of optimal control for stochastic systems with a nonlinear quality functional is solved. The main model in the work [4] is the linear autonomous SDE in the form
g ( x , u ) = A 0 x + B 0 u + w k ( A 1 x + B 1 u ) ,
where w k i . i . d with E w k = 0 , V a r ( w k ) = 1 . Notably, the results of the paper are devoted to the quadratic quality functional and necessary and sufficient conditions for the stability of the systems (1), and the stability conditions are formulated in terms of Riccati-type jumping operator properties. Note that the linear case (1) is most often considered as a first approximation of the dynamics of a real phenomenon, since the optimal control in this case can be found in closed form and the approximations of the optimal control can be compared with the exact values.
In article [5], the main attention is focused on SDE with external switches. In this paper, a number of important remarks are made on the calculation of the product (infinitesimal) operator of a random process given by a SDE with Markov switches.
In article [6], SDE with external switches and Poisson perturbations are discussed. Using the Belman equation, sufficient conditions for the existence of optimal control for a general quality function are found, and a closed form of optimal control is found for the linear case with a quadratic quality function.
The transition from difference equations to differential equations gives rise to a number of complications in the synthesis of optimal control, since it leads to a transition from solving more complex dynamical systems such as the analog of the Lyapunov equation for nonautonomous systems. Further, the presence of random variables of different structures leads to the use of the infinitesimal operator [7,8], the calculation of which depends on the nature of the random variables that affect the calculation of the infinitesimal operator.
In this paper, we will focus on the use of semi-Markov processes [8] as the main source of external random disturbances. Noteworthy, the use of semi-Markov processes significantly extends the range of application of theoretical results for many applied problems, since the condition of exponential distribution of the time spent in the state θ n for continuous Markov processes ξ ( t ) ,   t 0 , is very strict for many applied problems. A description based on a Markov process will be incorrect if, for example, an assumption is made about the minimum time spent in a particular state P ( θ n > t m i n ) = 1 , where t m i n is the minimum time spent in the state. In this case, the use of semi-Markov processes is more efficient, since it allows us to control the properties of the residence time, the size of the jump ξ ^ n and the interdependence between them based on the semi-Markov kernel Q ( y , B , t ) [8]. The use of Markov processes greatly simplifies the study of the system, since it requires estimation only on the basis of the intensities of the Markov process λ ( y ) , and usually when studying systems with external Markov disturbances it is assumed that
0 < λ m i n λ ( y ) λ m a x < , y Y .
In this study, we will not focus on the asymptotic properties of the random process x ( t ) , t 0 , but will consider the problem of synthesizing optimal control on a finite interval [ 0 , T ] , so the article will not consider additional conditions on the semi-Markov process that ensure its ergodicity [7,9,10].
Note that the study of optimal control with semi-Markovian processes has been receiving attention since the last quarter of the 20th century, when it became clear that the exponential distribution cannot be considered as the basic distribution for the transition time in many cases. For example, in the work [11], the optimization of the system in the presence of random shocks, which in turn are described by semi-Markov processes, is considered. The work gives an example of optimal replacement of the levels of the system and constructs a replacement algorithm for a countable set of states of the semi-Markov process. Models for finding the optimal general repair policy are considered in [12], and the semi-Markov process in this case describes construction errors, and the policy change is based on minimizing the budget on the one hand and fulfilling all construction requirements on the other. In [13], the authors considered the possibility of using semi-Markov models to solve the output-feedback control problem for general semi-Markov jump systems with a memory-dynamic event-triggered protocol. The authors of this article proposed a description of the system using a differential equation with a semi-Markovian component for the derivative of the process, i.e., the dynamics of the process will depend integrally on the dynamics of the semi-Markovian process. The main statements of this work provide sufficient conditions for the existence of optimal control for systems with semi-Markov parameters and an example of finding optimal control is given. A somewhat similar result for the case of asynchronous control was investigated in work [14]. The paper also provides several sufficient conditions for stochastic controllability of the system and considers a number of model examples with implementations of the semi-Markov process.
As noted in many works, the use of semi-Markov processes is of great practical importance, since the use of these models allows for a more accurate description of real processes. For example, in work [12] the authors used a semi-Markov process to find the optimal policy in construction, in work [15] the authors used semi-Markov processes with the corresponding problem of optimizing a portfolio of securities. Practically, this explains the use of Markov models with exponential time of stay in states is a good theoretical model, but for practical reasons it has a number of limitations associated with the moments of switching. As noted in paper [12], the policy change occurs at discrete moments of time and has a uniform distribution over a finite horizon, so the use of the Markov model in this case is inappropriate.
Instead, the main attention will be paid to the elementary study of the dynamics of the main process x ( t ) , t [ 0 , T ] , at a fixed value of the external perturbation ξ ( t ) = y , y Y , which allows us to more effectively study the optimal control based on the methods proposed for Markov processes. Section 2 of the article describes the problem statement and justifies the existence of a solution that is unique up to stochastic equivalence. In Section 3, sufficient conditions for the existence of a solution to the optimal control problem are established, and in Section 4, a method for finding it in the general case is described. Section 5 is devoted to the synthesis of optimal control for a linear system with a quadratic performance index, and in Section 6, a system of Riccati differential equations is presented for finding the matrix functions that form optimal control. The last section, Section 7, contains a model example that illustrates the stages of optimal control synthesis under the condition that the time spent in a state is determined by θ n = t m i n + P o i s s ( λ ( y ) ) , i.e., the time spent in a state is discrete and with probability 1 greater than t m i n .

2. Problem Statement

Consider a stochastic dynamical system defined on a probabilistic basis ( Ω , F , F , P ) [7,16]. The system is governed by the stochastic differential equation (SDE):
d x ( t ) = a ( t , ξ ( t ) , x ( t ) , u ( t ) ) d t + b ( t , ξ ( t ) , x ( t ) , u ( t ) ) d w ( t ) , t R + ,
with initial conditions
x ( 0 ) = x 0 R m , ξ ( 0 ) = y Y .
Here, ξ ( t ) , t 0 , is a semi-Markov process with values in Y , which is characterised by generator [9]
Q ( y , B , t ) = P ( y , B ) F y ( t ) , y Y , B β Y , t 0 ,
where P ( y , B ) specifies the distribution of jumps of the nested Markov chain ξ ^ n , n 0 [8], F y ( t ) is the conditional distribution of time spent in the state y. The representation of the generator based on the splitting (4) greatly simplifies the main calculations in proving the main theoretical results of this paper, but allows for generalization to the general case of a semi-Markov kernel Q ( y , B , t ) ; x : 0 , + × Ω R m ; w ( t ) , t 0 , is a standard Wiener process; a control u ( t ) : = u ( t , x ( t ) ) : [ 0 , T ] × R m R m is m-measured function from the set of admissible controls U [17]; and the processes w and ξ are independent [7,16].
As in works [16,18], we assume that the measured by the set of variables functions a : R + × Y × R m × U R m and b : R + × Y × R m × U R m satisfy the boundedness condition and the Lipschitz condition
| a ( t , y , x , u ) | 2 + | b ( t , y , x , u ) | 2 C ( 1 + | x | 2 ) ,
| a ( t , y , x 1 , u ) a ( t , y , x 2 , u ) | 2 + | b ( t , y , x 1 , u ) b ( t , y , x 2 , u ) | 2 L | x 1 x 2 | 2 , x 1 , x 2 R m .
The semi-Markov process ξ has the following effect on the trajectories of the process x. Suppose that on the interval [ t 0 = 0 , t 1 ) , the process ξ takes the value y 1 . Then, the movement will occur due to the system
d x ( t ) = a ( t , y 1 , x ( t ) , u ( t ) ) d t + b ( t , y 1 , x ( t ) , u ( t ) ) d w ( t ) , t R + , x ( 0 ) = x 0 R m .
According to [10,16], if the conditions (5) and (6) are met, the system (7) has on the interval [ 0 , t 1 ) a unique solution with a finite second moment up to stochastic equivalence.
Then, at time t 1 , the value of the process ξ changes: y 1 y 2 . Then, on the interval [ t 1 , t 2 ) , the motion will occur due to the system
d x ( t ) = a ( t , y 2 , x ( t ) , u ( t ) ) d t + b ( t , y 2 , x ( t ) , u ( t ) ) d w ( t ) , t R + , x ( t 1 ) = x ( t 1 ) R m .
According to [10,16], if conditions (5) and (6) are met, system (8) has a unique solution with a finite second moment on the interval [ t 1 , t 2 ) .
Thus, conditions (5) and (6) guarantee the existence of a unique solution to the Cauchy problem, (2), and (3) on the interval [ 0 , + ) = k = 0 [ t k , t k + 1 ) , 0 = t 0 , t 1 , , t k , , the second moment for which is finite. Thus, for the existence of a solution on [ 0 , ) , we will assume that the semi-Markov process is defined on [ 0 , ) , i.e., T > 0 : P ( ξ ( T ) exist ) = 1 .

3. Sufficient Conditions for Optimality

We introduce a sequence of functions v k ( t , x ) : [ t k , T ] × R m R 1 , k 0 , and class V : = { v k ( t , x ) C 1 , 2 ( R + × R m ) } .
On the functions v k ( t , x ) V , we define the weak infinitesimal operator (WIO)
L v k ( t , x ) = lim Δ 0 + 1 Δ { E v k ( t + Δ , x ( t + Δ , t k , y , u ) ) v k ( t , x ) } ,
where x ( t ) = x ( t , t k , y , u ) is the strong solution (2) on the interval t [ t k , t k + 1 ) with control u = u k U . Notice that WIO is natural extension of the right-hand side of the ordinary differential equation
d x d t = f ( t , x )
in the case of stochastic systems.
The problem of optimal control is to find a control u k 0 , k 0 , from the set U that minimizes the scalar quality functional [17]
I ( u k 0 ) = I u k 0 ( t , x ) = inf u U k = 0 N E t k , x k F ( x ( T ) ) + t T G ( s , x ( s ) , u ( s ) ) d s ,
for some fixed T > 0 , F ( x ) 0 ,   G ( t , x , u ) 0 , E t , x { f } = E { f / x ( t ) = x } , and N = m a x { k : t k < T } .
To obtain sufficient conditions for optimality, we need to prove several auxiliary statements.
Lemma 1.
Let:
(1) 
There exists a unique solution to the Cauchy problem (2), and (3), whose second moment is finite for each t;
(2) 
A sequence of functions v k : [ 0 , T ] × R m R 1 , k 0 , from class V exists;
(3) 
For s [ t k , T ) , the WIO L v k ( s , x ) is defined on the solutions of (2) and (3).
Then, for t ˜ 1 , t ˜ 2 [ t 0 , T ] the following equality holds
E t k , x k v k ( t ˜ 2 , x ( t ˜ 2 ) ) E t k , x k v k ( t ˜ 1 , x ( t ˜ 1 ) ) = t ˜ 1 t ˜ 2 E t k , x k L v k ( s , x ( s ) ) d s .
Proof. 
For the Markov process x ( t , t k , y , u ) , with respect to the σ -algebra F t ˜ 1 , t ˜ 2 constructed on the interval [ t ˜ 1 , t ˜ 2 ] , t ˜ 1 , t ˜ 2 [ t k , T ] , the following Dynkin formula [7] holds
E t k , x k v k ( t ˜ 1 + τ ( t ˜ 2 ) , x ( t ˜ 1 + τ ( t ˜ 2 ) , t k , y , u ) )
= v k ( t , x ) + E t k , x k t 0 τ r ( t ˜ 2 ) L v k ( t ˜ 1 + τ , x ( t ˜ 1 + τ , t k , y , u ) ) d τ ,
where τ r = inf t { | x ( t ) | > r } and τ r ( t ˜ 2 ) = min { τ r , t ˜ 2 } .
If lim r τ r ( t ˜ 2 ) = t ˜ 2 , then for the solution of the problems (2) and (3), we obtain the following equality
E t k , x k v k ( t ˜ 1 + t ˜ 2 , x ( t ˜ 1 + t ˜ 2 , t k , y , u ) ) = v k ( t , x ) + E t k , x k t 0 t ˜ 2 L v k ( t ˜ 1 + τ , x ( t ˜ 1 + τ , t k , y , u ) ) d τ .
Similarly, write the Dynkin formula on the interval [ t 0 , t 1 ] and, subtracting it from (12), we obtain (11). Lemma 1 is proved. □
Lemma 2.
Let:
(1) 
Conditions (1) and (2) of Lemma 1 are fulfilled;
(2) 
For t [ t 0 , T ] , v k V , k 0 has the sense of the equation
L v k ( t , x ) + G ( t , x , u k ( t , x ) ) = 0 ,
with a boundary condition
v k ( T , x ) = 0 , k 0 ,
where L v k ( t , x ) is the WIO defined by (9).
Then, v k V , k 0 , can be written
v k ( t , x ) = E t k , x k t T G ( s , x , u ( s , x ) ) d s , t [ t k , T ] .
Proof. 
Consider the solution x ( t ) R m of the problems (2) and (3) for t [ t k , T ] , constructed according to the corresponding initial condition.
We integrate (13) with respect to s from t k to T and calculate the mathematical expectation. We obtain
E t k , x k t k T L v k ( s , x ) d s + E t k , x k t k T G ( s , x , u k ( s , x ) ) d s = 0 .
According to Lemma 1, there exists a first term (16) that is equal to the increment (11):
E t k , x k t k T L v k ( s , x ) d s = E t k , x k v k ( T , x ( T ) ) E t k , x k v k ( t , x ) = v k ( t , x ) ,
where E t k , x k v k ( T , x ( T ) ) = v k ( T , x ( T ) ) = 0 according to (14), and E t k , x k v k ( t , x ( t ) ) = v k ( t , x ) . Thus,
E t k , x k t k T L v k ( s , x ) d s = v k ( t , x ) .
Substituting (17) into (16), we obtain the statement of Lemma 2. □
Theorem 1.
Let:
(1) 
There exists a unique solution to the Cauchy problem (2), and (3), whose second moment is finite for each t;
(2) 
There exists a sequence of functions v k V , k 0 , and an optimal control u k 0 U , k 0 , that satisfy the equation
L v k ( t , x ) + G ( t , x , u k 0 ( t , x ) ) = 0 ,
with a boundary condition
v k ( T , x ) = F ( x ( T ) ) ;
(3) 
t [ 0 , T ] , u k U , k 0 , the following inequality holds
L v k ( t , x ) + G ( t , x , u k ( t ) ) 0 ,
where L = L ( t , x , u ) is the WIO (9) on the solutions of (2), (3).
Then, the control u k 0 is optimal, for t [ t 0 , T ] ,
I u k 0 ( t , x ) = inf u U I u ( t , x ) = v k ( t , x ) .
The sequence of functions v k ( t , x ) is called the control cost or Bellman function, and Equation (18) can be written as the Bellman equation
inf u U L ( t , x k , u ) v k ( t , x k ) + G ( t , x k , u ) = 0 .
Proof. 
An optimal control is also an admissible control. Therefore, there exists a solution x ( t k , t 0 , y , u k 0 ) for which (18) takes the form
L v k ( t , x ( t , t k , y , u k 0 ) ) + G ( t , x ( t , t k , y , u k 0 ) , u k 0 ( t , x ) ) = 0 ,
where u k 0 is taken at the point ( t , x ( t , t k , y , u k 0 ) ) , t [ t k , t k + 1 ) .
We integrate (23) from t to T, calculate the mathematical expectation, and, taking into account (19), we obtain
v k ( t , x ) = E t k , x k F ( x ( T ) ) + t T L v k ( τ , x ( τ , t 0 , y , u k 0 ) ) d τ .
Now, let u k = u k ( t ) be an arbitrary control from the class of admissible controls U. Then, according to condition (3) of the theorem, the following inequality holds u k U , k 0 ,
L v k ( τ , x ( τ , t 0 , y , u k ) ) + G ( τ , x ( τ , t 0 , y , u k ) , u k ( τ ) ) 0 ,
We integrate (24) over τ [ t 0 , T ] , and calculate the mathematical expectation E at fixed τ and initial value x. Accounting for Lemmas 1 and 2, we obtain
v k ( t , x ( t , t 0 , y , u k 0 ) )
E t k , x k F ( x ( T , u k ) ) + t k T G ( τ , x ( τ , t 0 , y , u k ) , u k ( τ ) ) d τ = v k ( t , x ( t , t 0 , y , u k ) ) .
And this, in fact, is the definition of optimal control u k 0 ( t , x ) in the sense of minimizing the quality functional I u ( t , x ) . Theorem 1 is proved. □

4. General Solution of the Optimal Control Problem

The following theorem holds.
Theorem 2.
The weak infinitesimal operator on the solutions of the Cauchy problem (2), (3) on the functions v k ( t , x ) V is calculated by the formula
L v k ( t , x ) = v k ( t , x ) t + ( v k ( t , x ) , a ( t , y , x , u ) )
+ 1 2 S p ( b T ( t , y , x , u ) · 2 v k ( t , x ) · b ( t , y , x , u ) )
+ i j N R m v k ( t , z ; j ) p i j ( t , z | x ) d z v k ( t , x ; i ) q i j ,
where ( · , · ) is a scalar product, ( v k ) : = ( ( v k / x 1 ) , . . . , ( v k / x m ) ) T , ( 2 v k ) : = [ 2 v k / x i x j ] i , j = 1 m , k 0 , < < T > > is a transpose sign, S p is a matrix trace,
q i j = lim t 0 Q ( i , j , t ) t = P ( i , j ) F x ( 0 ) ,
p i j ( t , z | x ) is a conditional density of the process x ( t ) for ξ ( t ) = i , ξ ( t ) = j :
P x ( t ) [ z , z + d z ] | x ( t ) = x , ξ ( t ) = i , ξ ( t ) = j = p i j ( t , z | x ) d z + o ( d z ) .
In the last term in the formula (25) in the function v k ( t , x ; i ) , the last argument indicates the value of the semi-Markov process at a given time, i.e., ξ ( t ) = y i .
Proof. 
The first three terms can be obtained in the same way as in [19]. We derive the form of the last term, which is related to the semi-Markov parameter ξ . To do this, consider the hypotheses H i j = { ξ ( t ) = i , ξ ( t ) = j } , i , j Y . Considering the hypotheses H i j , we determine that the last term will correspond to the hypotheses H i j , i j , i.e.,
E v k ( t + Δ , x ( t + Δ , t k , y , u ) ) = i , j Y E v k ( t + Δ , x ( t + Δ , t k , y , u ) ) I H i j
= i = j Y E v k ( t + Δ , x ( t + Δ , t k , y , u ) ) I H i j + i j Y E v k ( t + Δ , x ( t + Δ , t k , y , u ) ) I H i j .
For the first term, in the absence of a change in the state of the semi-Markov process, the conditions imposed on the coefficients of the initial equation allow us to write
lim Δ 0 1 Δ i = j Y E v k ( t + Δ , x ( t + Δ , t k , y , u ) ) I H i j v k ( t , x )
= v k ( t , x ) t + ( v k ( t , x ) , a ( t , y , x , u ) ) + 1 2 S p ( b T ( t , y , x , u ) · 2 v k ( t , x ) · b ( t , y , x , u ) ) .
Further, when the state of the semi-Markov process ξ ( t ) changes at time t, taking into account the form of the conditional transition probability P x ( t ) [ z , z + d z ] | x ( t ) = x , ξ ( t ) = i , ξ ( t ) = j (26), we obtain
lim Δ 0 1 Δ i j Y E v k ( t + Δ , x ( t + Δ , t k , y , u ) ) I H i j v k ( t , x )
= i j N R m v k ( t , z ; j ) p i j ( t , z | x ) d z v k ( t , x ; i ) q i j .
Theorem 2 is proved. □
The first equation for finding v k 0 ( t , x ) , k 0 , can be obtained by substituting (25) into (18). We obtain
v k 0 ( t , x ; i ) t + v k 0 ( t , x ; i ) x T · a ( t , y i , x , u )
+ 1 2 S p b T ( t , y i , x , u ) · 2 v k 0 ( t , x ; i ) x 2 · b ( t , y i , x , u )
+ i j N R m v k 0 ( t , x ; j ) p i j ( t , z | x ) d z v k 0 ( t , x ; i ) q i j + G ( t , x , u ) = 0
with a boundary condition
v k 0 ( T , x ) = F ( x ) .
The second equation, for finding the optimal control u k 0 ( t , x ) , can be obtained from (27) by differentiating respect to u, since u = u k 0 , k 0 , delivers the minimum of the left-hand side of (27):
v k 0 ( t , x ; i ) x T · a ( t , y i , x , u ) u
+ 1 2 S p b ( t , y i , x , u ) u T · 2 v k 0 ( t , x ; i ) x 2 · b ( t , y i , x , u )
+ b T ( t , y i , x , u ) · 2 v k 0 ( t , x ; i ) x 2 · b ( t , y i , x , u ) u
+ G ( t , x , u ) u | u = u k 0 = 0 ,
where ( a / u ) is a m × m -Jacobian, which consists of the elements { ( a n / u s ) , n = 1 , m ¯ , s = 1 , m ¯ } ( ( b / u ) —similarly), ( G / u ) ( ( G / u 1 ) , , ( G / u r ) ) , k 0 .
The solution of the systems (27) and (29) is a very difficult task even with the use of modern computer technologies. Therefore, it is advisable to consider a simplified version of the problems (2), (3), and (10); namely, a linear system with a quadratic quality functional.

5. Synthesis of Optimal Control for a Linear Stochastic System

When studying complex systems, the main approach is to break these systems into simpler components, thus considering simple idealized objects that will reveal many of their properties [20]. The uniqueness and exceptional importance of the linear case for stochastic systems, as an “idealization” approach, is explained by a number of reasons.
First, in this case, as already mentioned, there is always the possibility of obtaining the form of the control function in a closed form. As well as the application of various methods for studying the properties of the solution, such as stability, etc.
Secondly, in theoretical articles, such as ours, the most important aspect is the analysis and justification of the stochastic components. They can be most clearly observed in linear systems—the solutions of which are sufficiently smooth—and the reader will clearly understand that all fluctuations in the solution are due exclusively to stochastic components.
Thirdly, it is linear systems that are usually systems of first approximation, which under certain conditions are a reflection of the behavior (stability or instability) of the solution of the general system.
Consider the problem of optimal control for a linear stochastic dynamical system given by the stochastic differential equation
d x ( t ) = [ A ( t , ξ ( t ) ) x ( t ) + B ( t , ξ ( t ) ) u ( t ) ] d t + σ ( t , ξ ( t ) ) x ( t ) d w ( t ) , t R + ,
with the initial conditions
x ( 0 ) = x 0 R m , ξ ( 0 ) = y Y .
Here A , B , σ are piecewise continuous integral-matrix functions of appropriate dimension.
The problem of optimal control for the systems (30) and (31) is to find a control u k 0 , k 0 , from the set of admissible controls U so that minimizes the quadratic quality functional
I ( u k ) = I u k ( t , x ) = k = 0 N ¯ E t k , x k x T ( T ) M 0 ( ξ ( t ) ) x ( T )
+ t T [ u T ( s ) M 1 ( s , ξ ( s ) ) u ( s ) + x T ( s ) M 2 ( s , ξ ( s ) ) x ( s ) ] d s ,
M 1 ( t , ξ ( t ) ) is a uniformly positive definite with respect to the t [ 0 , T ] m × m matrix, and M 0 ( ξ ( t ) ) and M 2 ( t , ξ ( t ) ) are non-negatively definite m × m matrices. To simplify, we introduce the notation
A i ( t ) = A ( t , y i ) , B i ( t ) = B ( t , y i ) , σ i ( t ) = σ ( t , y i ) ,
M 0 i = M 0 ( y i ) , M 1 i ( t ) = M 1 ( t , y i ) , M 2 i ( t ) = M 2 ( t , y i ) .
Theorem 3.
The optimal control for the problem (30)–(32) has the following form:
u k 0 ( t , x ; i ) = M 1 i 1 ( t ) B i T ( t ) P i ( t ) x ( t ) ,
where the non-negatively definite m × m matrix P i ( t ) = P ( t , ξ ( t ) ) defines the Bellman functional
v k 0 ( t , x ; i ) = x T ( t ) P i ( t ) x ( t ) + g ( t ) , v k 0 ( T , x ; i ) = x T ( t k ) M 0 i x ( t k ) .
Here, g is a non-negative scalar function.
Proof. 
Bellman’s equation for (30)–(32) has the form
inf u U [ L v k 0 ( t , x ; i ) + u k T ( t , x ; i ) M 1 i ( t ) u k ( t , x ; i ) + x T ( t ) M 2 i ( t ) x ( t ) ) ] = 0 ,
where
L v k 0 ( t , x ; i ) = v k 0 ( t , x ; i ) t + g ˙ ( t ) + [ A i ( t ) x ( t ) + B i ( t ) u k ( t , x ; i ) ] T v k 0 ( t , x ; i )
+ 1 2 S p ( σ i T ( t ) · 2 v i 0 ( t , x ; i ) · σ i ( t ) )
+ i j N v k 0 ( t , x ; j ) v k 0 ( t , x ; i ) q i j .
Substitute (36) into (35):
v k 0 ( t , x ; i ) t + g ˙ ( t ) + [ A i ( t ) x ( t ) + B i ( t ) u k ( t , x ; i ) ] T v k 0 ( t , x ; i )
+ 1 2 S p ( σ i T ( t ) · 2 v k 0 ( t , x ; i ) · σ i ( t ) )
+ i j N v k 0 ( t , x ; j ) v k 0 ( t , x ; i ) q i j + u k T ( t , x ; i ) M 1 i ( t ) u k ( t , x ; i ) + x T ( t ) M 2 i ( t ) x ( t ) ) = 0 .
The form of the optimal control is obtained by differentiating (37), since u k ( t , x ; i ) = u k 0 ( t , x ; i ) minimizes the left side (37):
u k 0 ( t , x ; i ) = 1 2 M 1 i 1 ( t ) B i T ( t ) v k 0 ( t , x ; i ) ,
where
v k 0 ( t , x ; i ) = 2 P i ( t ) x ( t ) .
Therefore,
u k 0 ( t , x ; i ) = M 1 i 1 ( t ) B i T ( t ) P i ( t ) x ( t ) .
Theorem 3 is proved. □

6. Construction of the Bellman Equation

Substituting (33) and (34) into Equation (35), we obtain the following equation for t [ t k , t k + 1 ) :
x T ( t ) d P i ( t ) d t x ( t ) + g ˙ ( t )
+ 2 [ A i ( t ) x ( t ) B i ( t ) M 1 i 1 ( t ) B i T ( t ) x ( t ) ] P i ( t ) x ( t ) + S p ( σ i T ( t ) P i ( t ) σ i ( t ) )
+ i j N x T ( t ) P j ( t ) x ( t ) x T ( t ) P i ( t ) x ( t ) q i j
+ [ M 1 i 1 ( t ) B i T ( t ) P i ( t ) x ( t ) ] T M 1 i ( t ) M 1 i 1 ( t ) B i T ( t ) P i ( t ) x ( t ) + x T ( t ) M 2 i ( t ) x ( t ) = 0 .
Equating to zero the quadratic form in x and expressions that do not depend on x, and taking into account the matrix equality 2 x T P i x = x T ( P i A i + A i T P i ) x , we obtain a system of differential equations for finding the matrices P i ( t ) , t [ t k , t k + 1 ) , k 0 :
d P i ( t ) d t + A i T ( t ) P i ( t ) + P i ( t ) A i ( t ) B i ( t ) M 1 i 1 ( t ) B i T ( t ) P i ( t ) + i j N P j ( t ) P i ( t ) q i j
+ [ M 1 i 1 ( t ) B i T ( t ) P i ( t ) ] T B i T ( t ) P i ( t ) + M 2 i ( t ) = 0 ,
S p ( σ i T ( t ) P i ( t ) σ i ( t ) ) + g ˙ ( t ) = 0 ,
with a boundary condition
P i ( T ) = M 0 i .
Thus, we can formulate the following theorem.
Theorem 4.
If the quality functional for the systems (30) and (31) is (32), and the control cost is (34), then the system of differential equations for finding the matrices P i ( t ) , t [ t k , t k + 1 ) , k 0 , has the form (39)–(41).
Next, we will prove the solvability of the system (39)–(41). We use the Bellman iteration method [21]. We consider the interval [ t k , t k + 1 ) where ξ ( t ) = y i ) and omit the index k for v , u and P. We define the zero approximation
u 0 ( t , x ) = M 1 1 ( t ) B T ( t ) P 0 ( t ) x ( t ) ,
where P 0 ( t ) 0 is a bounded piecewise continuous matrix. Substitute (42) into (27) and find the value of v 1 ( t , x ) from the resulting equation, which corresponds to the control of (42).
Next, we substitute v 1 ( t , x ) into the Bellman equation (22) and find the control u 1 ( t , x ) that minimizes (22).
Continuing this process, we obtain a sequence of controls u n ( t , x ) and functional v n ( t , x ) in the form
u n ( t , x ) = M 1 1 ( t ) B T ( t ) P n ( t ) x ( t ) , v n ( t , x ) = x T ( t ) P n ( t ) x ( t ) , v n ( T , x ) = x T ( t k ) M 0 x ( t k ) ,
where P n ( t ) , t [ t k , t k + 1 ) is the solution of the boundary-value problem (39)–(41) for T : = t k + 1 .
For n 0 , the next estimate is correct:
v n 1 ( t , x ) v n ( t , x ) , t [ t k , t k + 1 ) .
The convergence of the functions v n ( t , x ) to v 0 ( t , x ) , the controls u n ( t , x ) to u 0 ( t , x ) , and the convergence of the sequence of matrices P n ( t ) to P ( t ) can be proved using (44) [17].
The next estimate is correct:
max t [ t k , t k + 1 ) P ( t ) P n ( t ) C n ! , C < + , k 1 .
Thus, we state the following theorem.
Theorem 5.
The approximate solution of the optimal control synthesis problem for the problem (30)–(32) is carried out using successive Bellman approximations, where the n-th approximation of the optimal control and the Bellman functional for each interval [ t k , t k + 1 ) , k 0 , is given by the formula (43). In this case, the error is estimated by the inequality (46).

7. Model Example

Consider the system
d x ( t ) = [ a ( ξ ( t ) ) x ( t ) + b ( ξ ( t ) ) u ( t ) ] d t + σ ( ξ ( t ) ) x ( t ) d w ( t ) , t R + ,
with the initial condition
x ( 0 ) = 1 , ξ ( 0 ) = y 1 .
Here, ξ is a semi-Markov process with two states y 1 = 1 ,   y 2 = 2 with transition probabilities for a nested Markov chain q 12 = q 21 = 1 and time in state θ n = 0.2 + P o i s ( λ ( y i ) ) , where λ ( y 1 ) = 0.5 ,   λ ( y 2 ) = 1.3 . a ( y 1 ) = a 1 = 1 ,   a ( y 2 ) = a 2 = 0.1 ,   b ( y 1 ) = b 1 = 2 , b ( y 2 ) = b 2 = 0.5 ,   σ ( y 1 ) = σ 1 = 0.5 ,   σ ( y 2 ) = σ 2 = 0.8 .
The matrices from the quality functional (32) are assumed to be equal to
M 01 = 3 , M 02 = 1 , M 11 = M 12 = 2 , M 21 = 1 , M 22 = 2
The Bellman functional will be found in the form
v k 0 ( t , x ; i ) = x 2 ( t ) P i + g ( t ) .
In this case, the system (39)–(41) has the form
2 a i P i b i 2 M 1 i 1 P i + P j ( t ) P i ( t ) q i j + M 1 i 1 b i 2 ( t ) P i 2 + M 2 i ( t ) = 0 , i , j = 1 , 2 , i j ,
σ i 2 P i + g ˙ ( t ) = 0 ,
with a boundary condition
x 2 ( T ) P i + g ( T ) = x 2 ( T ) M 0 i .
Or
2 P 1 2 P 1 + P 2 + 1 = 0 , 0.125 P 2 2 0.925 P 2 + P 1 + 2 = 0 ,
where P 1 0.847 , P 2 1.587 .
The optimal control is as follows
u k 0 ( t , x ; 1 ) = 0.847 x ( t ) , u k 0 ( t , x ; 2 ) = 0.397 x ( t ) .
Simulated trajectory of the solution (46) and (47) without control and under the influence of optimal control is shown in Figure 1.

8. Discussion

The main focus of this paper is on theoretical derivations of the optimal control system for stochastic differential equations in the presence of external perturbations described by semi-Markov processes. This generalization allows us to more accurately describe the dynamics of real processes under various restrictions on the spend time θ n in states, which is impossible in the case of a Markov process. In Theorem 2, we find an explicit form of the infinitesimal operator, which is determined on the basis of the coefficients of the original equation and the characteristics of the semi-Markov process. This representation allows us to synthesize the optimal control u 0 ( t , x ) based on the Bellman equation (18) with the boundary condition (19). For the linear case of the system of the form (30), the search for optimal control is carried out on the basis of solving the Riccati equation (39), which also arises in the case of the presence of a Markovian external perturbation.
As we can see from Figure 1, the controlled solution visually resembles an asymptotically stable one, but from Figure 2 it is clear that the solution is not. Thus, asymptotic stability may not be achieved by constructing an optimal control. The problem of stability and stability of the solution will be the subject of research in subsequent articles.
The main focus in the following works for dynamical systems with semi-Markovian external perturbations will be on taking into account the ergodic properties of the semi-Markovian process ξ ( t ) , t 0 when analyzing the asymptotic behavior of the system. In contrast to systems with Markovian external switches, where the ergodic properties of ξ ( t ) were described on the basis of intensities λ ( y ) , for the semi-Markovian case, conditions on the times of steady state and jumps will play an important role. Thus, the parameter estimate of model (2) will have not only an estimate of the parameters a ( · ) , b ( · ) , but also an estimate of the distribution of the residence time in the states. Therefore, the following algorithm can be proposed for system analysis and parameter estimation:
  • Estimation of switching moments
    τ n = i = 1 n θ n ;
    This estimation can be realized using a generalized unit-root test developed for time series [22];
  • Estimation state space for semi-Markov process ξ ( t ) ,
    Y = { 1 , 2 , . . . , N } ;
  • Estimation coefficients a ( · ) , b ( · ) for SDE (2);
The presented framework of stochastic dynamic systems with semi-Markov parameters offers a promising tool for systems biology and systems medicine. In systems biology, it can help model complex molecular interactions, such as oxidative stress and mitochondrial dysfunction, influenced by stochastic perturbations. In systems medicine, this approach supports personalized treatment strategies by capturing patient-specific dynamics. For instance, it can predict disease progression and optimize therapies in conditions like Parkinson’s Disease. By integrating theoretical modeling with clinical data, this framework bridges the gap between understanding disease mechanisms and advancing precision medicine.

9. Conclusions

We propose optimal control strategies for stochastic dynamical systems with semi-Markov parameters. A linear optimization algorithm is derived and its convergence substantiated. This work focused primarily on the generalization of the Markov model, which would allow for a more accurate description of real processes in which switching occurs at random times, not subject to an exponential distribution. The developed theory currently has a number of advantages, but there are a variety of limitations to consider when interpreting the results. First, semi-Markov switching models require defining a distribution of switching times, which complicates estimation of model parameters. Second, the use of the semi-Markov process requires the verification of the finiteness of the necessary moments, and in the case of the Markov switching process, this verification was reduced to the inequality 0 < λ m i n < λ m a x < , where λ m i n , λ m a x —minimal and maximal intensity for the Markov switching process. Future work in this direction will develop algorithms for estimating optimal control for nonlinear systems with quadratic and quadratic-exponential performance functions.

Author Contributions

Conceptualization, T.L., I.V.M., A.A.H. and V.P.S.; methodology, T.L., I.V.M. and V.P.S.; formal analysis, T.L., I.V.M. and A.A.H.; investigation, T.L.; writing—original draft preparation, T.L. and I.V.M.; writing—review and editing, T.L. and I.V.M. and A.A.H.; visualization, I.V.M. and A.A.H.; supervision, V.P.S.; project administration, V.P.S.; funding acquisition, V.P.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by ELIXIR-LU https://elixir-luxembourg.org/ (accessed on 24 March 2025), the Luxembourgish node of ELIXIR, with funding and infrastructure provided by the Luxembourg Centre for Systems Biomedicine (LCSB). LCSB’s support contributed to the computational analyses and methodological development presented in this study.

Data Availability Statement

Data are contained within the article.

Acknowledgments

The authors would like to acknowledge the institutional support provided by the Luxembourg Centre for Systems Biomedicine (LCSB) at the University of Luxembourg and Yuriy Fedkovych Chernivtsi National University, which facilitated the completion of this work.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lindquist, A. Optimal control of linear stochastic systems with applications to time lag systems. Inf. Sci. 1973, 5, 81–124. [Google Scholar]
  2. Kumar, P.R.; van Schuppen, J.H. On the optimal control of stochastic systems with an exponential-of-integral performance index. J. Math. Anal. Appl. 1981, 80, 312–332. [Google Scholar]
  3. Buckdahn, R.; Labed, B.; Rainer, C.; Tamer, L. Existence of an optimal control for stochastic control systems with nonlinear cost functional. Stochastics 2010, 82, 241–256. [Google Scholar]
  4. Dragan, V.; Popa, I.-L. The Linear Quadratic Optimal Control Problem for Stochastic Systems Controlled by Impulses. Symmetry 2024, 16, 1170. [Google Scholar] [CrossRef]
  5. Das, A.; Lukashiv, T.O.; Malyk, I.V. Optimal control synthesis for stochastic dynamical systems of random structure with the markovian switchings. J. Autom. Inf. Sci. 2017, 4, 37–47. [Google Scholar] [CrossRef]
  6. Antonyuk, S.V.; Byrka, M.F.; Gorbatenko, M.Y.; Lukashiv, T.O.; Malyk, I.V. Optimal Control of Stochastic Dynamic Systems of a Random Structure with Poisson Switches and Markov Switching. J. Math. 2020, 2020, 9457152. [Google Scholar] [CrossRef]
  7. Dynkin, E.B. Markov Processes; Academic Press: New York, NY, USA, 1965. [Google Scholar]
  8. Koroliuk, V.; Limnios, N. Stochastic Systems in Merging Phase Space; World Scientific Publishing Co., Pte Ltd.: Hackensack, NJ, USA, 2005. [Google Scholar]
  9. Ibe, O. Markov Processes for Stochastic Modelling, 2nd ed.; Elsevier: London, UK, 2013. [Google Scholar]
  10. Gikhman, I.I.; Skorokhod, A.V. Introduction to the Theory of Random Processes; W. B. Saunders: Philadelphia, PA, USA, 1969. [Google Scholar]
  11. Feldman, R. Optimal Replacement With Semi-Markov Shock Model. J. Appl. Probab. 1976, 13, 108–117. [Google Scholar]
  12. Love, C.; Zhang, G.; Zitron, M.; Renkuan, G. Discrete semi-Markov decision model to determine the optimal repair/replacement policy under general repairs. Eur. J. Oper. Res. 2000, 125, 398–409. [Google Scholar] [CrossRef]
  13. Cheng, J.; Xie, L.; Park, J.; Yan, H. Protocol-Based Output-Feedback Control for Semi-Markov Jump Systems. IEEE Trans. Autom. Control 2022, 67, 4346–4353. [Google Scholar] [CrossRef]
  14. Xu, Z.; Su, H.; Shi, P.; Wu, Z.-G. Asynchronous H control of semi-Markov jump linear systems. Appl. Math. Comput. 2019, 349, 270–280. [Google Scholar]
  15. Jallow, M.A.; Nafiu, L.; Weke, P. Stochastic Forecasting of Stock Prices of Capital Assets Using Semi-Markov Model. Int. J. Stat. Econ. 2022, 23, 50–61. [Google Scholar]
  16. Øksendal, B. Stochastic Differential Equation; Springer: New York, NY, USA, 2013. [Google Scholar]
  17. Kolmanovskii, V.B.; Shaikhet, L.E. Control of Systems with Aftereffect; American Mathematical Society: Providence, RI, USA, 1996. [Google Scholar]
  18. Jacod, J.; Shiryaev, A.N. Limit Theorems for Stochastic Processes; Fizmatlit: Moscow, Russia, 1994; Volumes 1 and 2. (In Russian) [Google Scholar]
  19. Lukashiv, T. One Form of Lyapunov Operator for Stochastic Dynamic System with Markov Parameters. J. Math. 2016, 2016, 1694935. [Google Scholar]
  20. Stewart, I.A.N. Broken symmetries and biological patterns. In On Growth, Form and Computers; Kumar, S., Bentley, P.J., Eds.; Elsevier: Amsterdam, The Netherlands, 2003; pp. 181–202. [Google Scholar]
  21. Bellman, R. Dynamic Programming; Princeton University Press: Princeton, NJ, USA, 1972. [Google Scholar]
  22. Narayan, P.; Popp, S. A New Unit Root Test with Two Structural Breaks in Level and Slope at Unknown Time. J. Appl. Stat. 2010, 37, 1425–1438. [Google Scholar] [CrossRef]
Figure 1. The solution of the systems (46) and (47) with the given values of the coefficients.
Figure 1. The solution of the systems (46) and (47) with the given values of the coefficients.
Symmetry 17 00498 g001
Figure 2. Log−transformed controlled solution of the systems (46) and (47).
Figure 2. Log−transformed controlled solution of the systems (46) and (47).
Symmetry 17 00498 g002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lukashiv, T.; Malyk, I.V.; Hemedan, A.A.; Satagopam, V.P. Optimal Control of Stochastic Dynamic Systems with Semi-Markov Parameters. Symmetry 2025, 17, 498. https://doi.org/10.3390/sym17040498

AMA Style

Lukashiv T, Malyk IV, Hemedan AA, Satagopam VP. Optimal Control of Stochastic Dynamic Systems with Semi-Markov Parameters. Symmetry. 2025; 17(4):498. https://doi.org/10.3390/sym17040498

Chicago/Turabian Style

Lukashiv, Taras, Igor V. Malyk, Ahmed Abdelmonem Hemedan, and Venkata P. Satagopam. 2025. "Optimal Control of Stochastic Dynamic Systems with Semi-Markov Parameters" Symmetry 17, no. 4: 498. https://doi.org/10.3390/sym17040498

APA Style

Lukashiv, T., Malyk, I. V., Hemedan, A. A., & Satagopam, V. P. (2025). Optimal Control of Stochastic Dynamic Systems with Semi-Markov Parameters. Symmetry, 17(4), 498. https://doi.org/10.3390/sym17040498

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop