Next Article in Journal
Resource Allocation Effect of Green Credit Policy: Based on DID Model
Next Article in Special Issue
Modelling of Fuel- and Energy-Switching Prices by Mean-Reverting Processes and Their Applications to Alberta Energy Markets
Previous Article in Journal
New Series Solution of the Caputo Fractional Ambartsumian Delay Differential Equationation by Mittag-Leffler Functions
Previous Article in Special Issue
Insurance Contracts for Hedging Wind Power Uncertainty
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Controlled Discrete-Time Semi-Markov Random Evolutions and Their Applications

by
Anatoliy Swishchuk
1,† and
Nikolaos Limnios
2,*,†
1
Department of Mathematics and Statistics, Faculty of Science, University of Calgary, Calgary, AB T2N 1N4, Canada
2
Applied Mathematics Lab, Université de Technologie de Compiègne, Sorbonne University Alliance, 60203 Compiègne, France
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2021, 9(2), 158; https://doi.org/10.3390/math9020158
Submission received: 9 December 2020 / Revised: 4 January 2021 / Accepted: 11 January 2021 / Published: 13 January 2021
(This article belongs to the Special Issue New Trends in Random Evolutions and Their Applications)

Abstract

:
In this paper, we introduced controlled discrete-time semi-Markov random evolutions. These processes are random evolutions of discrete-time semi-Markov processes where we consider a control. applied to the values of random evolution. The main results concern time-rescaled weak convergence limit theorems in a Banach space of the above stochastic systems as averaging and diffusion approximation. The applications are given to the controlled additive functionals, controlled geometric Markov renewal processes, and controlled dynamical systems. We provide dynamical principles for discrete-time dynamical systems such as controlled additive functionals and controlled geometric Markov renewal processes. We also produce dynamic programming equations (Hamilton–Jacobi–Bellman equations) for the limiting processes in diffusion approximation such as controlled additive functionals, controlled geometric Markov renewal processes and controlled dynamical systems. As an example, we consider the solution of portfolio optimization problem by Merton for the limiting controlled geometric Markov renewal processes in diffusion approximation scheme. The rates of convergence in the limit theorems are also presented.

1. Introduction

Random evolutions were introduced over 40 years ago, see, e.g., in [1,2] and for its asymptotic theory in [3,4,5,6] and references therein. Discrete-time random evolutions, induced by discrete-time Markov chains, are introduced by Cohen [7] and Keepler [8], and discrete-time semi-Markov random evolutions (DTSMRE) by Limnios [9]. See also [10]. Koroliuk and Swishchuk [4], Swishchuk and Wu [5], Anisimov [11,12,13,14], and Koroliuk and Limnios [3], studied discrete-time random evolutions induced by the embedded Markov chains of continuous time semi-Markov processes. This is equivalent to discrete-time Markov random evolution stopped at random time (continuous). One of the examples of discrete-time random evolutions is the geometric Markov renewal process (GMRP). Applications of GMRP in finance have been considered in [15,16,17]. Optimal stopping of GMRP and pricing of European and American options for underlying assets modelled by GMRP have been studied in [18].
Discrete-time semi-Markov chains (SMC) have only recently been used in applications. Especially, in DNA analysis, image and speech processing, reliability theory, etc., see in [19] and references therein. These applications have stimulated a research effort in this area. While the literature in discrete-time Markov chains theory and applications is quite extensive, there is only a small amount of the literature on SMC and most of them are related to hidden semi-Markov models for estimation.
The present article is a continuation of our previous work [20]. Thus, we keep all our notation and definitions the same as in the latter paper. Compared with our previous work [20], where we studied random evolutions of semi-Markov chains, here we considered additionally a control on the random evolution, which we call controlled discrete-time semi-Markov random evolution (CDTSMRE) in a Banach space, and we presented time-rescaled convergence theorems. In particular, we get weak convergence theorems in Skorokhod space D [ 0 , ) for càdlàg stochastic processes, see, e.g., in [21]. The limit theorems include averaging, diffusion approximation, and diffusion approximation with equilibrium. For the above limit theorems we also presented rates of convergence results. Finally, we give some applications regarding the above mentioned results, especially to controlled additive functionals (CAF), CGMRP, and controlled dynamical systems (CDS), and optimization problems.
Regarding the optimization problems, we provide dynamical principles for discrete-time dynamical systems such as CAF and CGMRPs (see Section 2.4), see, e.g., [22,23,24]. We also produce dynamic programming equations (Hamilton–Jacobi–Bellman equations) for the limiting processes in diffusion approximation such as CAF, CGMRP, and CDS. As an example, we consider the solution of portfolio optimization problem by Merton for the limiting CGMRP in DA (see Section 4.4). Merton problem, or Merton portfolio’s problem, is a problem in continuous-time finance associated with portfolio choice. In ( B , S ) -security market, which consists of a stock and a risk-free asset, an investor must choose how much to consume, and must allocate his wealth between the stock and the risk-free asset in a such way that maximizes expected utility. The problem was formulated and first solved by Robert Merton in 1969, and published in 1971 [25].
Results presented here are new and deals with CDTSMRE on Banach spaces. This paper contains new and original results on dynamical principle for CDTSMRE and DPE (HJB equations) for the limiting processes in DA. One of the new remarkable results is the solution of Merton portfolio problem for the limiting CGMRP in DA. The method of proofs was based on the martingale approach together with convergence of transition operators of the extended semi-Markov chain via a solution of a singular perturbation problem [3,4,26]. As in our previous work [20], the tightness of these processes is proved via Sobolev’s embedding theorems [27,28,29]. It is worth mentioning that, as in the Markov case, the results presented here cannot be deduced directly from the continuous-time case. We should also note that that DTSMREs have been completely studied in [20]. For semi-Markov processes see, e.g., [30,31,32,33]. For Markov chains and additive functionals see, e.g., [34,35,36,37,38].
The paper is organized as follows. Definition and properties of discrete-time semi-Markov random evolutions and Controlled DTSMREs, as well as particular stochastic systems as applications, are introduced in Section 2. The main results of this paper, limit theorems of CDTSMRE, as averaging, diffusion approximation and diffusion approximation with equilibrium of controlled DTSMREs are considered in Section 3. In Section 4, we provide three applications of averaging, diffusion approximation, and diffusion approximation with equilibrium of controlled DTSMREs: controlled additive functionals, controlled GMRP, and controlled dynamical systems. Section 5 deals with the analysis of the rates of convergence in the limit theorems, presented in the previous sections, for controlled DTSMREs and for CAF and CGMRP. In Section 6, we give the proofs of theorems presented in the previous sections. The last section concludes the paper and indicates some future works.

2. Controlled Discrete-Time Semi-Markov Random Evolutions

2.1. Semi-Markov Chains

The aim of this section is to present some notation and to make this paper as autonomous as possible. The reader may refer to our article in [20] for more details.
Let ( E , E ) be a measurable space with countably generated σ -algebra and ( Ω , F , ( F n ) n I N , P ) be a stochastic basis on which we consider a Markov renewal process ( x n , τ n , n I N ) in discrete time k I N , with state space ( E , E ) . Notice that I N is the set of non-negative integer numbers. The semi-Markov kernel q is defined by (see, e.g., in [9,19]),
q ( x , B , k ) : = P ( x n + 1 B , τ n + 1 τ n = k x n = x ) , x E , B E , k , n I N .
We will denote also q ( x , B , Γ ) = n Γ q ( x , B , n ) , where Γ I N . The process ( x n ) is the embedded Markov chain of the MRP ( x n , τ n ) with transition kernel P ( x , d y ) . The semi-Markov kernel q is written as
q ( x , d y , k ) = P ( x , d y ) f x y ( k ) ,
where f x y ( k ) : = P ( τ n + 1 τ n = k x n = x , x n + 1 = y ) , the conditional distribution of the sojourn time in state x given that the next visited state is y.
Define also the counting process of jumps ν k = max { n : τ n k } , and the discrete-time semi-Markov chain z k by z k = x ν k , for k I N . Define now the backward recurrence time process γ k : = k τ ν k , k 0 , and the filtration F k : = σ ( z , γ ; k ) , k 0 .
Let us consider a separable Banach space B of real-valued measurable functions defined on E × I N , endowed with the sup norm · and denote by B its Borel σ -algebra. The Markov chain ( z k , γ k ) , k 0 , has the following transition probability operator P on B
P φ ( x , k ) = 1 H ¯ x ( k ) E \ { x } q ( x , d y , k + 1 ) φ ( y , 0 ) + H ¯ x ( k + 1 ) H ¯ x ( k ) φ ( x , k + 1 ) ,
where φ B , and its stationary distribution, if there exist, is given by
π ( d x × { k } ) = ρ ( d x ) H ¯ x ( k ) / m ,
where
m : = E ρ ( d x ) m ( x ) , m ( x ) = k 0 H ¯ x ( k ) ,
and ρ ( d x ) is the stationary distribution of the EMC ( x n ) , H x ( k ) : = q ( x , E , [ 0 , k ] ) , and H ¯ x ( k ) : = 1 H x ( k ) = q ( x , E , [ k + 1 , ) ) . The probability measure π defined by π ( B ) = π ( B × I N ) is the stationary probability of the SMC ( z k ) . Define also the r-th moment of holding time in state x E ,
m r ( x ) : = k 1 k r q ( x , E , k ) , r = 1 , 2 ,
Of course, m ( x ) = m 1 ( x ) , for any x E .
Define now the stationary projection operator Π on the null space of the (discrete) generating operator Q : = P I ,
Π φ ( x , s ) = 0 E π ( d y × { } ) φ ( y , ) 1 ( x , s ) ,
where 1 ( x , s ) = 1 for any x E , and s I N . This operator satisfies the equations
Π Q = Q Π = 0 .
The potential operator of Q , denoted by R 0 , is defined by
R 0 : = ( Q + Π ) 1 Π = k 0 [ ( P ) k Π ] .

2.2. General Definition and Properties of DTSMREs

We define here controlled discrete-time semi-Markov random evolutions. Let U denote a compact Polish space representing the control, and let u k be U-valued control process and we suppose that it is a Markov chain. We note that we could also define the process u ν k which is a semi-Markov control process, considered in many papers (see, e.g., in [39,40]). We suppose that homogeneous Markov chain u k is independent of z k , and transition probability kernel P u = P ( u k + 1 d y u k = u ) = Q ( u , d y ) .
Let us consider a family of bounded contraction operators D ( z , u ) , z E , u U , defined on B, where the maps D ( z , u ) φ : E × U B are E × U -measurable, φ B . Denote by I the identity operator on B. Let Π B = N ( Q ) be the null space, and ( I Π ) B = R ( Q ) be the range values space of operator Q . We will suppose here that the Markov chain ( z k , γ k , k I N ) is uniformly ergodic, that is, ( ( P ) n Π ) φ 0 , as n , for any φ B . In that case, the transition operator is reducible-invertible on B. Thus, we have B = N ( Q ) R ( Q ) , the direct sum of the two subspaces. The domain of an operator A on B is denoted by D ( A ) : = { φ B : A φ B } .
Definition 1.
A controlled discrete-time semi-Markov random evolution (CDTSMRE) Φ k u , k I N , on the Banach space B, is defined by
Φ k u φ = D ( z k , u k ) D ( z k 1 , u k 1 ) D ( z 2 , u 2 ) D ( z 1 , u 1 ) φ ,
for k 1 , Φ 0 u = I , u 0 = u U , and for any φ B 0 : = x E , u U D ( D ( x , u ) ) . Thus we have Φ k = D ( z k , u k ) Φ k 1 .
The process ( z k , γ k , u k ) is a Markov chain on E × I N × U , adapted to the filtration F k u : = σ ( z , γ , u ; k ) , k 0 . We also note that ( Φ k u φ , z k , γ k , u k ) is a Markov chain on B × E × I N × U with discrete generator
L u φ = [ P ˜ + E U P ˜ ( · , d v ) ( D ( v , u ) I ) ] φ ,
where φ : = φ ( x , z , s , u ) , and
P ˜ φ ( z , s , u ) : = P P u φ ( z , s , u ) = s I N E × U P ( z , s ; d z , s ) P u ( u ; d u ) φ ( z , s , u ) .
The process M k u defined by
M k u : = Φ k u I = 0 k 1 E [ Φ + 1 u Φ u F u ] , k 1 , M 0 = 0 ,
on B, is an F k u -martingale. The random evolution Φ k u can be written as follows
Φ k u : = I + = 0 k 1 [ D ( z + 1 , u + 1 ) I ] Φ u ,
and then, the martingale (5) can be written as follows,
M k u : = Φ k u I = 0 k 1 E [ ( D ( z + 1 , u + 1 ) I ) Φ u F u ] ,
or
M k u : = Φ k u I = 0 k 1 [ E ( D ( z + 1 , u + 1 ) F u ) I ] Φ u .
Finally, as E [ D ( z + 1 , u + 1 ) Φ u φ F u ] = [ ( P + P u ) D ( · ) Φ φ ] ( z , γ , u ) , one takes
M k u : = Φ k u I = 0 k 1 [ P ˜ D ( · , u ) I ] Φ u .

2.3. Some Examples

Example 1. 
Controlled Additive Functional or Markov Decision Process.
Let define the following controlled additive functional,
y k u = l = 0 k a ( z l , u l ) , k 0 , y 0 = y .
If we define the operator D ( z , u ) on C 0 ( I R ) in the following way,
D ( z , u ) φ ( y ) : = φ ( y + a ( z , u ) ) ,
then the controlled discrete-time semi-Markov random evolution Φ k φ has the following presentation,
Φ k u φ ( y ) = φ ( y k u ) .
Process y k u is usually called in the literature the Markov decision process (see, e.g., in [41,42,43,44]).
Example 2. 
Controlled geometric Markov renewal process.
The CGMRP is defined in the following way,
S k u : = S 0 l = 1 k ( 1 + a ( z l , u l ) ) , k I N , S 0 = s .
We suppose that k = 1 0 = 1 .
If we define the operator D ( z , u ) on C 0 ( I R ) in the following way,
D ( z , u ) φ ( s ) : = φ ( s ( 1 + a ( z , u ) ) ) ,
then the controlled discrete-time semi-Markov random evolution Φ k u φ can be given as follows,
Φ k u φ ( s ) = φ ( S k u ) .
To the authors opinion, this process is defined for the first time in the literature and the notion of controlled GMRP is a new one as well.

2.4. Dynamic Programming for Controlled Models

Here, we present dynamic programming for controlled models given in Examples in previous section. Let us consider a Markov control model (see in [45]) ( E , A , { A ( z ) | z E } , Q , c ) . Here, E is the state space; A is the control or action set; Q is the transition kernel, i.e., a stochastic kernel on E given K , where K : = { ( z , u ) | z E , u A ( z ) } ; and c : K R is a measurable function called the cost-per-stage function.
We are interested in is to minimize the finite-horizon performance criterion either (see Example 1)
J 1 ( π , z ) : = E z π [ l = 0 N 1 a ( z l , u l ) + a N ( z N ) ]
or (see Example 2)
J 2 ( π , z ) : = E z π [ ln ( l = 1 k ( 1 + a ( z l , u l ) ) ( 1 + a N ( z N ) ) ) ] ,
where a N ( z N ) is the terminal cost function, π Π is the set of control policies.
In this way, denoting by J * the value function
J i * ( z ) : = inf Π J i ( π , z ) , z E , i = 1 , 2 ,
the problem is to find a policy π * Π such that
J i ( π * , z ) = J i * ( z ) , z E , i = 1 , 2 , .
Example 3. 
Controlled Additive Functional.
Let us provide an algorithm for finding both the value function J * and an optimal policy π * for the example with function J 1 ( π , z ) (see Example 1).
Let J 0 , 1 , J 1 , 1 , , J N , 1 be the functions on E defined from l = N to l = 0 by (backwards)
J N , 1 ( z ) : = a N ( z ) , l = N
and
J l , 1 ( z ) : = min A ( z ) [ a ( z , u ) + E J l + 1 , 1 ( y ) Q ( u , d y ) ] , l = N 1 , N 2 , , 0 .
Suppose that there is a selector f t F such that f l ( z ) A ( z ) attains the minimum in the above expression for J l ( z ) for all z E , meaning for any z E and l = 0 , , N 1 ,
J l , 1 ( z ) = a ( z , f l ) + E J l + 1 , 1 ( y ) Q ( f l , d y ) .
Then, the optimal policy is the deterministic Markov one π * = { f 0 , , f N 1 } , and the value function J * equals J 0 , i.e.,
J 1 * ( z ) = J 0 ( z ) = J 1 ( π * , z ) , z E .
Example 4. 
Controlled Geometric Markov Renewal Chain.
Let us provide an algorithm for finding both the value function J * and an optimal policy π * for the example with function J 2 ( π , z ) (see Example 2). We will modify the expression for S k u in Example 2. Let ln ( S k u S 0 ) be a log-return, then
ln ( S k u S 0 ) = l = 1 k ln ( 1 + a ( z l , u l ) .
Thus, we are interested in minimizing the finite-horizon performance criterion for
J 2 ( π , z ) : = E z π [ l = 0 N 1 ln ( 1 + a ( z l , u l ) + ln ( 1 + a N ( z N ) ]
Let J 0 , 2 , J 1 , 2 , , J N , 2 be the functions on E defined from l = N to l = 0 by (backwards)
J N , 2 ( z ) : = ln ( 1 + a N ( z ) ) , l = N
and
J l , 2 ( z ) : = min A ( z ) [ ln ( 1 + a ( z , u ) ) + E J l + 1 , 1 ( y ) Q ( u , d y ) ] , l = N 1 , N 2 , , 0 .
Suppose that there is a selector f t F such that f l ( z ) A ( z ) attains the minimum in the above expression for J l ( z ) for all z E , meaning for any z E and l = 0 , , N 1 ,
J l , 2 ( z ) = ln ( 1 + a ( z , f l ) ) + E J l + 1 , 2 ( y ) Q ( f l , d y | z ) .
Then, the deterministic Markov policy π * = { f 0 , , f N 1 } is optimal, and the value function J * equals J 0 , i.e.,
J 2 * ( z ) = J 0 ( z ) = J 2 ( π * , z ) , z E .

3. Limit Theorems for Controlled Semi-Markov Random Evolutions

In this section, we present averaging, diffusion approximation, and diffusion approximation with equilibrium results for the controlled discrete-time semi-Markov random evolutions. It is worth noticing that the main scheme of results are almost the same as in our previous works in particular [20]. Nevertheless, the additional component of the control allows us to study more interesting problems.

3.1. Averaging of CDTSMREs

We consider here CDTSMREs defined in Section 2. Let us now set k : = [ t / ε ] and consider the continuous time process M t ε
M t ε , u : = M [ t / ε ] u = Φ [ t / ε ] ε , u I = 0 [ t / ε ] 1 [ P ˜ D ε ( · , u ) I ] Φ ε , u
We will prove here asymptotic results for this process as ε 0 .
The following assumptions are needed for averaging.
A1:
The MC ( z k , γ k , k I N ) is uniformly ergodic with ergodic distribution π ( B × { k } ) , B E , k I N .
A2:
The moments m 2 ( x ) , x E , are uniformly integrable.
A3:
The perturbed operators D ε ( x ) have the following representation on B
D ε ( x , u ) = I + ε D 1 ( x , u ) + ε D 0 ε ( x , u ) ,
where operators D 1 ( x , u ) on B are closed and B 0 : = x E , u U D ( D 1 ( x , u ) ) is dense in B, B ¯ 0 = B . Operators D 0 ε ( x , u ) are negligible, i.e., lim ε 0 D 0 ε ( x , u ) φ = 0 for any φ B 0 .
A4:
We have E U [ π ( d x ) π 1 ( d u ) ] D 1 ( x , u ) φ 2 < . (See A7.)
A5:
There exists Hilbert spaces H and H * such that compactly embedded in Banach spaces B and B * , respectively, where B * is a dual space to B .
A6:
Operators D ε ( x ) and ( D ε ) * ( x ) are contractive on Hilbert spaces H and H * , respectively.
A7:
The MC ( u k , k I N ) , is independent of ( z k ) , and is uniformly ergodic with stationary distribution π 1 ( d u ) , k I N .
We note that if B = C 0 ( I R ) , then H = W l , 2 ( I R ) is a Sobolev space, and W l , 2 ( I R ) C 0 ( I R ) and this embedding is compact (see [29]). For the spaces B = L 2 ( I R ) and H = W l , 2 ( I R ) the situation is the same.
We also note, that semi-Markov chain ( z k , u k ) is uniformly ergodic on E × U with stationary probabilities π ( d x ) π 1 ( d u ) , which follows from conditions A1 and A7.
Theorem 1.
Under Assumptions A1–A7, the following weak convergence takes place,
Φ [ t / ε , u ] ε Φ ¯ ( t ) , ε 0 ,
where the limit random evolution Φ ¯ ( t ) is determined by the following equation,
Φ ¯ ( t ) φ φ 0 t I L ^ Φ ¯ ( s ) φ d s = 0 , 0 t T , φ B 0 ,
or, equivalently,
d d t Φ ¯ ( t ) φ = I L ^ Φ ¯ ( t ) φ ,
where the limit contracted operator is then given by
I L ^ = D ^ 1 = E U [ π ( d x ) π 1 ( d u ) ] D 1 ( x , u ) .
This result generalize the classical Krylov–Bogolyubov averaging principle [46] on a Banach and a controlled spaces.

3.2. Diffusion Approximation of DTSMREs

For the diffusion approximation of CDTSMREs, we will consider a different time-scaling and some additional assumptions.
D1:
Let us assume that the perturbed operators D ε ( x , u ) have the following representation in B,
D ε ( x , u ) = I + ε D 1 ( x , u ) + ε 2 D 2 ( x , u ) + ε 2 D 0 ε ( x , u ) ,
where operators D 2 ( x , u ) on B are closed and B 0 : = x E , u U D ( D 2 ( x , u ) ) is dense in B, B ¯ 0 = B ; operators D 0 ε ( x , u ) are a negligible operator, i.e., lim ε 0 D 0 ε ( x , u ) φ = 0 .
D2:
The following balance condition holds,
Π D 1 ( x , u ) Π = 0 ,
where
Π φ ( x , k , u ) : = l 0 E U π ( d y × ) π 1 ( d u ) φ ( y , , u ) 1 ( x , k ) .
D3:
The moments m 3 ( x ) , x E , are uniformly integrable.
Theorem 2.
Under Assumptions A1, A5–A7 (see Section 3.1), and D1-D3, the following weak convergence takes place,
Φ [ t / ε 2 ] ε Φ 0 ( t ) , ε 0 ,
where the limit random evolution Φ 0 ( t ) is a diffusion random evolution determined by the following generator
I L = Π D 2 ( x ) Π + Π D 1 ( x ) R 0 D 1 ( x ) Π Π D 1 2 ( x ) Π ,
where
R 0 : = [ Q ˜ + Π ] 1 Π ,
and
Q ˜ : = P ˜ I .

3.3. Diffusion Approximation with Equilibrium

The diffusion approximation with equilibrium or the normal deviation is obtained by considering the difference between the rescaled initial processes and the averaging limit process. This is of great interest when we have no balance condition as previously in the standard diffusion approximation scheme.
Consider now the controlled discrete-time semi-Markov random evolution Φ [ t / ε , u ] ε , averaged evolution Φ ¯ ( t ) (see Section 3.1) and the deviated evolution
W t ε , u : = ε 1 / 2 [ Φ [ t / ε ] ε , u Φ ¯ ( t ) ] .
Theorem 3.
Under Assumptions A1, A5–A6 (see Section 3.1), and D3, with operators D ε ( x ) in A3, instead of D1, the deviated controlled semi-Markov random evolution W t ε , u weakly convergence, when ε 0 , to the diffusion random evolution W t 0 defined by the following generator
I L = Π ( D 1 ( x , u ) D ¯ 1 ) R 0 ( D 1 ( x , u ) D ¯ 1 ) Π ,
where Π is defined in (9).

4. Applications to Stochastic Systems

In this section, we give two applications in connection with the above results: additive functionals that has many application, e.g., in storage, reliability, and risk theories (see, e.g., in [3,4,19,47]), and to geometric Markov renewal processes, that also have many application including finance (see [15,16,17,18]). Our main goal here is to get the limiting processes and apply optimal control methods to receive the solutions of optimization problems. The limiting results for MC such as LLN and CLT were considered in [11,12].

4.1. Controlled Additive Functionals

Let us consider here the CAF, ( y k u ) , described previously in Example 1.
Averaging of CAF. Now, if we define the continuous time process
y t ε , u : = ε l = 0 [ t / ε ] a ( z l , u l ) ,
then from Theorem 1 it follows that this process has the following limit y 0 ( t ) = lim ε 0 y t ε
y 0 ( t ) = y + a ^ t ,
where a ^ = E U π ( d z ) π 1 ( d u ) a ( z , u ) . We suppose that E U π ( d z ) π 1 ( d u ) | a ( z , u ) | < + .
Diffusion Approximation of CAF. If we consider the continuous time process ξ t ε , u as follows
ξ t ε , u : = ε l = 0 [ t / ε 2 ] a ( z l , u l ) , ξ 0 ε = y ,
then under balance condition E U π ( d z ) π 1 ( d u ) a ( z , u ) = 0 and E U π ( d z ) π 1 ( d u ) | a ( z , u ) | 2 < + we get that the limit process ξ 0 ( t ) = lim ε 0 ξ t ε has the following form,
ξ 0 ( t ) = y + b w t ,
where b 2 = 2 a ^ 0 a ^ 2 , and
a ^ 0 = E U π ( d z ) π 1 ( d u ) a ( z , u ) R 0 a ( z , u ) , a ^ 2 = E U π ( d z ) π 1 ( d u ) a 2 ( z , u ) ,
and w t is a standard Wiener process.
Diffusion Approximation with Equilibrium of CAF. Let us consider the following normalized additive functional,
w t ε , u : = ε 1 / 2 [ y t ε , u a ^ t ] .
Then, this process converges to the following process, σ w t , where
σ 2 = E U π ( d z ) π 1 ( d u ) ( a ( z , u ) a ^ ) R 0 ( a ( z , u ) a ^ ) ,
and w t is a standard Wiener process.
In this way, the AF y t ε may be presented in the following approximated form,
y t ε a ^ t + ε σ w t .

4.2. Controlled Geometric Markov Renewal Processes

The CGMRP is defined in the following way (see in [15,16]),
S k u : = S 0 l = 1 k ( 1 + a ( z l , u l ) ) , k I N , τ 0 = s .
We suppose that k = 1 0 = 1 .
If we define the operator D ( z ) on C 0 ( I R ) in the following way,
D ( z , u ) φ ( s ) : = φ ( s ( 1 + a ( z , u ) ) ) ,
then the discrete-time semi-Markov random evolution Φ k u φ has the following presentation,
Φ k u φ ( s ) = φ ( S k u ) .
Averaging of CGMRP. Now, define the following sequence of processes,
S t ε , u : = S 0 k = 1 [ t / ε ] ( 1 + ε a ( z k , u k ) ) , t I R + , S 0 = s .
Then, under averaging conditions the limit process S ¯ t has the following form,
S ¯ t = S 0 e a ^ t ,
where a ^ = E U π ( d z ) π 1 ( d u ) a ( z , u ) .
Diffusion Approximation of CGMRP. If we define the following sequence of processes,
S ε , u ( t ) : = S 0 k = 1 [ t / ε 2 ] ( 1 + ε a ( z k , u k ) ) , t I R + , S 0 = s ,
then, in the diffusion approximation scheme, we have the following limit process, S 0 ( t )
S 0 ( t ) = S 0 e t a ^ 2 / 2 e σ a w ( t ) ,
where
a ^ 2 : = E U π ( d z ) π 1 ( d u ) a 2 ( z , u ) ,
σ a 2 : = E U π ( d z ) π 1 ( d u ) [ a 2 ( z , u ) / 2 + a ( z , u ) R 0 a ( z , u ) ] .
It means that S 0 ( t ) satisfies the following stochastic differential equation,
d S 0 ( t ) S 0 ( t ) = 1 2 ( σ a 2 a ^ 2 ) d t + σ a d w t ,
where w t is a standard Wiener process.
Diffusion Approximation with Equilibrium of CGMRP. Let us consider the following normalized GMRP:
w t ε , u : = ε 1 / 2 [ ln ( S t ε , u / S 0 ) a ^ t ] .
It is worth noticing that in finance the expression ln ( S t ε , u / S 0 ) represents the log-return of the underlying asset (e.g., stock) S t ε , u .
Then, this process converges to the following process, σ w t , where
σ 2 = E U π ( d z ) π 1 ( d u ) ( a ( z , u ) a ^ ) R 0 ( a ( z , u ) a ^ ) ,
and w t is a standard Wiener process.
In this way, the GMRP S t ε may be presented in the following approximated form,
S t ε S 0 e a ^ t + ε σ w t .

4.3. Controlled Dynamical Systems

We consider here discrete-time CDS and their asymptotic behaviour in series scheme: average and diffusion approximation ([9]).
Define the measurable function C on I R × E × U . Let us consider the difference equation
y k + 1 ε , u = y k ε , u + ε C ( y k ε ; z k + 1 , u k + 1 ) , k 0 , and y 0 ε = u ,
switched by the SMC ( z k ) .
The perturbed operators D ε ( z , u ) , x E , are defined now by
D ε ( z , u ) φ ( u ) = φ ( z + ε C ( z , x , u ) ) .
Averaging of CDS. Under averaging assumptions the following weak convergence takes place,
y [ t / ε ] ε , u y ¯ ( t ) , as ε 0 ,
where y ¯ ( t ) , t 0 is the solution of the following (deterministic) differential equation,
d d t y ¯ ( t ) = C ¯ ( y ¯ ( t ) ) , and y ¯ ( 0 ) = u ,
where C ¯ ( z ) = E U π ( d x ) π 1 ( d u ) C ( z , x , u ) .
Diffusion Approximation of CDS. Under diffusion approximation conditions the following weak convergence takes place
y [ t / ε 2 ] ε , u x t , as ε 0 ,
where x t , t 0 , is a diffusion processes, with initial value x 0 = u , determined by the operator
I L φ ( z ) = a ( z ) φ ( z ) + 1 2 b 2 ( z ) φ ( z ) ,
provided that b 2 ( z ) > 0 , and drift and diffusion coefficients are defined as follows,
b 2 ( z ) : = 2 C ¯ 0 ( z ) C ¯ 2 ( z ) , a ( z ) : = C ¯ 01 ( z ) C ¯ 1 ( z ) ,
with:
C ¯ 0 ( z ) : = E U π ( d x ) π 1 ( d u ) C 0 ( z , x , u ) , C 0 ( z , x , u ) : = C ( z , x , u ) R 0 C ( z , x , u ) ,
C ¯ 2 ( z ) : = E U π ( d x ) π 1 ( d u ) C * ( z , x , u ) C ( z , x , u ) , where C * means transpose of the vector C,
C ¯ 01 ( z ) : = E U π ( d x ) π 1 ( d u ) C 01 ( z , x , u ) , C 01 ( z , x , u ) : = C ( z , x , u ) R 0 C z ( z , x , u ) ,
C ¯ 1 ( z ) : = E U π ( d x ) π 1 ( d u ) C 1 ( z , x , u ) , C 1 ( z , x , u ) : = C ( z , x , u ) C z ( z , x , u ) .

4.4. The Dynamic Programming Equations for Limiting Models in Diffusion Approximation

In this section, we consider the DPE, i.e., HJB Equations, for the limiting models in DA from Section 4.1, Section 4.2 andSection 4.3. As long as all limiting processes in DA in Section 4.1, Section 4.2 and Section 4.3 are diffusion processes, then we will set up a general approach to control for diffusion processes, see in [48].
Let x t u be a diffusion process satisfying the following stochastic differential equation,
d x t u = μ ( x t u , u t ) d t + σ ( x t u , u t ) d w t ,
where u t is the control process, w t is a standard Wiener process. Let us also introduce the following performance criterion function, J u ( t , x )
J u ( t , x ) : = E t , x [ G ( x T u ) + t T F ( s , x s u , u s ) d s ] ,
where G ( x ) : R R is a terminal reward function (uniformly bounded), F ( t , x , u ) : R + × R 2 R is a running penalty/reward function (uniformly bounded), 0 t T . The problem is to maximize this performance criteria, i.e., to find the value function
J ( t , x ) : = sup u U t , T J u ( t , x ) ,
where U t , T is the admissible set of strategies/controls which are F -predictable, non-negative, and bounded.
The Dynamic Programming Principle (DPP) for diffusions states that the value function J ( t , x ) satisfies the DPP
J ( t , x ) = sup u U t , T E t , x [ J u ( T , x T u ) + t T F ( s , x s u , u s ) d s ]
for all ( t , x ) [ 0 , T ] × R .
Moreover, the value function J ( t , x ) above satisfies the Dynamic Programming Equation (DPE) or Hamilton–Jacobi–Bellman (HJB) Equation:
J ( t , x ) t + sup u U t , T [ L t u J ( t , x ) + F ( t , x , u ) ] = 0 J ( T , x ) = G ( x ) ,
where L t u is an infinitesimal generator of the diffusion process x t u above, i.e.,
L t u = μ ( x , u ) x + σ 2 ( x , u ) 2 2 x 2 .
DPE/HJB Equation for the Limiting CAF in DA (see Section 4.1)
We remind that the limiting process ξ 0 ( t ) = lim ε 0 ξ t ε in this case has the following form
ξ 0 ( t ) = y + b w t ,
where b 2 = 2 a ^ 0 a ^ 2 , and
a ^ 0 = E U π ( d z ) π 1 ( d u ) a ( z , u ) R 0 a ( z , u ) , a ^ 2 = E U π ( d z ) π 1 ( d u ) a 2 ( z , u ) ,
and w t is a standard Wiener process.
In this case, the DPE or HJB Equation (16) reads with the generator
L t u = 1 2 b 2 ( u ) 2 x 2 ,
with b 2 ( u ) : = 2 a ^ 0 ( u ) a ^ 2 ( u ) , and
a ^ 0 ( u ) : = E π ( d z ) a ( z , u ) R 0 a ( z , u ) , a ^ 2 ( u ) : = E π ( d z ) a 2 ( z , u ) .
DPE/HJB Equation for the Limiting CGMRP in DA (see Section 4.2)
We recall that we have the following limiting process S 0 ( t ) in this case:
S 0 ( t ) = S 0 e t a ^ 2 / 2 e σ a w ( t ) ,
where
a ^ 2 : = E U π ( d z ) π 1 ( d u ) a 2 ( z , u ) ,
σ a 2 : = E U π ( d z ) π 1 ( d u ) [ a 2 ( z , u ) / 2 + a ( z , u ) R 0 a ( z , u ) ] .
Furthermore, S 0 ( t ) satisfies the following stochastic differential equation (SDE),
d S 0 ( t ) S 0 ( t ) = 1 2 ( σ a 2 a ^ 2 ) d t + σ a d w t ,
where w t is a standard Wiener process.
In this case, the DPE or HJB Equation (16) reads with the generator
L t u = 1 2 ( σ a 2 ( u ) a 2 ( u ) ) s + 1 2 σ a 2 ( u ) 2 s 2 ,
and a ^ 2 ( u ) : = E π ( d z ) a 2 ( z , u ) , σ a 2 ( u ) : = E π ( d z ) [ a 2 ( z , u ) / 2 + a ( z , u ) R 0 a ( z , u ) ] .
DPE/HJB Equation for the Limiting CDS in DA (see Section 4.3)
We remind that in the diffusion approximation the limiting process is a diffusion process x t with a generator
L φ ( z ) = a ( z ) φ ( z ) + 1 2 b 2 ( z ) φ ( z ) ,
provided that b 2 ( z ) > 0 , and drift and diffusion coefficients are defined as follows,
b 2 ( z ) : = 2 C ¯ 0 ( z ) C ¯ 2 ( z ) , a ( z ) : = C ¯ 01 ( z ) C ¯ 1 ( z ) ,
with
C ¯ 0 ( z ) : = E U π ( d x ) π 1 ( d u ) C 0 ( z , x , u ) , C 0 ( z , x , u ) : = C ( z , x , u ) R 0 C ( z , x , u ) ,
C ¯ 2 ( z ) : = E U π ( d x ) π 1 ( d u ) C * ( z , x , u ) C ( z , x , u ) , where C * means transpose of the vector C,
C ¯ 01 ( z ) : = E U π ( d x ) π 1 ( d u ) C 01 ( z , x , u ) , C 01 ( z , x , u ) : = C ( z , x , u ) R 0 C z ( z , x , u ) ,
C ¯ 1 ( z ) : = E U π ( d x ) π 1 ( d u ) C 1 ( z , x , u ) , C 1 ( z , x , u ) : = C ( z , x , u ) C z ( z , x , u ) .
In this case the DPE or HJB Equation (16) reads with the generator
L t u = a ( z , u ) φ ( z ) + 1 2 b 2 ( z , u ) φ ( z ) ,
and
b 2 ( z , u ) : = 2 C ¯ 0 ( z , u ) C ¯ 2 ( z , u ) , a ( z , u ) : = C ¯ 01 ( z , u ) C ¯ 1 ( z , u ) ,
with:
C ¯ 0 ( z , u ) : = E π ( d x ) C 0 ( z , x , u ) , C 0 ( z , x , u ) : = C ( z , x , u ) R 0 C ( z , x , u ) ,
C ¯ 2 ( z , u ) : = E π ( d x ) C * ( z , x , u ) C ( z , x , u ) , where C * means transpose of the vector C,
C ¯ 01 ( z , u ) : = E π ( d x ) C 01 ( z , x , u ) , C 01 ( z , x , u ) : = C ( z , x , u ) R 0 C z ( z , x , u ) ,
C ¯ 1 ( z , u ) : = E π ( d x ) C 1 ( z , x , u ) , C 1 ( z , x , u ) : = C ( z , x , u ) C z ( z , x , u ) .
Remark 1.
Our construction here is equivalent to some extend to “Recurrent Processes of a semi-Markov type (RPSM)” studied first in [13,14] including limit theorems. Those results were described in more detail in [11,12]. In particular, “RPSM with Markov switching” reflects the case of independent Markov components z k and u k , and “General case of RPSM” reflects the case when u k is dependent on z k .
The Merton Problem
This is an example of solution of DPE/HJB equation for the limiting CGMRP in DA. Let us consider the portfolio optimization problem proposed by Merton (1971), see in [25]. We will apply this approach to the limiting CGMRP in DA above. In this problem, the agent seeks to maximize expected wealth by trading in a risky asset and the risk-free bonds (or bank account). She/he places $ π t for a total wealth X t in the risky asset S 0 ( t ) and looks to obtain the value function (performance criterion)
J π ( t , S , x ) : = sup π U 0 , T E t , S , x [ U ( X T π ) ] ,
which depends on the current wealth x and asset price S , and the optimal trading strategy π , U ( x ) is the agent’s utility function (e.g., exponential ( e γ x ) or power x γ ). We suppose that the asset price S 0 ( t ) satisfies the following SDE
d S 0 ( t ) S 0 ( t ) = ( μ r ) d t + σ a d w t , S 0 ( 0 ) = S ,
where
μ : = 1 2 ( σ a 2 a ^ 2 ) ,
a ^ 2 : = E U π ( d z ) π 1 ( d u ) a 2 ( z , u ) ,
σ a 2 : = E U π ( d z ) π 1 ( d u ) [ a 2 ( z , u ) / 2 + a ( z , u ) R 0 a ( z , u ) ] .
Here, μ represents the expected continuously compounded rate of growth of the traded asset, r is the continuously compounded rate of return of the risk-free asset (bond or bank account).
The wealth process X t π follows the following SDE,
d X t π = ( π t ( μ r ) + r X t π ) d t + π t σ a d w t , X 0 π = x .
From the SDEs for S 0 ( t ) and for X t π above we conclude that the infinitesimal generator for the pair ( S 0 ( t ) , X t π ) is
L t π = ( r x + ( μ r ) π ) x + 1 2 σ a 2 π 2 x 2 + ( μ r ) S S + 1 2 σ a 2 S 2 2 S 2 + σ a π 2 x S .
From HJB equation for the limiting CGRMP in DA it follows that the value function
J ( t , S , x ) = sup π U t , T J π ( t , S , x )
should satisfy the equation
J ( t , S , x ) t + sup π [ L t π J ( t , S , x ) ] = 0
with terminal condition J ( T , S , x ) = U ( x ) .
The explicit solution of this PDE depends on the explicit form of the utility function U ( x ) . Let us take the exponential utility function
U ( x ) = e γ x , γ > 0 , x R .
In this case we can find that the optimal amount to invest in the risky asset is a deterministic function of time
π t * = ( σ a 2 a ^ 2 ) / 2 r γ σ a 2 e r ( T t ) .

5. Rates of Convergence in Averaging and Diffusion Approximations

The rate of convergence in a limit theorem is important in several ways, both theoretical and practical. We present here the rates of convergence of CDTSMRE in the averaging, diffusion approximation and diffusion approximation with equilibrium schemes and, as corollaries, we give the rates of convergence for CAF and CGMRP in the corresponding limits.
Proposition 1.
The Rate of Convergence of CDTSMRE in the Averaging has the following form,
| | E [ Φ [ t / ε ] ε , u φ ] Φ ¯ ( t ) φ | | ε A ( T , φ , | | R 0 | | , D 1 ( z , u ) ) ,
where A ( T , φ , | | R 0 | | , D 1 ( z , u ) ) is a constant, and 0 t T .
The proof of this proposition is given in Section 6.4.
Proposition 2.
The Rate of Convergence of CDTSMRE in the Diffusion Approximation takes the following form,
| | E [ Φ [ t / ε 2 ] ε , u φ ] Φ 0 ( t ) φ | | ε D ( T , | | φ | | , | | R 0 | | , | | D 1 | | , | | D 2 | | ) ,
where D ( T , | | φ | | , | | R 0 | | , | | D 1 | | , | | D 2 | | ) is a constant, and 0 t T .
Proposition 3.
The Rate of Convergence of CDTSMRE in Diffusion Approximation with Equilibrium has the following form,
| | E [ W t ε , u φ ] W t 0 φ | | ε N ( T , | | φ | | , | | R 0 | | , | | D 1 | | , | | D 1 2 | | ) ,
where N ( T , | | φ | | , | | R 0 | | , | | D 1 | | , | | D 1 2 | | ) is a constant and 0 t T .
The proofs of the above Propositions 2 and 3 are similar as the proof of Proposition 1. We give in what follows some rate of convergence results (Corollaries 1 and 2) concerning applications.
Corollary 1.
The Rate of Convergence in the Limit Theorems for CAF:
- Rate of Convergence in Averaging:
| | E y t ε , u y 0 ( t ) | | ε a ( T , | | R 0 | | , | | a | | ) ,
where a ( T , | | R 0 | | , | | a | | ) is a constant, and 0 t T .
- Rate of Convergence in Diffusion Approximation
| | E ξ t ε , u ξ 0 ( t ) | | ε d ( T , | | R 0 | | , | | a | | , | | a 2 | | ) ,
where d ( T , | | R 0 | | , | | a | | , | | a 2 | | ) is a constant, and 0 t T .
- Rate of Convergence in diffusion approximation with equilibrium for CAF
| | E W t ε , u w t | | ε n ( T , | | R 0 | | , | | a | | , | | a 2 | | ) ,
where n ( T , | | R 0 | | , | | a | | , | | a 2 | | ) is a constant, and 0 t T .
Corollary 2.
The Rate of Convergence in the Limit Theorems for CGMRP:
- Rate of Convergence in Averaging
| | E S t ε , u τ ¯ t | | ε a ( T , | | R 0 | | , | | a | | ) ,
where a ( T , | | R 0 | | , | | a | | ) is a constant, and 0 t T .
- Rate of Convergence in Diffusion Approximation
| | E S t ε , u τ 0 ( t ) ) | | ε d ( T , | | R 0 | | , | | a | | , | | a 2 | | ) ,
where d ( T , | | R 0 | | , | | a | | , | | a 2 | | ) is a constant, and 0 t T .
- Rate of Convergence in diffusion approximation with equilibrium
| | E W t ε , u w t | | ε n ( T , | | R 0 | | , | | a | | , | | a 2 | | ) ,
where n ( T , | | R 0 | | , | | a | | , | | a 2 | | ) is a constant, and 0 t T .

6. Proofs

The proofs here have almost the same general construction scheme as in our paper [20] except that we consider also the control process. Let C B [ 0 , ) be the space of B-valued continuous functions defined on [ 0 , ) .

6.1. Proof of Theorem 1

The proof of the relative compactness of CDTSMRE in the average approximation is based on the following four lemmas.
The CDTSMRE Φ [ t / ε ] ε , u φ , see (3), is weakly compact in D B [ 0 , ) with limit points into C B [ 0 , ) .
Lemma 1.
Under Assumptions A1–A7, the limit points of Φ [ t / ε ] ε , u φ , φ B 0 , as ε 0 , belong to C B [ 0 , ) .
Proof. 
Assumptions A5–A6 imply that the discrete-time semi-Markov random evolution Φ k u φ is a contractive operator in H and, therefore, | | Φ k u φ | | H is a supermartingale for any φ H , where | | · | | H is a norm in Hilbert space H ([4,9]) Obviously, the same properties satisfy the following family Φ [ t / ε ] ε , u . Using Doob’s inequality for the supermartingale | | Φ [ t / ε ] ε , u | | H we obtain
P { Φ [ t / ε ] ε , u K Δ } 1 Δ ,
where K Δ is a compact set in B and Δ is any small number. It means that sequence Φ [ t / ε ] ε , u is tight in B . Taking into account conditions A1–A6, we obtain that discrete-time semi-Markov random evolution Φ [ t / ε ] ε , u is weakly compact in D B [ 0 , + ) with limit points in C B [ 0 , + ) , φ B 0 .
Let J t ε , u : = J ( Φ [ t / ε ] ε , u ; [ t / ε ] ) : = sup k [ t / ε ] Φ [ t / ε ] + k ε , u φ Φ [ t / ε ] ε , u φ , and let K Δ be a compact set from compact containment condition Δ > 0 . It is sufficient to show that J t ε , u weakly converges to zero. This is equivalent to the convergence of J t ε , u in probability as ε 0 .
From the very definition of J t ε , u and A3, we obtain
J t ε , u 1 K Δ ε sup k [ t / ε ] sup φ S Δ ( D 1 ( z k , u k ) φ + D 0 ε ( z k , u k ) φ ) ,
where 1 K Δ is the indicator of the set K Δ , and S Δ is the finite δ -set for K Δ . Then, for δ < Δ , we have
P π × π 1 ( J t ε , u 1 K Δ > Δ ) P π × π 1 ( sup k [ t / ε ] D k > ( Δ δ ) / ε ) = i = 1 [ t / ε ] P π × π 1 ( { sup k [ t / ε ] D k > ( Δ δ ) / ε } D i ) ε 2 [ t / ε ] sup φ S Δ [ P ˜ [ t / ε ] ( D 1 ( x , u ) φ 2 + 2 D 1 ( x , u ) φ D 0 ε ( x , u ) φ + D 0 ε ( x , u ) φ 2 ) ] ,
where D k : = sup φ S Δ ( D 1 ( z k , u k ) φ + D 0 ε ( z k , u k ) φ ) , and
D i : = { ω : D k contains the maximum for the first time on the variable D i } .
It is worth noticing that the operator P ˜ k is bounded when k . So is the case for P ˜ [ t / ε ] when ε 0 .
Taking both ε and δ go to 0 we obtain the proof of the this lemma. □
Let us now consider the continuous time martingale
M t ε , u : = M [ t / ε ] ε = Φ [ t / ε ] ε , u I k = 0 [ t / ε ] 1 E π × π 1 [ Φ k + 1 ε , u Φ k ε , u F k ] .
Lemma 2.
The process
M t ε , u : = Φ [ t / ε ] ε , u I = 0 [ t / ε ] 1 [ P ˜ D ε ( · , u ) I ] Φ ε , u ,
is an F [ t / ε ] u -martingale.
Proof. 
As long as M k ε , u : = Φ k ε , u I = 0 k 1 [ P ˜ D ε ( · , u ) I ] Φ ε , u , is a martingale M t ε , u = M [ t / ε ] ε , u is an F [ t / ε ] u -martingale. Here, we have E π × π 1 [ M k + 1 ε F k u ] = M k ε , u which can be easily checked. □
Lemma 3.
The family ( k = 0 [ t / ε ] E π × π 1 [ Φ k + 1 ε , u φ Φ k ε , u φ F k u ] ) is relatively compact for all B 0 * , dual of the space B 0 .
Proof. 
Let
N t ε , u : = k = 0 [ t / ε ] E π × π 1 [ ( Φ k + 1 ε , u Φ k ε , u ) φ F k u ] .
Then,
N t ε , u = k = 0 [ t / ε ] [ P ˜ D ε ( · , u ) I ] Φ k ε , u .
As long as Φ k + 1 ε , u = D ε ( z k + 1 , u k + 1 ) Φ k ε , u , we obtain
E π × π 1 [ Φ k + 1 ε , u φ F k u ] = E π × π 1 [ D ε ( z k + 1 , u k + 1 ) Φ k ε , u φ F k u ] .
Then,
( k = [ t / ε ] + 1 [ ( t + η ) / ε ] E π × π 1 [ Φ k + 1 ε , u φ Φ k ε , u φ F k u ] ) = ( k = [ t / ε ] + 1 [ ( t + η ) / ε ] [ P ˜ D ε ( z k + 1 , z k + 1 ) I ] Φ k ε , u φ ) ε ( [ ( t + η ) / ε ] [ t / ε ] 1 ) P ˜ ( D 1 ( z k + 1 , u k + 1 ) + D 0 ε ( z k + 1 , u k + 1 ) ) φ ε η ε P ˜ ( D 1 ( · , u ) + D 0 ε ( · , u ) ) φ = η P ˜ ( D 1 ( · , u ) + D 0 ε ( · , u ) ) φ 0 , η 0 ,
as P ˜ ( D 1 ( · , u ) + D 0 ε ( · , u ) ) φ is bounded for any φ B 0 .
It means that the family ( k = 0 [ t / ε ] E π × π 1 [ Φ k + 1 ε , u φ Φ k ε , u φ F k u ] ) , is relatively compact for any B 0 * . □
Lemma 4.
The family ( M [ t / ε ] ε , u φ ) is relatively compact for any B 0 * , and any φ B 0 .
Proof. 
It is worth noticing that the martingale M [ t / ε ] ε , u can be represented in the form of the martingale differences
M [ t / ε ] ε , u = k = 0 [ t / ε ] 1 E π × π 1 [ Φ k + 1 ε , u φ E π × π 1 ( Φ k + 1 ε , u φ F k u ) ] .
Then, using the equality
E π × π 1 [ Φ k + 1 ε , u φ F k u ] = E π × π 1 [ D ε ( z k + 1 , u k + 1 ) Φ k ε , u φ F k u ] ,
we get
M [ ( t + η ) / ε ] ε , u φ M [ t / ε ] ε , u φ = k = [ t / ε ] + 1 [ ( t + η ) / ε ] [ D ε ( z k + 1 , u k + 1 ) Φ k ε , u φ E π × π 1 [ D ε ( z k + 1 , u k + 1 ) Φ k ε , u φ F k u ] } = k = [ t / ε ] + 1 [ ( t + η ) / ε ] [ D ε ( z k + 1 , u k + 1 ) Φ k ε , u φ P ˜ D ε ( z k + 1 , u k + 1 ) Φ k ε , u φ ] = k = [ t / ε ] + 1 [ ( t + η ) / ε ] [ D ε ( z k + 1 , u k + 1 ) P ˜ D ε ( z k + 1 , u k + 1 ) ] Φ k ε , u φ ,
for any η > 0 . Now, from the above, we get
E π × π 1 ( M [ ( t + η ) / ε ] ε , u φ M [ t / ε ] ε , u φ ) ( [ t + η ) / ε ] [ t / ε ] ) ε E π × π 1 ( D 1 ( z k + 1 , u k + 1 ) φ + D 0 ε ( z k + 1 , u k + 1 ) φ + P ˜ D 1 ( · , u ) φ + P ˜ D 0 ε ( · , u ) φ ) 2 η ( P ˜ D 1 ( · , u ) φ + P ˜ D 0 ε ( · , u ) φ ) 0 , η 0 ,
which proves the lemma. □
Now the proof of Theorem 1 is achieved as follows.
From Lemmas 2–4 and the representation (17) it follows that the family ( Φ [ t / ε ] ε , u φ ) is relatively compact for any B 0 * , and any φ B 0 .
Moreover, let I L ε ( x ) , x E , be a family of perturbed operators defined on B as follows,
I L ε , u ( x ) : = ε 1 Q ˜ + P ˜ D 1 ( x , u ) + P ˜ D 0 ε ( x , u ) .
Then, the process
M t ε , u = Φ [ t / ε ] ε , u I ε = 0 [ t / ε ] 1 I L ε , u Φ ε , u ,
is an F t ε , u -martingale.
The following singular perturbation problem, for the non-negligible part of compensating operator, I L ε , u , denoted by I L 0 ε , u ( x ) : = ε 1 Q ˜ + P ˜ D 1 ( x , u ) ,
I L 0 ε , u φ ε = I L φ + ε θ ε , u ,
on the test functions φ ε ( z , x ) = φ ( z ) + ε φ 1 ( z , x ) , has the solution (see [3] Proposition 5.1): φ N ( Q ˜ ) , φ 1 = R 0 D ˜ 1 φ , with D ˜ 1 ( x , u ) = P ˜ D 1 ( x , u ) D ^ 1 , D ^ 1 = E π × π 1 ( d x ) D 1 ( x , u ) , and θ ε , u ( x ) = ( P × P u ) D 1 ( x , u ) R 0 D ˜ 1 ( x , u ) φ .
The limit operator is then given by
I L Π = Π D 1 ( · , u ) Π ,
form which we get the contracted limit operator
I L ^ = D ^ 1 .
We note that martingale M t ε , u has the following asymptotic representation,
M t ε , u = Φ [ t / ε ] ε , u I ε = 0 [ t / ε ] 1 I L ^ Φ ε , u + O φ ( ε ) ,
where | | O φ ( ε ) | | 0 , as ε 0 . The families l ( M [ t / ε ] ) and l ( = 0 [ t / ε ] 1 [ ( P × P u ) D ε ( · , u ) I ] Φ ε , u ) are weakly compact for all l B 0 * in a dense subset B 0 * B . It means that family l ( Φ [ t / ε ] ε , u ) is also weakly compact. In this way, the sum ε = 0 [ t / ε ] 1 I L ^ Φ ε , u φ converges, as ε 0 , to the integral 0 t I L ^ Φ ¯ ( s ) φ d s . The quadratic variation of the martingale l ( M t ε , u φ ) tends to zero when ε 0 , thus, M t ε , u φ 0 when ε 0 , for any f B 0 and for any l B 0 * . Passing to the limit in (23), when ε 0 , we get Φ [ t / ε ] ε , u φ ε 0 Φ ¯ ( t ) φ , where Φ ¯ ( t ) is defined in (6).
The quadratic variation of the martingale M t ε , u , in the average approximation, is
( M [ t / ε ] ε , u = k = 0 [ t / ε ] E π × π 1 [ 2 ( M k + 1 ε , u φ ε M k ε , u φ ε ) F k u ] ,
where φ ε ( x ) = φ ( x ) + ε φ 1 ( x ) . Hence
( M k + 1 ε , u φ ε M k ε , u φ ε ) = ( ( M k + 1 ε , u M k ε , u ) φ ) + ε ( ( M k + 1 ε , u M k ε , u ) φ 1 ) ,
and
M k + 1 ε , u M k ε , u = Φ k + 1 ε , u Φ k ε , u E π × π 1 [ Φ k + 1 ε , u Φ k ε , u F k u ] .
Therefore,
( M k + 1 ε , u φ ε M k ε , u φ ε ) = ( ( D ( z k + 1 , u k + 1 ) ε I ) Φ k ε , u φ ) E π × π 1 [ ( D ε ( z k + 1 , u k + 1 ) I ) φ F k u ] + ε ( ( D ε ( z k + 1 , u k + 1 ) I ) φ 1 ) E π × π 1 [ ( D ( z k + 1 , u k + 1 ) ε I ) φ 1 F k u ] = ε ( ( D 1 ( z k + 1 , u k + 1 ) + D 0 ε ( z k + 1 , u k + 1 ) ) φ ) ε E π × π 1 [ ( D 1 ( z k + 1 , u k + 1 ) + D 0 ε ( z k + 1 , u k + 1 ) ) φ F k u ] + ε 2 ( ( D 1 ( z k + 1 , u k + 1 ) + D 0 ε ( z k + 1 , u k + 1 ) ) φ 1 ) ε 2 E π [ ( D 1 ( z k + 1 , u k + 1 ) + D 0 ε ( z k + 1 , u k + 1 ) ) φ 1 F k u ] .
Now, from (24) and (26) and from boundedness of all operators in (26) with respect to E π × π 1 , it follows that ( M [ t / ε ] ε , u goes to 0 when ε 0 , and the quadratic variation of limit process M t 0 , u , for the martingale M t ε , u , is equals to 0.
In this case, the limit martingale M t 0 equals to 0. Therefore, the limit equation for M t ε , u has the form (6). As long as the solution of the martingale problem for operator I L ^ is unique, then it follows that the solution of the Equation (6) is unique as well [49,50]. It is worth noticing that operator I L ^ is a first order operator ( D ^ 1 , see (22)). Finally, the operator I L ^ generates a semigroup, then Φ ¯ ( t ) φ = exp [ I L ^ t ] φ and the latter representation is unique.

6.2. Proof of Theorem 2

We can prove the relative compactness of the family Φ [ t / ε 2 ] ε , u exactly on the same way, and following the same steps as above. However, in the case of diffusion approximation the limit continuous martingale M 0 ( t ) for the martingale M t ε has quadratic variation that is not zero, that is,
M 0 ( t ) φ = Φ 0 ( t ) φ φ 0 t I L ^ Φ 0 ( s ) d s ,
and so ( M 0 ) 0 , for B 0 * .
Moreover, operator I L ^ defined in Theorem 2 is a second-order kind operator as it contains operator D ^ 2 and Π D 1 R 0 P ˜ D 1 Π , compare with the first-order operator I L ^ in (7).
Let I L ε , u ( x ) , x E , be a family of perturbed operators defined on B as follows,
I L ε , u ( x ) : = ε 2 Q ˜ + ε 1 P ˜ D 1 ( x , u ) + P ˜ D 2 ( x , u ) + P ˜ D 0 ε ( x , u ) .
Then, the process
M t ε , u = Φ [ t / ε ε ] ε , u I ε 2 k = 0 [ t / ε 2 ] 1 I L ε , u Φ k ε , u ,
is an F t ε , u -martingale with mean value zero.
For the non-negligible part of compensating operator, I L ε , u , denoted by I L 0 ε , u ( x ) : = ε 2 Q ˜ + ε 1 P ˜ D 1 ( x , u ) + P ˜ D 2 ( x , u ) , consider the following singular perturbation problem,
I L 0 ε , u φ ε = I L φ + ε θ ε , u ( x ) ,
where φ ε ( z , x ) = φ ( z ) + ε φ 1 ( z , x ) + ε 2 φ 2 ( z , x ) . The solution of this problem is realized by the vectors (see in [3], Proposition 5.2)
φ 1 = R 0 P ˜ D 1 ( x , u ) φ , φ 2 = R 0 A ˜ φ ,
with A ˜ ( x , u ) : = A ( x , u ) A ^ . Finally, the negligible term θ ε , u ( x ) is
θ ε , u ( x ) = [ P ˜ D 1 ( x , u ) + ε P ˜ D 2 ( x , u ) ] φ 2 + P ˜ D 2 ( x , u ) φ 1 .
Of course, φ N ( Q ˜ ) .
Now the limit operator I L is given by
I L = P ˜ D 2 ( · , u ) + P ˜ D 1 ( · , u ) R 0 P ˜ D 1 ( · , u ) ,
from which, the contracted operator on the null space N ( Q ˜ ) is
I L ^ = D ^ 2 Π + Π D 1 ( x , u ) R 0 P ˜ D 1 ( x , u ) Π .
Moreover, due to the balance condition (8) we get the limit operator.
We worth noticing that Assumptions A5–A7 and D1–D3 imply that discrete-time semi-Markov random evolution Φ [ t / ε 2 ] ε , u φ is a contractive operator in H and, therefore, | | Φ [ t / ε 2 ] ε , u φ | | H is a supermartingale for any φ H , where | | · | | H is a norm in Hilbert space H ([4,9]). By Doob’s inequality for the supermartingale | | Φ [ t / ε 2 ] ε , u | | H we obtain
P { Φ [ t / ε 2 ] ε , u K Δ 1 } 1 Δ ,
where K Δ 1 is a compact set in B and Δ is any positive small real number.
We conclude that under Assumptions A5–A7 and D1–D3, the family M t ε , u is tight and is weakly compact in D B [ 0 , + ) with limit points in C B [ 0 , + ) .
Moreover, under Assumptions A5–A6 and D1–D2, the martingale M t ε , u has the following asymptotic presentation:
M t ε , u φ = Φ [ t / ε 2 ] ε , u φ φ ε 2 k = 0 [ t / ε 2 ] 1 I L ^ Φ k ε , u φ + O φ ( ε ) ,
where | | O φ ( ε ) | | 0 , as ε 0 . The families l ( M t ε , u ϕ ) and l ( ε 2 k = 0 [ t / ε ε ] 1 I L ^ Φ k ε , u φ ) are weakly compact for all l B * and φ B 0 . It means that Φ [ t / ε 2 ] ε , u is also weakly compact and has a limit.
Let us denote the previous limit by Φ 0 ( t ) , then the sum ε 2 k = 0 [ t / ε ε ] 1 I L ^ Φ k ε , u φ converges to the integral 0 t I L ^ Φ 0 ( s ) φ d s . Let M 0 ( t ) also be a limit martingale for M t ε , u when ε 0 . Then, from the previous steps and (32), we obtain
M 0 ( t ) φ = Φ 0 ( t ) φ φ 0 t I L ^ Φ 0 ( s ) φ d s .
As long as martingale M t ε , u has mean value zero, the martingale M 0 ( t ) has also mean value zero. If we take the mean value from both parts of (33) we get
0 = E Φ 0 ( t ) φ φ 0 t I L ^ E Φ 0 ( t ) φ d s ,
or, solving it, we get
E Φ 0 ( t ) φ = exp [ I L ^ t ] φ .
The last equality means that the operator I L ^ generates a semigroup, namely, U ( t ) : = E Φ 0 ( t ) φ = exp [ I L ^ t ] φ . Now, the uniqueness of the limit evolution Φ 0 ( t ) in diffusion approximation approximation follows from the uniqueness of solution of the martingale problem for Φ 0 ( t ) (uniqueness of the limit process under weak compactness). As long as the solution of the martingale problem for operator I L ^ is unique, then it follows that the solution of the Equation (34) is unique as well [49,50].

6.3. Proof of Theorem 3

We note that W t ε , u in (12) has the following presentation,
W t ε , u = ε 1 / 2 { k = 1 [ t / ε ] [ D ε ( z k 1 , u k 1 ) I ] Φ k ε , u 0 t D ¯ 1 Φ ¯ ( s ) d s } .
As the balance condition Π ( D 1 D ^ 1 ) = 0 , holds, then we apply the diffusion approximation algorithm (see Section 3.2), i.e., to the right-hand side of (36) with the following operators, D 2 = 0 and ( D 1 ( z ) D ¯ 1 ) instead of D 1 ( z ) . It is worth mentioning that the family W t ε , u is weakly compact and the result is proved (see Section 6.1 and Section 6.2).

6.4. Proof of Proposition 1

The proof of this proposition is based on the estimation of
| | E π [ Φ [ t / ε ] ε , u φ ε ] Φ ¯ ( t ) φ | | ,
for any φ B 0 , where φ ε ( x ) = φ ( x ) + ε φ 1 ( x ) .
We note that
( P ˜ I ) φ 1 ( x ) = ( D ^ 1 P ˜ D 1 ( x , u ) ) φ .
As long as Π ( D ^ 1 ( P × P u ) D 1 ( x , u ) ) φ = 0 , φ B 0 , Equation (37) has the solution in domain R ( P ˜ I ) , φ 1 ( x ) = R 0 D ˜ 1 φ .
In this way,
E π × π 1 φ 1 ( x ) 2 R 0 E U π ( d z ) π 1 ( d u ) P ˜ D 1 ( · , u ) φ : = 2 C 1 ( φ 1 R 0 ) ,
where R 0 is a potential operator of Q ˜ : = P ˜ I .
From here we obtain
E π × π 1 ( Φ [ t / ε ] ε , u I ) φ 1 4 C 1 ( φ 1 R 0 ) ,
as Φ k ε , u are contractive operators.
We note also that
E π × π 1 [ ε k = 0 [ t / ε ] I L ^ Φ k ε , u φ 0 t I L ^ Φ ¯ ( s ) φ d s ] ε C 2 ( t , φ ) ,
where C 2 ( t , φ ) : = 4 T E U π ( d z ) π 1 ( d u ) P ˜ D 1 ( · , u ) φ , t [ 0 , T ] . This follows from standard argument about the convergence of Riemann sums in Bochner integral (see Lemma 4.14, p. 161, [4]).
We note that
| | E π × π 1 [ Φ [ t / ε ] ε , u φ ε ] Φ ¯ ( t ) φ | | | | E π × π 1 [ Φ [ t / ε ] ε , u φ Φ ¯ ( t ) φ | | + ε C 1 ( φ 1 R 0 ) ,
where we applied representation φ ε = φ + ε φ 1 .
We also note that Φ ¯ ( t ) satisfies the equation
Φ ¯ ( t ) φ φ 0 t I L ^ Φ ¯ ( s ) φ d s = 0 .
Let us introduce the following martingale,
M [ t / ε ] + 1 ε , u φ ε : = Φ [ t / ε ] ε , u φ ε φ ε k = 0 [ t / ε ] E π × π 1 [ Φ k + 1 ε , u φ ε Φ k ε , u φ ε F k u ] .
This is of zero mean-value martingale
E π × π 1 M [ t / ε ] ε , u φ ε = 0 ,
which comes directly from (43).
Again, from (43), we get the following asymptotic representation
M [ t / ε ] ε , u φ ε = Φ [ t / ε ] ε , u φ φ + ε [ Φ [ t / ε ] I ] φ 1 ε k = 0 [ t / ε ] I L ^ Φ k ε , u φ ε 2 k = 0 [ t / ε ] [ P ˜ D 1 ( · , u ) Φ k ε , u φ 1 + o φ ( 1 ) ] ,
where o φ ( 1 ) 0 , as ε 0 , for any φ B 0 .
Now, from Equation (6) and expressions (44) and (45), we obtain the following representation
E π × π 1 [ Φ [ t / ε ] ε , u φ Φ ¯ ( t ) φ ] = ε E π × π 1 [ Φ [ t / ε ] ε , u I ] φ 1 + E π × π 1 [ ε k = 0 [ t / ε ] I L ^ Φ k ε , u φ 0 t I L ^ Φ ¯ ( s ) φ d s ] + ε 2 E π × π 1 [ k = 0 [ t / ε ] 1 R k u ( φ 1 ) ] ,
where R k u ( φ 1 ) : = P ˜ D 1 ( · , u ) Φ k ε , u φ 1 + o φ ( 1 ) .
Let us estimate R k u ( φ 1 ) in (46).
R k u ( φ 1 ) sup g K Δ ( P ˜ D 1 ( z , u ) g + o g ( 1 ) ) : = C 2 ( z , g , K Δ , u ) ,
where K Δ is a compact set, Δ > 0 ,because Φ k ε , u φ 1 satisfies compactness condition for any ε > 0 and any k.
In this way, we get from (46) that
E π × π 1 [ k = 0 [ t / ε ] 1 R k u ( φ 1 ) ] T E π ( d z ) π 1 ( d u ) C 3 ( z , g , K Δ , u ) , t [ 0 , T ] .
Finally, from inequalities (38)–(41) and from (47)–(48), we obtain the desired rate of convergence of the CDTSMRE in averaging scheme
| | E π × π 1 [ Φ [ t / ε ] ε , u φ ε ] Φ ¯ ( t ) φ | | ε A ( T , φ , R 0 , D 1 ( z , u ) ) ,
where the constant
A ( T , φ , R 0 , D 1 ( z , u ) ) : = 5 C 1 ( φ , R 0 ) + C 2 ( T , φ ) + T E π ( d z ) π 1 ( d u ) C 3 ( z , g , K Δ , u ) ,
and C 3 ( z , g , K Δ , u ) is defined in (48). Therefore, the proof of Proposition 1 is done.
Remark 2.
In a similar way, we can obtain the rate of convergence results in diffusion approximation (see Propositions 2–3).

7. Concluding Remarks and Future Work

In this paper, we introduced controlled semi-Markov random evolutions in discrete-time in Banach space. The main results concerned time-rescaled limit theorems, namely, averaging, diffusion approximation, and diffusion approximation with equilibrium by martingale weak convergence method. We applied these results to various important families of stochastic systems, i.e., the controlled additive functionals, controlled geometric Markov renewal processes, and controlled dynamical systems. We provided dynamical principles for discrete-time dynamical systems such as controlled additive functionals and controlled geometric Markov renewal processes. We also produced dynamic programming equations (Hamilton–Jacobi–Bellman equations) for the limiting processes in diffusion approximation such as CAF, CGMRP, and CDS. As an example, we considered the solution of portfolio optimization problem by Merton for the limiting CGMRP in DA. We also point out the importance of convergence rates and obtained them in the limit theorems for CDTSMRE and CAF, CGMRP, and CDS.
The future work will be associated with the study of optimal control for the initial, not limiting models, such as CAF in Section 4.1, CGMRP in Section 4.2, and CDS in Section 4.3. Other optimal control problems would be also interesting to consider for diffusion models with equilibrium, e.g., CAF in Section 4.1 and CGMRP in Section 4.2. In our future work, the latter models will be considered for solutions of Merton portfolio’s problems as well. We will also consider in our future research the case of dependent SMC z k and the MC u k .

Author Contributions

These authors contributed equally to this work. All authors have read and agreed to the published version of the manuscript.

Funding

The research of the first author is partially supported by NSERC.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Acknowledgments

We thank to four anonymous referees for valuable remarks and suggestions that improved the paper. The research of the first author is partially supported by NSERC. The first author also thanks to the Laboratory of Applied Mathematics of the Université de Technologie de Compiègne, Compiègne, France, very much for their hospitality during his visit.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

SMCDiscrete-time semi-Markov chain;
DTSMREDiscrete-time semi-Markov random evolution;
CDTSMREControlled discrete-time semi-Markov random evolution;
CGMRPControlled geometric Markov renewal processes;
CAFControlled additive functionals;
CDSControlled dynamical systems;
HJBHamilton–Jacobi–Bellman (equation);
DPEDynamic programming equation;
DPPDynamic programming principle;
DADiffusion approximation;
SDEStochastic differential equation.

References

  1. Griego, R.; Hersh, R. Random evolutions, Markov chains, and Systems of partial differential equations. Proc. Natl. Acad. Sci. USA 1969, 62, 305–308. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Hersh, R. Random evolutions: A Survey of results and problems. Rocky Mountain J. Math. 1974, 4, 443–477. [Google Scholar] [CrossRef]
  3. Koroliuk, V.S.; Limnios, N. Stochastic Systems in Merging Phase Space; World Scientific: Singapore, 2005. [Google Scholar]
  4. Korolyuk, V.S.; Swishchuk, A. Evolution of System in Random Media; CRC Press: Boca Raton, FL, USA, 1995. [Google Scholar]
  5. Swishchuk, A.; Wu, J. Evolution of Biological Systems in Random Media: Limit Theorems and Stability; Kluwer: Dordrecht, The Netherlands, 2003. [Google Scholar]
  6. Swishchuk, A.V. Random Evolutions and their Applications; Kluwer AP: Dordrecht, The Netherlands, 1995. [Google Scholar]
  7. Cohen, J.E. Random Evolutions in Discrete and Continuous Time. Stoch. Proc. Appl. 1979, 9, 245–251. [Google Scholar] [CrossRef] [Green Version]
  8. Keepler, M. Random evolutions processes induced by discrete time Markov chains. Port. Math. 1998, 55, 391–400. [Google Scholar]
  9. Limnios, N. Discrete-time random evolutions—Difference equations and additive functionals. Commun. Statist. Theor. Methods 2011, 40, 1–11. [Google Scholar] [CrossRef]
  10. Yin, G.G.; Zhang, Q. Discrete-Time Markov Chains. Two-Time-Scale Methods and Applications; Springer: New York, NY, USA, 2005. [Google Scholar]
  11. Anisimov, V.V. Switching processes: Averaging principle, diffusion approximation and applications. Acta Appl. Math. 1995, 40, 95–141. [Google Scholar] [CrossRef]
  12. Anisimov, V.V. Switching Processes in Queueing Models; J. Wiley & Sons: London, UK; ISTE: London, UK, 2008. [Google Scholar]
  13. Anisimov, V.V. Averaging principle for switching recurrent sequences. Theory Probab. Math. Stat. 1991, 45, 1–8. [Google Scholar]
  14. Anisimov, V.V. Averaging principle for switching processes. Theory Probab. Math. Stat. 1992, 46, 1–10. [Google Scholar]
  15. Swishchuk, A.V.; Islam, M.S. The Geometric Markov Renewal Processes with application to Finance. Stoch. Anal. Appl. 2010, 29, 4. [Google Scholar] [CrossRef]
  16. Swishchuk, A.V.; Islam, M.S. Diffusion Approximations of the Geometric Markov Renewal Processes and option price formulas. Int. J. Stoch. Anal. 2010, 2010, 347105. [Google Scholar] [CrossRef] [Green Version]
  17. Swishchuk, A.V.; Islam, M.S. Normal Deviation and Poisson Approximation of GMRP. Communic. Stat. Theory Methods 2013, 1488–1501. [Google Scholar] [CrossRef]
  18. Swishchuk, A.V.; Limnios, N. Optimal stopping of GMRP and pricing of European and American options. In Proceedings of the 15th Intern. Congress on Insurance: Mathematics and Economics (IME 2011), Trieste, Italy, 14–17 June 2011. [Google Scholar]
  19. Barbu, V.; Limnios, N. Semi-Markov Chains and Hidden Semi-Markov Models. Toward Applications. Their Use in Reliability and DNA Analysis; Lecture Notes in Statistics; Springer: New York, NY, USA, 2008; Volume 191. [Google Scholar]
  20. Limnios, N.; Swishchuk, A. Discrete-time Semi-Markov Random Evolutions and their Applications. Adv. Appl. Probab. 2013, 45, 214–240. [Google Scholar] [CrossRef]
  21. Jacod, J.; Shiryaev, A.N. Limit Theorems for Stochastic Processes; Springer: Berlin/Heidelberg, Germany, 1987. [Google Scholar]
  22. Bertsekas, D.; Shreve, S. Stochastic Optimal Control: The Discrete-Tiime Case; Athena Scientific Publisher: Belmont, MA, USA, 1996. [Google Scholar]
  23. Fleming, W.H.; Rishel, R.W. Deterministic and Stochastic Optimal Control; Springer: New York, NY, USA, 1975. [Google Scholar]
  24. Kushner, H.J. Introduction to Stochastic Control Theory; Holt, Rinehart, Winston: New York, NY, USA, 1971. [Google Scholar]
  25. Merton, R. Optimum consumption and portfolio rules in a continuous-time model. J. Econ. Theory 1971, 3, 373–413. [Google Scholar] [CrossRef] [Green Version]
  26. Sviridenko, M.N. Martingale approach to limit theorems for semi-Markov processes. Theor. Probab. Appl. 1986, 34, 540–545. [Google Scholar] [CrossRef]
  27. Adams, R. Sobolev Spaces; Academic Press: New York, NY, USA, 1979. [Google Scholar]
  28. Rudin, W. Functional Analysis; McGraw-Hill: New York, NY, USA, 1991. [Google Scholar]
  29. Sobolev, S. Some Applications of Functional Analysis in Mathematical Physics, 3rd ed.; American Mathematical Society: Providence, RI, USA, 1991; Volume 90. [Google Scholar]
  30. Pyke, R. Markov renewal processes: Definitions and preliminary properties. Ann. Math. Statist. 1961, 32, 1231–1242. [Google Scholar] [CrossRef]
  31. Pyke, R. Markov renewal processes with finitely many states. Ann. Math. Statist. 1961, 32, 1243–1259. [Google Scholar] [CrossRef]
  32. Pyke, R.; Schaufele, R. Limit theorems for Markov renewal processes. Ann. Math. Statist. 1964, 35, 1746–1764. [Google Scholar] [CrossRef]
  33. Shurenkov, V.M. On the theory of Markov renewal. Theor. Probab. Appl. 1984, 19, 247–265. [Google Scholar] [CrossRef]
  34. Maxwell, M.; Woodroofe, M. Central limit theorems for additive functionals of Markov chains. Ann. Probab. 2000, 28, 713–724. [Google Scholar]
  35. Nummelin, E. General Irreducible Markov Chains and Non-Negative Operators; Cambridge University Press: Cambridge, UK, 1984. [Google Scholar]
  36. Revuz, D. Markov Chains; North-Holland: Amsterdam, The Netherlands, 1975. [Google Scholar]
  37. Skorokhod, A.V. Asymptotic Methods in the Theory of Stochastic Differential Equations; AMS: Providence, RI, USA, 1989; Volume 78. [Google Scholar]
  38. Skorokhod, A.V.; Hoppensteadt, F.C.; Salehi, H. Random Perturbation Methods with Applications in Science and Engineering; Springer: New York, NY, USA, 2002. [Google Scholar]
  39. Jaśkiewicz, A.; Nowak, A. Average optimality for semi-Markov control processes. Morfismos 2007, 11, 15–36. [Google Scholar]
  40. Vega-Amaya, O.; Luque-Vásquez, F. Sample-path average cost optimality for semi-Markov control processes on Borel spaces: Unbounded cost and mean holding times. Appl. Math. 2000, 27, 343–367. [Google Scholar] [CrossRef] [Green Version]
  41. Altman, E.; Shwartz, A. Markov decision problems and state-action frequencies. SIAM J. Control. Optimikzation 1991, 29, 786–809. [Google Scholar] [CrossRef]
  42. Beutler, F.; Ross, K. Optimal policies for controlled Markov chains with a constraint. J. Mathem. Analysis Appl. 1985, 112, 236–252. [Google Scholar] [CrossRef] [Green Version]
  43. Borkar, V. Dynamic programming for ergodic control with partial observations. Stoch. Proc. Appl. 2003, 103, 293–310. [Google Scholar] [CrossRef]
  44. Boussement, M.; Limnios, N. Markov decision processes with asymptotic average failure rate constraint. Communic. Statis. Theory Methods 2004, 33, 1689–1714. [Google Scholar] [CrossRef]
  45. Hernández-Lerma, O.; Lasserre, J.B. Discrete-Time Markov Control Processes; Springer: New York, NY, USA, 1996. [Google Scholar]
  46. Krylov, N.; Bogolyubov, N. Introduction to Non-linear Mechanics; Princeton University Press: Princeton, NJ, USA, 1947. [Google Scholar]
  47. Limnios, N.; Oprişan, G. Semi-Markov Processes and Reliability; Birkhäuser: Boston, MA, USA, 2001. [Google Scholar]
  48. Cartea, A.; Jaimungal, S.; Penalva, J. It Algorithmic and High-Frequency Trading; Cambridge University Press: Cambridge, UK, 2015. [Google Scholar]
  49. Ethier, S.N.; Kurtz, T.G. Markov Processes: Characterization and Convergence; J. Wiley: New York, NY, USA, 1986. [Google Scholar]
  50. Stroock, D.W.; Varadhan, S.R.S. Multidimensional Diffusion Processes; Springer: Berlin/Heidelberg, Germany, 1979. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Swishchuk, A.; Limnios, N. Controlled Discrete-Time Semi-Markov Random Evolutions and Their Applications. Mathematics 2021, 9, 158. https://doi.org/10.3390/math9020158

AMA Style

Swishchuk A, Limnios N. Controlled Discrete-Time Semi-Markov Random Evolutions and Their Applications. Mathematics. 2021; 9(2):158. https://doi.org/10.3390/math9020158

Chicago/Turabian Style

Swishchuk, Anatoliy, and Nikolaos Limnios. 2021. "Controlled Discrete-Time Semi-Markov Random Evolutions and Their Applications" Mathematics 9, no. 2: 158. https://doi.org/10.3390/math9020158

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop