Next Article in Journal
Improved Bounds for Integral Jensen’s Inequality Through Fifth-Order Differentiable Convex Functions and Applications
Previous Article in Journal
Characterization of the Best Approximation and Establishment of the Best Proximity Point Theorems in Lorentz Spaces
Previous Article in Special Issue
The m-CCE Inverse in Minkowski Space and Its Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Operator Newton Method for Large-Scale Coupled Riccati Equations Arising from Jump Systems

School of Science, Hunan University of Technology, Zhuzhou 412007, China
*
Author to whom correspondence should be addressed.
Axioms 2025, 14(8), 601; https://doi.org/10.3390/axioms14080601 (registering DOI)
Submission received: 27 June 2025 / Revised: 26 July 2025 / Accepted: 28 July 2025 / Published: 1 August 2025
(This article belongs to the Special Issue Advances in Linear Algebra with Applications, 2nd Edition)

Abstract

Consider a class of coupled discrete-time Riccati equations arising from jump systems. To compute their solutions when systems reach a steady state, we propose an operator Newton method and correspondingly establish its quadratic convergence under suitable assumptions. The advantage of the proposed method lies in the fact that its subproblems are solved using the operator Smith method, which allows it to maintain quadratic convergence in both the inner and outer iterations. Moreover, it does not require the constant term matrix of the equation to be invertible, making it more broadly applicable than existing inverse-free iterative methods. For large-scale problems, we develop a low-rank variant by incorporating truncation and compression techniques into the operator Newton framework. A complexity analysis is also provided to assess its scalability. Numerical experiments demonstrate that the presented low-rank operator Newton method is highly effective in approximating solutions to large-scale structured coupled Riccati equations.

1. Introduction

Consider a discrete-time jump control system [1,2,3], modeled by:
x k + 1 = A i x k + B i u k , y k = C i x k , i = 1 , , m , k = 1 , 2 , ,
where A i R N × N , B i R N × m b , and C i R m c × N , with m b , m c N . To obtain the optimal control of the above system, one has to minimize the quadratic cost function
J ( x , u ) = k = 1 x k Q i x k + u k R i u k ,
where Q i = C i C i 0 and R i > 0 are symmetric positive semi-definite (SPSD) and symmetric positive definite (SPD) matrices, respectively, representing the state and control weighting terms in the cost function. The corresponding optimal control feedback gain is given by:
u k = ( R i + B i E i ( X ) B i ) 1 B i E i ( X ) A i x k ,
where E i ( X ) = j = 1 m p i j X j R N × N is defined as a convex combination of X j via the stochastic weights p i j . To achieve the optimal control u k , one has to solve the coupled discrete-time algebraic Riccati equation (CDARE):
D C i ( X ) = X i + A i E i ( X ) A i + Q i A i E i ( X ) B i R i + B i E i ( X i ) B i 1 B i E i ( X ) A i = 0 ,
for i = 1 , , m .
Among the various approaches developed for solving the CDARE (1), iterative methods remain prominent. Commonly used iterations typically reformulate the problem as either optimization-based methods [4] or fixed-point iterations [5,6], drawing on classical linear–quadratic regulator theory. Enhanced variants based on linear matrix inequality (LMI) formulations have also been proposed [7,8]. However, such methods often exhibit slow convergence of the objective function and limited numerical precision, particularly in large-scale or tightly coupled scenarios. Fixed-point schemes address the CDARE directly by recasting it as a fixed-point problem. Notably, Ivanov [9] introduced two distinct fixed-point iterations, which were later accelerated by incorporating extrapolation techniques that leverage information from the current iteration to replace the previous one, thereby obviously improving the convergence rate.
To circumvent the computational cost associated with explicit matrix inversion, inverse-free fixed-point methods have been introduced. These schemes are inspired by the Schulz iteration [10], which recasts matrix inversion as a Newton-type process, replacing inversions with matrix multiplications [11,12,13]. While inverse-free methods demonstrate notable computational efficiency in practice, their theoretical convergences are generally linear.
To further accelerate convergence, Newton-type methods have garnered attention. For continuous-time coupled Riccati equations, Feng and Chu [14] proposed a Newton-based approach that extends the block-diagonal structure originally developed in [15] and generalizes the pseudo-Newton strategies given in [13]. For the discrete-time case in (1), Newton-type iteration was explored in [16], though the convergence analysis therein is restricted to settings where each weighting matrix Q i is symmetric positive definite. In practical large-scale systems, however, the output matrix C i often has low rank, rendering the associated Q i only positive semi-definite. In such a case, the above Newton variants, including the inverse-free version [17], become inapplicable. For other methods applicable to similar types of equations, readers may also refer to [18,19].
The key to applying Newton’s method to large-scale CDARE (1) lies in efficiently computing the solution of large-scale coupled Stein equations (CSEs). Recently, a novel operator Smith algorithm (OSA) was introduced for solving CSEs with SPSD constant matrices [20]. This method interprets the coupling unknowns as operator-valued expressions and employs a doubling strategy to accelerate the iteration, demonstrating well-behaved numerical performance on large-scale problems. Another approach to handling large-scale CSEs with low-rank structure is to represent the low-rank matrix in the HODLR structured form [21,22]. However, after performing matrix operations with sparse matrices, the HODLR-structured matrix requires restructuring; otherwise, the dense matrix will not be adapted to large-scale computations.
Inspired by the development of OSA in [20], we propose an operator Newton method (ONM) tailored to solving large-scale CDAREs (1). The main contributions are summarized as follows:
  • We develop an operator Newton iteration scheme grounded in the structure of the OSA for solving the CDARE (1) and rigorously establish the convergence as well as the convergence rate. Crucially, unlike existing inverse-free schemes [11,12,13] that require invertible Q i to initiate the iteration, our method allows for SPSD initial Q i .
  • A low-rank variant of the operator Newton method is constructed to address the large-scale system. When the matrix Q i admits low-rank representation, the proposed method ensures that the rank of the initial residual remains fixed across iterations, effectively mitigating rank inflation.
  • We employ a doubling-based operator formulation for Newton’s subproblem, i.e., the coupled Stein equations, and embed the truncation–compression (TC) technique to control the growth of the column of low-rank factors. This enables efficient low-rank approximation to the solution without compromising numerical stability.
  • We propose a scalable residual evaluation strategy for large-scale CDARE and validate the proposed method on practical problems from engineering applications. Numerical experiments demonstrate that, for a comparable level of residual accuracy, the presented operator Newton method significantly reduces CPU time relative to the standard Newton’s method with the incorporation of the HODLR structure [21,22].
The paper is structured as follows. Section 2 reviews the operator Smith algorithm, and presents several lemmas required for constructing the operator Newton method. Section 3 introduces the iterative scheme of the operator Newton method for solving the CDAREs and establishes corresponding theorems on the convergence and convergence rate. Section 4 develops a low-rank variant of the operator Newton method tailored for large-scale problems with low-rank structures. By incorporating the truncation and compression technique, the scheme effectively controls the growth of the iterative matrix sequence. A detailed analysis of the computational complexity per iteration is also provided. Section 5 demonstrates the effectiveness of the proposed operator Newton method in solving large-scale practical CDAREs.

2. Preliminaries

In this section, we begin by reviewing the iterative framework of the operator Smith algorithm (OSA), which serves as the foundation for solving the subproblems. We then establish several useful lemmas that underpin the convergence analysis of the proposed method.

2.1. Operator Smith Algorithm for Coupled Discrete-Time Stein Equations

The operator Smith algorithm [20] for solving the coupled discrete-time Stein equations
X i A i E i ( X ) A i W i = 0 , i = 1 , , m
is given by the iterative scheme:
X i ( k + 1 ) = X i ( k ) + F i ( k ) ( X ( k ) ) , X i ( 0 ) = W i , k = 0 , 1 , 2 , ,
where the operator F i ( k ) at the k-th iteration is defined as
F i ( k ) ( · ) = A i E i A E ( A E 2 k 1 ( · ) A ) A 2 k 1 A i : = A i E i ( A E ) 2 k 1 ( · ) A 2 k 1 A i
with A E ( · ) A = j = 1 m A j E i ( · ) A j being a linear combination about j. After the k-th iteration, the expression of the iteration matrix is provided by the following theorem.
Theorem 1
([20]). Let E i ( W ) = s = 1 m p i , s W s . Then the k-th iteration X i ( k ) produced by (3) admits the representation
X i ( k ) = W i + A i E i j = 0 2 k 2 ( A E ) j ( W ) A j A i , i = 1 , , m ,
and the solution to (2) is given by
X i * = W i + A i E i j = 0 ( A E ) j ( W ) A j A i .
When each A i is d-stable, the following theorem establishes the quadratic convergence.
Theorem 2
([20]). Let ρ : = max i = 1 m ρ ( A i ) and p : = max i , j = 1 m p i j such that m p ρ m < 1 . Then, for any matrix norm, the following bound holds:
X i ( k ) X i * ( m p ρ m ) 2 k 1 m p ρ m W ,
where W : = max i = 1 m W i .

2.2. Some Lemmas

In this subsection, we introduce some lemmas which are useful for constructing the convergence of the operator Newton method.
Lemma 1.
Let A i be d-stable. Let L i ( S ) = j = 1 m c i j S j be a linear operator with c i j > 0 . Define
V i = S i A i L i ( S ) A i .
Then, S i 0 if V i 0 for all i = 1 , , m .
Proof. 
From (7), it follows that
S i = V i + A i L i j = 0 ( A L ) j ( V ) A j A i .
Since c i j > 0 for all i , j = 1 , , m , the SPSD V i implies that S i 0 .    □
Lemma 2.
Let A and B be SPSD matrices. If I + A B is nonsingular, then
( I + A B ) 1 2 1 .
Proof. 
Let y = ( I + A B ) 1 x for an arbitrary nonzero vector x. Then x = y + A B y , and hence
x 2 2 = y ( I + A B ) ( I + A B ) y   y y + y ( A B + B A ) y + y ( B A 2 B ) y   y 2 2 .
Therefore,
( I + A B ) 1 2 2 = max x 0 y 2 x 2 2 1 .
  □
Lemma 3.
The Sherman–Morrison–Woodbury formula [23] is
A + U C V 1 = A 1 A 1 U C 1 + V A 1 U 1 V A 1 ,
where A R n × n , U R n × k , C R k × k , and V R k × n are conformable matrices.

3. Operator Newton Method

In this section, we first introduce the iterative framework of the operator Newton method (ONM) for solving the CDARE (1), and then analyze its convergence.

3.1. Iteration Format

The Fréchet derivative of the nonlinear operator D C i ( X ) in (1) corresponds to the linear part of the increment D C i ( X + H ) D C i ( X ) , and is given by:
D C i ( X ) ( H ) = H i A i ( I + E i ( X ) B i R i 1 B i ) 1 E i ( H ) ( I + B i R i 1 B i E i ( X ) ) 1 A i ,
assuming all matrix inverses exist.
Given X i ( 0 ) , Newton’s method based on the above derivative, can be written as
D C i ( X ( k ) ) X ( k + 1 ) X ( k ) = D C i ( X ( k ) ) ,
which yields a sequence of linear subproblems of the form
X i ( k + 1 ) ( A ^ i ( k ) ) E i ( X ( k + 1 ) ) A ^ i ( k ) = Q i + L i ( k ) R i ( L i ( k ) ) , i = 1 , , m ,
where
A ^ i ( k ) = A i B i ( L i ( k ) ) , L i ( k ) = A i E i ( X ( k ) ) B i R i + B i E i ( X ( k ) ) B i 1 ,
and
E i ( X ( k ) ) = s = 1 m p i s X s ( k ) .
Obviously, each subproblem corresponds to coupled Stein equations (8). When solved by using the OSA (3), and under some conditions specified in Theorem 3, the convergence of each inner iteration is guaranteed to be quadratic by Theorem 2. The resulting iterative scheme is referred to as the ONM.
Remark 1.
In many large-scale applications, the coefficient matrix A i ( i = 1 , , m ) is large and sparse with dimensions N × N , while the control matrix B i is tall and skinny, meaning it has significantly fewer columns than rows. Consequently, the gain matrix L i ( k ) computed during the ONM also exhibits a tall and skinny structure.
An advantage of the ONM emerges when the right-hand side matrix Q i in (8) admits low-rank factorization of the form Q i = C i C i , where the number of columns in C i is far less than N. In such a case, the column space of the right-hand side of (8) remains low-dimensional throughout the iteration, and the generated matrix tends to maintain a low-rank structure. This allows for efficient computation of the approximate solution with significantly reduced storage cost.

3.2. Convergence and Convergence Rate

The following theorem concludes the monotone convergence of the iteration sequence { X i ( k ) } generated by ONM (8) under some mild conditions.
Theorem 3.
Let Q i 0 , R i > 0 for all i m . Assume that there exists an SPD matrix X ^ i such that D c i ( X ^ ) 0 and the Euclidean norm of A i is less than one for each i. Then the sequence { X i ( k ) } k = 1 generated by (8) satisfies
1.
X i ( k ) X i ( k + 1 ) X ^ i , k 1 ;
2.
σ ( A ^ i ( k ) ) D < , k 1 .
So there is a solution X i * 0 to D c i ( X ) = 0 such that σ ( A ^ i * ) D for each i, where
A ^ i * = I B i ( R i + B i E i ( X * ) B i ) 1 B i E i ( X * ) A i .
Proof. 
We prove items 1–2 by induction. Let X i ( 1 ) be the solution of
X i ( 1 ) ( A i ) E i ( X ( 1 ) ) A i = Q i
for i = 1 , , m . As A i is d-stable and Q i 0 , it follows from Lemma 1 that X i ( 1 ) 0 and E i ( X ( 1 ) ) 0 . Moreover, direct calculation shows that
( X i ( 1 ) X ^ i ) A i E i ( X ( 1 ) ) E i ( X ^ ) A i = D c i ( X ^ ) + E ^ i R ^ i 1 E ^ i ,
where
R ^ i = R i + B i E i ( X ^ ) B i , E ^ i = B i E i ( X ^ ) A i .
Using the assumption R i > 0 and Lemma 1 again, one has X i ( 1 ) X ^ i for i = 1 , , m .
The SMW formula (Lemma 3) directly yields
A ^ i ( 1 ) = A i B i ( L i ( 1 ) ) = I B i ( R i + B i E i ( X ( 1 ) ) B i ) 1 B i E i ( X ( 1 ) ) A i = I + B i R i 1 B i E i ( X ( 1 ) ) 1 A i
with
( L i ( 1 ) ) = ( R i + B i E i ( X ( 1 ) ) B i ) 1 B i E i ( X ( 1 ) ) A i : = ( R ^ i ( 1 ) ) 1 B i E i ( X ( 1 ) ) A i .
It then follows Lemma 2 that
ρ ( A ^ i ( 1 ) ) ( I + B i R i 1 B i E i ( X ( 1 ) ) ) 1 2 A i 2 < 1
for i = 1 , , m .
Now with the available X i ( 1 ) , A ^ i ( 1 ) , and L i ( 1 ) , we construct sequences { X i ( l ) } l = 1 k , { A ^ i ( l ) } l = 1 k , and { L i ( l ) } l = 1 k satisfying the induction assumptions
X i ( 1 ) X i ( k ) X ^ i ,
and
σ ( A ^ i ( l ) ) = σ A i B i ( L i ( l ) ) D < ,
for i m and l k . Here,
( L i ( l ) ) = ( R i + B i E i ( X ( l ) ) B i ) 1 B i E i ( X ( l ) ) A i : = ( R ^ i ( l ) ) 1 B i E i ( X ( l ) ) A i .
We will show they are true for k + 1 .
By using Newton’s iteration, one has
( X i ( k ) X i ( k + 1 ) ) ( A ^ i ( k ) ) ( E i ( X ( k ) ) E i ( X ( k + 1 ) ) ) A ^ i ( k ) = ( X i ( k ) ( A ^ i ( k ) ) E i ( X ( k ) ) A ^ i ( k ) ) ( X i ( k + 1 ) ( A ^ i ( k ) ) E i ( X ( k + 1 ) ) A ^ i ( k ) ) = ( X i ( k ) ( A ^ i ( k 1 ) ) E i ( X ( k ) ) A ^ i ( k 1 ) ) + ( A ^ i ( k 1 ) ) E i ( X ( k ) ) A ^ i ( k 1 ) ( A ^ i ( k ) ) E i ( X ( k ) ) A ^ i ( k ) L i ( k ) R i ( L i ( k ) ) Q i = L i ( k 1 ) R i ( L i ( k 1 ) ) L i ( k ) R i ( L i ( k ) ) + ( A ^ i ( k 1 ) ) E i ( X ( k ) ) A ^ i ( k 1 ) ( A ^ i ( k ) ) E i ( X ( k ) ) A ^ i ( k ) .
By replacing A ^ i ( k 1 ) and A ^ i ( k ) with A i B i ( L i ( k 1 ) ) and A i B i ( L i ( k ) ) , respectively, and noting B i E i ( X ( k ) ) A i = R ^ i ( k ) ( L i ( k ) ) , the above expression becomes
L i ( k 1 ) R ^ i ( k ) ( L i ( k 1 ) ) L i ( k ) R ^ i ( k ) ( L i ( k ) ) ( L i ( k 1 ) L i ( k ) ) B i E i ( X ( k ) ) A i A i E i ( X ( k ) ) B i ( L i ( k 1 ) L i ( k ) ) = L i ( k 1 ) R ^ i ( k ) ( L i ( k 1 ) ) L i ( k ) R ^ i ( k ) ( L i ( k ) ) ( L i ( k 1 ) L i ( k ) ) R ^ i ( k ) ( L i ( k ) ) L i ( k ) R ^ i ( k ) ( L i ( k 1 ) L i ( k ) ) = ( L i ( k 1 ) L i ( k ) ) R ^ i ( k ) ( L i ( k 1 ) L i ( k ) ) .
Since A ^ i ( k ) is d-stable and ( L i ( k 1 ) L i ( k ) ) R ^ i ( k ) ( L i ( k 1 ) L i ( k ) ) 0 (as R i > 0 ), it follows from Lemma 1 that X i ( k ) X i ( k + 1 ) for i = 1 , , m .
We next show that X i ( k + 1 ) X ^ i holds for all i m . In fact, one has
X ^ i ( A ^ i ( k + 1 ) ) E i ( X ^ ) A ^ i ( k + 1 ) = X ^ i ( A i ) E i ( X ^ ) A i + L i ( k + 1 ) E ^ i + ( L i ( k + 1 ) E ^ i ) L i ( k + 1 ) ( R ^ i R i ) ( L i ( k + 1 ) ) = D c i ( X ^ ) + Q i ( R ^ i L i ( k + 1 ) E ^ i ) R ^ i 1 ( R ^ i L i ( k + 1 ) E ^ i ) + L i ( k + 1 ) R i ( L i ( k + 1 ) ) .
On the other hand, one can use almost the same assertion with (10) that ρ ( A ^ i k + 1 ) < 1 for all i m . Now Newton’s iteration indicates that
X i ( k + 1 ) ( A ^ i ( k + 1 ) ) E i ( X ( k + 1 ) ) A ^ i ( k + 1 ) = X i ( k + 1 ) A ^ i ( k ) + B i ( L i ( k ) L i ( k + 1 ) ) E i ( X ( k + 1 ) ) A ^ i ( k ) + B i ( L i ( k ) L i ( k + 1 ) ) = X i ( k + 1 ) ( A ^ i ( k ) ) E i ( X ( k + 1 ) ) A ^ i ( k ) B i ( L i ( k ) L i ( k + 1 ) ) E i ( X ( k + 1 ) ) A ^ i ( k ) ( A ^ i ( k ) ) E i ( X ( k + 1 ) ) B i ( L i ( k ) L i ( k + 1 ) ) + B i ( L i ( k ) L i ( k + 1 ) ) E i ( X ( k + 1 ) ) B i ( L i ( k ) L i ( k + 1 ) ) = L i ( k ) R i ( L i ( k ) ) + Q i B i ( L i ( k ) L i ( k + 1 ) ) E i ( X ( k + 1 ) ) ( A i + B i ( L i ( k ) ) ) A i + B i ( L i ( k ) ) E i ( X ( k + 1 ) ) B i ( L i ( k ) L i ( k + 1 ) ) + B i ( L i ( k ) L i ( k + 1 ) ) E i ( X ( k + 1 ) ) B i ( L i ( k ) L i ( k + 1 ) ) .
Subtracting (13) from (14) and recollecting some terms with the equality B i E i ( X ( k ) ) A i = R ^ i ( k ) ( L i ( k ) ) , one has
X i ( k + 1 ) X ^ i ( A ^ i ( k + 1 ) ) E i ( X ( k + 1 ) ) E i ( X ^ ) A ^ i ( k + 1 ) = D c i ( X ^ ) + ( R ^ i ( L i ( k + 1 ) ) E ^ i ) R ^ i 1 ( R ^ i ( L i ( k + 1 ) ) E ^ i ) + ( L i ( k + 1 ) L i ( k ) ) ( R i + B i E i ( X ( k + 1 ) ) B i ) ( L i ( k + 1 ) L i ( k ) ) .
Since D c i ( X ^ ) 0 and X ( k + 1 ) 0 (follows from Newton’s iteration Format (8) and Lemma 1), one has
X i ( k + 1 ) X ^ i
for all i m by using Lemma 1.
We now have obtained a non-increasing sequence { X i ( k ) } of SPSD matrices bounded below by X ^ i . Then
X i * : = lim k X i ( k )
exists and is an SPSD matrix satisfying X i * X ^ i . Moreover, R i + B i X i * B i R i + B i X ^ i B i > 0 . By taking the limit in (8) while k , and, for brevity, writing
L i * = A i E i ( X * ) B i R i + B i E i ( X * ) B i 1 ,
one has
X i * A i B i ( L i * ) E i ( X * ) A i B i ( L i * ) = L i * R i ( L i * ) + Q i .
By noting
L i * R i + B i E i ( X * ) B i ( L i * ) = L i * B i E i ( X * ) A i = A i E i ( X * ) B i ( L i * ) ,
(16) is equivalent to D c i ( X * ) = 0 . At last, since A ^ i ( k ) is d-stable for all i m , the limit
σ ( A ^ i * ) = σ A i B i ( L * ) D .
  □
Corollary 1.
Given the assumptions in Theorem 3, the sequence { X i ( k ) } k = 1 generated by (8) also satisfies D c i ( X ( k ) ) 0 for all i m .
Proof. 
We have shown that at the k-th step (see (12)),
( X i ( k ) X i ( k + 1 ) ) ( A ^ i ( k ) ) ( E i ( X ( k ) ) E i ( X ( k + 1 ) ) ) A ^ i ( k ) 0 .
By using Newton’s iteration, it is equivalent to
X i ( k ) A i B i ( L i ( k ) ) E i ( X ( k ) ) A i B i ( L i ( k ) ) L i ( k ) R i ( L i ( k ) ) Q i = X i ( k ) A i E i ( X ( k ) ) A i + 2 A i E i ( X ( k ) ) B i ( R ^ i ( k ) ) ( 1 ) B i E i ( X ( k ) ) A i L i ( k ) R ^ i ( k ) ( L i ( k ) ) Q i = D c i ( X ( k ) ) 0 .
So D c i ( X ( k ) ) 0 for i m .    □
The following theorem describes the convergence rate of the operator Newton method.
Theorem 4.
Given the assumptions in Theorem 3 and that the sequence { X i ( k ) } k = 0 of SPSD matrices generated by (8) converges to the solution X i * for i m . Let
δ X ( k ) : = X i ( k ) X i * a n d δ X ( k ) : = max i m X i ( k ) X i * .
Let L i * be defined by (15). If A ^ i * = A i B i ( L i * ) is d-stable, then there exists a constant r > 0 such that
δ X ( k + 1 ) c δ X ( k ) 2 .
Proof. 
Let
R ^ i * = R i + B i E i ( X * ) B i , R ^ i ( k ) = R i + B i E i ( X ( k ) ) B i
and L i ( k ) , L i * be defined in (11) and (15), respectively. Then
L i ( k ) L i * = A i E i ( X ( k ) ) B i ( R ^ i ( k ) ) 1 A i E i ( X * ) B i ( R ^ i * ) 1 = A i E i ( X ( k ) X * ) B i ( R ^ i ( k ) ) 1 + A i E i ( X * ) B i ( R ^ i * ) 1 B i E i X ( k ) X * B i ( R ^ i ( k ) ) 1 .
Since ρ ( A i ) < 1 and X i ( k ) X i * for all i m and k = 1 , 2 , , one has
( R ^ i ( k ) ) 1 ( R ^ i * ) 1 ,
and therefore for some norm, one has
L i ( k ) L i * c 1 j = 1 m p i j ( X j ( k ) X j * ) c 1 δ X ( k ) ,
where the constant
c 1 = ( R ^ i * ) 1 · B i + ( R ^ i * ) 1 2 · B i 2 · B i E i ( X * ) A i .
By using Newton’s iteration, it follows that
( X i ( k + 1 ) X i * ) ( A ^ i ( k ) ) ( E i ( X ( k + 1 ) ) E i ( X * ) ) A ^ i ( k ) = ( X i ( k + 1 ) ( A ^ i ( k ) ) E i ( X ( k + 1 ) ) A ^ i ( k ) ) ( X i * + ( A ^ i * ) E i ( X * ) A ^ i * ) + ( A ^ i ( k ) ) E i ( X * ) A ^ i ( k ) ( A ^ i * ) E i ( X * ) A ^ i * = L i ( k ) R i ( L i ( k ) ) L i * R i ( L i * ) + ( A ^ i ( k ) ) E i ( X * ) A ^ i ( k ) ( A ^ i * ) E i ( X * ) A ^ i * .
Note that A ^ i ( k ) = A i B i ( L i ( k ) ) and A ^ i * = A i B i ( L i * ) . Then (18) is equivalent to the coupled Stein equations
δ X i ( k + 1 ) ( A ^ i ( k ) ) E i ( δ X ( k + 1 ) ) A ^ i ( k ) = W i ( k ) ,
where
W i ( k ) = ( L i ( k ) L i * ) R ^ i * ( L i ( k ) L i * ) .
With the current A ^ i ( k ) and W i ( k ) at the k-th step, Theorem 1 indicates that the solution to the above equation can be written as
δ X i ( k + 1 ) = W i ( k ) + ( A ^ i ( k ) ) E i j = 0 ( A ^ ( k ) ) E j ( W ( k ) ) ( A ^ ( k ) ) j A ^ i ( k ) .
By taking the term-by-term norm on the right-hand side, one then has
δ X i ( k + 1 ) W i ( k ) + A ^ i ( k ) 2 ( j = 1 m p i j W j ( k ) ) + + A ^ i ( k ) 2 j = 1 m p i , i l A ^ i l ( k ) 2 ( j = 1 m p i 1 , j W j ( k ) ) + W ( k ) + ( ρ ( k ) ) 2 W ( k ) + + ( ρ ( k ) ) 2 i l W ( k ) ) + W ( k ) 1 ( ρ ( k ) ) 2 ,
where ρ ( k ) = max i m A ^ i ( k ) and W ( k ) = max i m W i ( k ) . Since A ^ i ( k ) converges to A ^ i * and A ^ i * is d-stable, this implies that ρ ( k ) is bounded. This, together with (17) and (19), indicates that
δ X ( k + 1 ) = max i m δ X i ( k + 1 ) c δ X ( k ) 2 ,
where the constant c does not associate with k.    □

4. Structured ONM for Large-Scale Problems

In many engineering applications, the matrix A i is typically sparse, while the matrix Q i often possesses a low-rank structure. Therefore, in this section, we adapt the ONM (8) to a low-rank framework that is particularly suitable for large-scale problems.

4.1. Structured Iteration Scheme

We first review the low-rank OSA (3) for the coupled discrete-time Stein equations (CDSEs)
X i A i E i ( X ) A i W i = 0 , i = 1 , , m ,
where W i = L i W ( L i W ) .
Lemma 4
([20]). Let X i ( 0 ) = W i = L i W ( L i W ) with L i W R N × l i and l i N . The sequences F i ( k ) ( X ( k ) ) and X i ( k ) generated by (3) are factorized as in the following format
X i ( k ) = L i , k W K i , k W ( L i , k W ) , F i ( k ) ( X ( k ) ) = L i , k F 2 k K i , k F 2 k ( L i , k F 2 k ) , k = 0 , 1 , 2 , ,
where
  L i , k F 2 k = A i [ L 1 , k F 2 k 1 , , L m , k F 2 k 1 ] , K i , k F 2 k = p i , 1 K 1 , k F 2 k 1 p i , m K m , k F 2 k 1 ,     L i , k W = [ L i , k 1 W , L i , k 1 F 2 k 1 ] , K i , k W = K i , k 1 W K i , k 1 F 2 k 1
and
L i , k F 1 = A i [ L 1 , k W , , L m , k W ] , K i , k F 1 = p i , 1 K 1 , k W p i , m K m , k W , K i , 0 W = I , L i , 0 W = L i W .
The above lemma shows that the iteration of OSA applied into CDSEs has a low-ranked format when W i is of low-rank form. For the operator Newton method, each iteration step requires solving low-rank coupled Stein equations of the form
X i ( k + 1 ) ( A ^ i ( k ) ) E i ( X ( k + 1 ) ) A ^ i ( k ) = Q ¯ i ( k ) K ¯ i ( Q ¯ i ( k ) ) , i = 1 , , m ,
where
Q ¯ i ( k ) = [ L i ( k ) , C i ] , K ¯ i = R i I , A ^ i ( k ) = A i B i ( L i ( k ) ) ,
and L i ( k ) is defined in (9).
We will demonstrate that when Q i = C i C i with C i R m c × N ( m c N ) and the current iteration term X i ( k ) in (21) admits a low-rank approximation with accuracy ϵ , then so does the next iteration term X i ( k + 1 ) .
Specifically, if X ( 0 ) = 0 , Equation (21) reduces to
X i ( 1 ) A i E i ( X ( 1 ) ) A i = Q ¯ i ( 0 ) K ¯ i ( Q ¯ i ( 0 ) ) , i = 1 , , m ,
where Q ¯ i ( 0 ) = C i and K ¯ i = I . By Lemma 4, X i ( 1 ) admits an approximate low-rank representation:
X i ( 1 ) Y i , l 1 ( 0 ) = L i , l 1 Q ( 0 ) K i , l 1 Q ( L i , l 1 Q ( 0 ) ) ,
with accuracy ϵ (after l 1 iterations), where
  L i , l 1 Q ( 0 ) = [ L i , l 1 1 Q ( 0 ) , L i , l 1 1 F 2 l 1 1 ( 0 ) ] , K i , l 1 Q = K i , l 1 1 Q K i , l 1 1 F 2 l 1 1 ,   L i , l 1 1 F 2 l 1 1 ( 0 ) = A i [ L 1 , l 1 1 F 2 l 1 1 1 ( 0 ) , , L m , l 1 1 F 2 l 1 1 1 ( 0 ) ] , , L i , l 1 1 F 1 ( 0 ) = A i [ L 1 , l 1 1 Q ( 0 ) , , L m , l 1 1 Q ( 0 ) ] ,   K i , l 1 1 F 2 l 1 1 = p i , 1 K 1 , l 1 1 F 2 l 1 1 1 p i , m K m , l 1 1 F 2 l 1 1 1 , , K i , l 1 1 F 1 = p i , 1 K 1 , l 1 1 Q p i , m K m , l 1 1 Q ,       L i , 0 F 1 ( 0 ) = A i [ C 1 , , C m ] , K i , 0 F = p i , 1 I p i , m I .
Assume that the k-th iteration term X i ( k ) is approximated by
X i ( k ) Y i , l k ( k 1 ) = L i , l k Q ( k 1 ) K i , l k Q ( L i , l k Q ( k 1 ) ) ,
with tolerance ϵ (after l k iterations). Then,
E i ( X ( k ) ) E i ( Y · , l k ( k 1 ) ) = p i , 1 Y 1 , l k ( k 1 ) p i , m Y m , l k ( k 1 ) : = L l k Q ( k 1 ) K l k Q ( L l k Q ( k 1 ) ) ,
where L l k Q ( k 1 ) = [ L 1 , l k Q ( k 1 ) , , L m , l k Q ( k 1 ) ] and K l k Q = p i , 1 K 1 , l k Q p i , m K m , l k Q .
Then the matrices L i ( k ) and A ^ i ( k ) in (9), as well as Q ¯ i ( k ) in (21), are approximated by
L i ( k ) a = A i L l k Q ( k 1 ) K l k Q ( L l k Q ( k 1 ) ) B i R i + B i L l k Q ( k 1 ) K l k Q ( L l k Q ( k 1 ) ) B i 1 ,
A ^ i ( k ) a = A i B i ( L i ( k ) a ) , Q ¯ i ( k ) a = [ L i ( k ) a , C i ] .
We now solve the approximated low-rank coupled Stein equations
X i A i B i ( L i ( k ) a ) E i ( X ) A i B i ( L i ( k ) a ) = Q ¯ i ( k ) a K ¯ i ( Q ¯ i ( k ) a ) , i = 1 , , m
with K ¯ i = R i I . Applying Lemma 4 again, we approximate X i ( k + 1 ) as
Y i , l k + 1 ( k ) = L i , l k + 1 Q ( k ) K i , l k + 1 Q ( L i , l k + 1 Q ( k ) ) ,
with tolerance ϵ (after l k + 1 iterations), where
  L i , l k + 1 Q ( k ) = [ L i , l k + 1 1 Q ( k ) , L i , l k + 1 1 F 2 l k + 1 1 ( k ) ] , K i , l k + 1 Q = K i , l k + 1 1 Q K i , l k + 1 1 F 2 l k + 1 1 ,   L i , l k + 1 1 F 2 l k + 1 1 ( k ) = ( A ^ i ( k ) a ) [ L 1 , l k + 1 1 F 2 l k + 1 1 1 ( k ) , , L m , l k + 1 1 F 2 l k + 1 1 1 ( k ) ] , ,   L i , l k + 1 1 F 1 ( k ) = ( A ^ i ( k ) a ) [ L 1 , l k + 1 1 Q ( k ) , , L m , l k + 1 1 Q ( k ) ] ,     ,   L i , 0 F 1 ( k ) = ( A ^ i ( k ) a ) [ Q ¯ 1 ( k ) a , , Q ¯ m ( k ) a ] ,
  K i , l k + 1 1 F 2 l k + 1 1 = p i , 1 K 1 , l k + 1 1 F 2 l k + 1 1 1 p i , m K m , l k + 1 1 F 2 l k + 1 1 1 ,   ,   K i , l k + 1 1 F 1 = p i , 1 K 1 , l k + 1 1 Q p i , m K m , l k + 1 1 Q ,     K i , 0 F 1 = p i , 1 K ¯ 1 p i , m K ¯ m .
We can then construct
L i ( k + 1 ) a = A i L l k + 1 Q ( k ) K l k + 1 Q ( L l k + 1 Q ( k ) ) B i R i + B i L l k + 1 Q ( k ) K l k + 1 Q ( L l k + 1 Q ( k ) ) B i 1 ,
A ^ i ( k + 1 ) a = A i B i ( L i ( k + 1 ) a ) , Q ¯ i ( k + 1 ) a = [ L i ( k + 1 ) a , C i ] ,
and resume the iteration (23) successively.
Note that the size of the kernel matrix K ¯ i = R i I remains invariant throughout the iterations. Consequently, when the number of columns in R i and C i is small relative to the dimension N, a distinctive feature of the large-scale ONM is that the rank of the approximated solution resets to a small, fixed number after each iteration before growing again. Given the quadratic convergence of ONM, this property—combined with the truncation and compression discussed later—this enables effective control over rank growth and prevents excessive increase in the solution’s column dimension during the iterative process.

4.2. Computation of the Residual

Assume that the subproblem (21) in the ONM is solved after l k + 1 inner iterations, yielding an approximated solution Y i , l k + 1 ( k ) in (22) to the coupled Stein equations with error ϵ . Below, we provide a detailed description of the residual computation at the current ONM iteration.
Substitute (22) into (1). The residual of the large-scale equation can be expressed as follows:
R d ( Y i , l k + 1 ( k ) ) = Y i , l k + 1 ( k ) + A i E i ( Y · , l k + 1 ( k ) ) A i + C i C i A i E i ( Y · , l k + 1 ( k ) ) B i R + B i E i ( Y · , l k + 1 ( k ) ) B i 1 B i E i ( Y · , l k + 1 ( k ) ) A i = L i R ( k + 1 ) K i R ( k + 1 ) ( L i R ( k + 1 ) ) ,
where
L i R ( k + 1 ) = [ L i , l k + 1 Q ( k ) , A i L 1 , l k + 1 Q ( k ) , , A i L m , l k + 1 Q ( k ) , C i ] , K i R ( k + 1 ) = K i , l k + 1 Q K ¯ i , l k + 1 Q I
and
K ¯ i , l k + 1 Q = p i 1 K 1 , l k + 1 Q             p i m K m , l k + 1 Q p i 1 K 1 , l k + 1 Q             p i m K m , l k + 1 Q     ( L 1 , l k + 1 Q ( k ) ) B 1 ( L m , l k + 1 Q ( k ) ) B m R i + [ B 1 L 1 , l k + 1 Q ( k ) , , B m L m , l k + 1 Q ( k ) ] D i K ¯ Q ( L 1 , l k + 1 Q ( k ) ) B 1 ( L m , l k + 1 Q ( k ) ) B m 1     [ B 1 L 1 , l k + 1 Q ( k ) , , B m L m , l k + 1 Q ( k ) ] p i 1 K 1 , l k + 1 Q             p i m K m , l k + 1 Q .
Clearly, the residual in (29) possesses a low-rank structure. By applying the truncation and compression technique discussed in the next subsection, the computation of the residual can be reduced to a small-scale problem whose size matches that of the kernel matrix  K i R ( k + 1 ) .

4.3. Truncation and Compression

The truncation and compression (TC) technique [24,25] is applied at two critical stages in our algorithm: during the iteration process and in the computation of the residual. Specifically, in each ONM iteration, we solve the subproblem using the operator Smith algorithm. From the iterative structure in (26), it can be observed that the number of columns in L i , l k 1 F 2 l k 1 ( k ) grows approximately as O ( m 2 k 1 l i ) , where l i , a fixed constant independent of k, denotes the initial column count of the factor Q ¯ i ( k ) a . To mitigate the resulting growth in computational and memory cost, we apply the TC technique to reduce the column dimension of L i , l k 1 F 2 l k 1 ( k ) in (26).
Concretely, we impose QR decompositions with the column pivoting on L i , l k 1 F 2 l k 1 ( k ) as
L i , l k 1 F 2 l k 1 ( k ) P i F 2 l k 1 = [ T i F 2 l k 1 Q ˜ i F 2 l k 1 ] U 1 F 2 l k 1 U 2 F 2 l k 1 0 U ˜ F 2 l k 1 ,
such that U ˜ F 2 l k 1 , after l k iterations, satisfies
U ˜ F 2 l k 1 < u 0 f τ ,
where P i F 2 l k 1 is the permutation matrix ensuring that the diagonal elements of the decomposed block triangular matrices decrease in absolute value. Additionally, u 0 f represents a constant, and τ is some small tolerance controlling TC, respectively. Denote m f 2 l k 1 by the column number of L i , l k 1 F 2 l k 1 ( k ) , bounded above by a given m max . Then, it admits that
r f 2 l k 1 : = rank ( L i , l k 1 F 2 l k 1 ( k ) ) m f 2 l k 1 m max ,
with m max N . We then truncate U ˜ F 2 l k 1 in (31) and get the approximated factor as
L i , l k 1 F 2 l k 1 ( k ) t = T i F 2 l k 1 .
The correspondingly compressed kernel is
K i , l k 1 F 2 l k 1 t : = [ U i , 1 F 2 l k 1 U i , 2 F 2 l k 1 ] ( P i F 2 l k 1 ) K i , l k 1 F 2 l k 1 P i F 2 l k 1 [ U i , 1 F 2 l k 1 U i , 2 F 2 l k 1 ] .
In addition, TC is also employed in the evaluation of the residual of (29). Specifically, we implement the QR decomposition with pivoting on L i R ( k ) in (30) as
L i R ( k ) P i R ( k ) = [ T i R ( k ) Q ˜ i R ( k ) ] U i , 1 R ( k ) U i , 2 R ( k ) 0 U ˜ i R ( k ) ,
such that U ˜ i R ( k ) satisfies
U ˜ i R ( k ) < u 0 r τ ,
where P i R ( k ) is a pivoting matrix and u 0 r is some constant. Then the compressed kernel of the residual is
K i R ( k ) t = [ U i , 1 R ( k ) U i , 2 R ( k ) ] ( P i R ( k ) ) K i , k R P i R ( k ) [ U i , 1 R ( k ) U i , 2 R ( k ) ] ,
and the termination condition of the whole algorithm is that
R e l _ R e s = max i K i R ( k ) t K i R ( 0 ) t ϵ
with ϵ being some tolerance.

4.4. Algorithm and Complexity

The low-rank structured ONM (ONM_lr), enhanced with truncation and compression (TC), is outlined in Algorithm 1 and the corresponding flowchart is given in Figure 1.
Algorithm 1: ONM_lr. Solve large-scale CDAREs with sparse A i and low-ranked Q i = C i C i .
Inputs: Sparse matrices A i , low-rank factors B i and C i as well as small matrices R i for i = 1 , , m , probability matrix Π R m × m , truncation tolerance τ , upper bound m max and the iteration tolerance ϵ .
Outputs: Low-ranked matrix L i S and the kernel matrix K i S with the solution X i * L i S K i S ( L i S ) .
1. Set Q ¯ i ( 0 ) = C i and K ¯ i = I for i = 1 , , m .
2. For k = 0 , 1 , , until convergence is reached:
3.  Compute L i , l k + 1 1 F 2 l k + 1 1 ( k ) and K i , l k + 1 1 F 2 l k + 1 1 as in (26) and (26), respectively.
4.  Truncate and compress L i , l k + 1 1 F 2 l k + 1 1 ( k ) as in (31) with accuracy u 0 f τ .
5.  Construct compressed L i , l k + 1 1 F 2 l k + 1 1 ( k ) and K i , l k + 1 1 F 2 l k + 1 1 as in (32) and (33), respectively.
6.  Construct L i , l k + 1 Q ( k ) and K i , l k + 1 Q as in (25).
7.  Compute residual matrices L i R ( k ) and K i R as in (30).
8.  Truncate and compress L i R ( k ) as in (34) with accuracy u 0 r τ .
9.  Construct compressed residual matrix K i R ( k ) t as in (35).
10.  Evaluate the relative residual Rel_Res in (36).
11.  If Rel_Res < ϵ , break. End.
12.  Construct L i ( k + 1 ) a and Q ¯ i ( k + 1 ) a as in (27) and (28), respectively. Set K ¯ i = R i I .
13.   k : = k + 1 .
14. End (For)
15. Output K i S : = K i , l k + 1 Q , L i S : = L i , l k + 1 Q ( k ) .
Remark 2.
Note that in lines 3–6, we solve the CDSEs (23) via the operator Smith algorithm with a truncation. Assume that the subproblem is solved after k + 1 iterations with the tolerance ϵ, the obtained approximated solution to the subproblem is (24).
We next show the computational complexity of ONM_lr at each iteration. We always assume that the matrix A i ( i = 1 , , m ) is sufficiently sparse. This allows us to consider the cost of both the product A i B i and solving the equation A i X = U , which are both within the range of c N floating-point operations (flops), where U is an N × m u matrix with m u N , and c is a constant. Additionally, for i = 1 , , m , the maximal numbers of the columns of initial matrices B i and C i , truncated factors L i , l j F 2 l j ( k ) and L i , l j Q 2 l j ( k ) , as well as residual matrix L i R ( k + 1 ) , are denoted by m b , m c , m j f ( k ) , m j q ( k ) , and m r ( k + 1 ) , respectively. The flops and memory of the k-th iteration are summarized in Table 1.
Remark 3.
1. We assume that l k + 1 iterations of OSA are required to solve the subproblem (21) with tolerance ϵ, the complexity of each iteration in solving the subproblem is given in [20], and the resulting complexities of L i , l k + 1 1 F 2 l k + 1 1 ( k ) and K i , l k + 1 1 F 2 l k + 1 1 in each ONM iteration are listed in the first two rows of Table 1.
2. The QR decomposition and finding the inverse in (27) to form L i ( k + 1 ) a are implemented by the Householder method and the LU decomposition, respectively [26].

5. Numerical Examples

In this section, we demonstrate the effectiveness of the proposed ONM_lr algorithm for computing the solution of large-scale CDARE (1), through examples drawn from [27,28,29,30]. The implementation of ONM_lr was coded by MATLAB R2019a on a 64-bit Windows 10 desktop, equipped with a 3.0 GHz Intel Core i5 processor (6 cores/6 threads) and 32 GB of RAM. The machine precision was set to eps = 2.22 × 10 16 . Especially, the HODLR structure [21,22] used in the standard Newton’s method was also coded by MATLAB and can be viewed at https://github.com/numpi/hm-toolbox (accessed on 30 July 2025.).
The maximum number of columns in the low-rank factors was restricted to m max = 1000 , and the truncation–compression (TC) tolerance was chosen as τ = 10 16 . The residuals were evaluated as in (36), and the stopping criterion was set to a tolerance of ϵ = 10 13 . We did not compare the proposed method with the recently developed inverse-free fixed-point methods [12,13], as those approaches require the initial matrices Q i to be nonsingular—an assumption that is clearly not satisfied in large-scale engineering problems with low-rank structure.
Example 1.
This example is adapted from a slightly modified all-pass single-input single-output (SISO) system originally studied in [29], generating an all-pass SISO system. In this setting, the controllability and observability Gramians satisfy a quasi-inverse relation, i.e., W c i W o i = σ i I for some σ i > 0 . Consequently, the system exhibits a single Hankel singular value with multiplicity equal to the system order.
The derived system matrices are as follows:
A 1 = 0.4 A ¯ 1 R N × N , A 2 = 0.5 A ¯ 2 R N × N , B 1 = [ 1 , , 0 , 0 ] R N × 1 , B 2 = [ 0 , , 0 , 1 ] R N × 1 , R 1 = 1 , R 2 = 1 , C 1 = [ 1 , 0 , , 0 , 1 ] R 1 × N , C 2 = [ 0 , 1 , 0 , , 0 , 1 , 0 ] R 1 × N ,
where A ¯ 1 and A ¯ 2 are both tri-diagonal matrices tridiag ( 1 , 0 , 1 ) but with A ¯ 1 ( 1 , 1 ) = 0.5 and A ¯ 2 ( 1 , 1 ) = 0.8 , respectively. We consider m = 2 and select the probability matrix Π = 0.244 0.756 0.342 0.658 .
We first compare the performance of the ONM_lr algorithm with the standard Newton’s method incorporating the HODLR structure (SN_HODLR) for CDARE of dimensions 10 , 000 and 20 , 000 . The results are reported in Table 2, where columns It., CPU, and Rel_Res report the iteration number, elapsed CPU time, and the relative residuals of the CDARE, respectively, when the algorithm terminates. For N = 10 , 000 , ONM_lr achieves the prescribed residual level in approximately 6.2 s, while the SN_HODLR requires about 1320 s to reach termination, roughly 212 times longer than ONM_lr. For N = 20 , 000 , ONM_lr reaches the prescribed residual in about 6.7 s, whereas the SN_HODLR is out of memory during iterations and fails to complete the computation.
We then assess the performance of the ONM_lr algorithm on larger CDARE with dimensions N = 50 , 000 , 70 , 000 , 90 , 000 , and 110 , 000 , and summarize the numerical results in Table 3. The quantities δ t k and t k denote the CPU time of the k-th iteration and the cumulative runtime up to iteration k, respectively. The column Rel_Res reports the relative residuals of the CDARE at each iteration, while the NC column indicates the maximum number of columns in L i , l k + 1 Q ( k ) . As shown in Table 3, ONM_lr consistently achieves a residual on the order of 10 13 after 4 iterations. Moreover, the column Rel_Res clearly demonstrates the algorithm’s quadratic convergence rate.
To further illustrate the convergence characteristics of the OSA used to solve each subproblem, we depict its convergence trajectories under various dimensions in Figure 2. In each subplot, the outer iterations 1 through 4 are represented by red, yellow, green, and orange, respectively. Within each color, a gradient from light to dark corresponds to increasing values drawn from the interval ( 0 , 10 ) . The concentric rings indicate logarithmic scales from 10 1 to 10 15 . The convergence behavior of the OSA is marked by black circles, blue pentagrams, purple stars, and brown diamonds across the four outer iterations. From the number of concentric levels traversed by each marker, it is evident that the OSA achieves near-quadratic convergence across all subproblems. This reinforces the effectiveness of the proposed operator Newton framework in delivering rapid and robust convergence from inner iterations to the overall CDARE solution.
Example 2.
Consider a structural model of a vertically mounted stand, representative of machinery control systems. This model corresponds to a segment of a machine tool frame, where a series of guide rails is fixed along one surface to facilitate the motion of a tool slide during operation [28,31]. The geometry has been modeled and meshed using ANSYS, and the spatial discretization employs linear Lagrange elements within the finite element framework implemented in FEniCS.
The resulting system matrices are
A 1 = r 1 A R 16 , 626 × 16 , 626 , A 2 = r 2 A R 16 , 626 × 16 , 626 , B 1 = B 2 = B R 16 , 626 × 1 R 1 = R 2 = 1 ,
where random scalars r 1 , r 2 ( 0 , 1 ) are used for certain parameterizations, and the vector B has 392 nonzero elements, with a maximum entry of at most 0.00251 . Due to the structural similarity between A 1 and A 2 , only the sparsity pattern of A 1 is depicted on the left panel of Figure 3. Vectors C 1 and C 2 R 16 , 626 are mostly zero, with five nonzero entries located at rows (3341, 6743, 8932, 11,324, 16,563) for C 1 and (1046, 2436, 6467, 8423, 12,574) for C 2 , respectively. Full matrix data are available from [27] and at the MOR Wiki repository (https://morwiki.mpi-magdeburg.mpg.de/morwiki/index.php/Vertical_Stand) (accessed on 30 July 2025). For this example, we set m = 2 and define the mode transition probability matrix as Π = 0.564 0.436 0.785 0.215 .
We apply the ONM_lr algorithm to solve the resulting coupled CDARE. We omit the comparison with the standard Newton’s method incorporating the HODLR structure, as it is out of memory capacity during iterations at this problem dimension. Table 4 summarizes the numerical results of ONM_lr. The residuals decrease to O ( 10 15 ) after only three outer iterations. The columns t k and δ t k report the cumulative and per-iteration CPU time, respectively. The NC column shows that the column dimension of L i , l k + 1 Q ( k ) increases by more than twice in the first iteration but grows more slowly in subsequent iterations. The Rel_Res column confirms that ONM_lr retains a quadratic convergence.
To further investigate the convergence of the OSA for the subproblems, we present their convergence histories in the right panel of Figure 3. The red, yellow, and green colors correspond to the first, second, and third subproblem solving, respectively, with increasing intensities representing values in ( 0 , 10 ) . Concentric circles indicate residual magnitudes from 10 1 to 10 18 . The convergence trajectories are plotted using black circles, blue pentagrams, and purple stars. The number of magnitude levels traversed by these markers provides clear evidence of nearly quadratic convergence for the OSA. This validates the robustness and efficiency of the ONM_lr algorithm in conjunction with the OSA strategy.
Example 3.
Consider a semi-discretized heat transfer model arising from the optimal cooling of steel profiles in automated control systems, as studied in [27]. The dimension of the resulting dynamical system depends on the level of refinement applied to the computational mesh. Spatial discretization is performed using linear Lagrange elements via the ALBERTA-1.2 finite element toolbox [30].
We slightly modify the model matrices as follows:
A 1 = r 1 A ¯ 1 R N × N , A 2 = r 2 A ¯ 2 R N × N , B 1 = B 2 R N × 7 , C 1 = C 2 R 7 × N ,
where A ¯ = A / A and A is estimated by ‘normest’ in Matlab. We take r 1 = 0.98 and r 2 = 0.92 for N = 20 , 209 and r 1 = 0.87 and r 2 = 0.97 for N = 79 , 841 . R 1 = R 2 = I 7 . In this experiment, we take C 1 = r 3 C and C 2 = r 4 C with r 3 and r 4 being random numbers in (0,1). Matrices A ¯ R N × N , B R N × 1 , and C R 1 × N can be found at [27], or the MOR Wiki repository (https://morwiki.mpi-magdeburg.mpg.de/morwiki/index.php/Steel_Profile) (accessed on 30 July 2025). The probability matrix is defined as Π = 0.713 0.287 0.584 0.416 .
To assess the performance of the proposed ONM_lr algorithm, we solve Equation (1) for two system sizes: N = 20 , 209 and N = 79 , 841 . Again, we omit the comparison with the standard Newton’s method incorporating the HODLR structure, as it is out of memory capacity during iterations at this problem dimension. The numerical results of ONM_lr are reported in Table 5. In both cases, ONM_lr attains a residual norm on the order of 10 15 within just three outer iterations. The cumulative CPU times are approximately 10.5 s and 19.5 s. The It. column records the number of outer iterations, while the NC column reflects the maximum number of columns in L i , l k + 1 Q ( k ) at each step. The modest growth in NC, remaining below two times across iterations, highlights the efficiency of the truncation–compression (TC) strategy employed within both the outer ONM_lr iterations and the inner OSA iterations. Furthermore, the Rel_Res column confirms the quadratic convergence of ONM_lr.
To further visualize the convergence characteristics of the subproblem solvers, we depict the residual histories for both problem sizes in the right panel of Figure 4. Each subplot uses red, yellow, and green markers to represent the first through third subproblem solving. Within each color, darker shades correspond to higher iteration indices. Concentric rings denote residual levels ranging from 10 1 to 10 16 . The convergence paths of the operator Smith iteration are illustrated using black circles, blue pentagrams, and purple stars. The number of magnitude rings traversed by these markers clearly indicates that each subproblem is solved with nearly quadratic convergence. This further confirms the rapid and robust performance of the ONM_lr algorithm when combined with the OSA subproblem strategy.

6. Conclusions

We have developed an operator Newton method for computing the solutions to a class of coupled discrete-time algebraic Riccati equations (CDARE) arising from jump systems. The proposed framework leverages an inner operator Smith algorithm to solve subproblems, and under appropriate assumptions, guarantees locally quadratic convergence in both the inner and outer iterations. To efficiently address large-scale systems in engineering applications, we further present a low-rank variant, ONM_lr, which incorporates truncation and compression strategies to control memory and computational costs. Compared to the standard Newton’s method incorporating the HODLR structure [21,22] and the recently developed inverse-free fixed-point iterations [12,13], which require invertible constant matrices and therefore fail to accommodate low-rank structures in large-scale scenarios, the proposed ONM_lr algorithm demonstrates its effectiveness in handling large-scale problems through numerical experiments. However, when CDARE approaches the critical case, the presented ONM method tends to exhibit near-linear convergence, leading to a significant increase in computational time for large-scale problems. Overcoming this limitation is an important direction for further research and is currently under investigation. Another future research avenue may be to extend the ONM method to coupled continuous-time Riccati equations and other types of equations arising in large-scale stochastic jump systems.

Author Contributions

Conceptualization, B.Y.; methodology, B.Y.; software, Y.L.; validation, N.D.; and formal analysis, N.D. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the NSF of Hunan Province (2023JJ50164, 2023JJ50165, 2026JJ50180), the foundation of the degree and postgraduate education reform project of the Hunan University of Technology (JGYB23009), and the basic education teaching reform research project of Hunan Province (Y2025452).

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

Abbreviations /NotationsDefinition
D < ( D ) The inner region of the open (closed) unit disk on the complex plane
I m Identity matrix with size m
σ ( A ) Spectral of matrix A
ρ ( A ) Spectral radius of matrix A
· -norm of a matrix
· 2 Euclidean norm of a matrix
A B Block diagonal matrix A 0 0 B
r a n k ( · ) The rank of a matrix
d-stableThe spectral radius of a matrix is less than 1, i.e., ρ ( · ) <1
SPSD (or A 0 )Symmetric and Positive Semi-Definite matrix A
SPD (or A > 0 )Symmetric and Positive Definite matrix A
ONMOperator Newton Method
OSAOperator Smith Algorithm
TCTruncation and compression
ϵ Tolerance for iteration termination
τ Tolerance for truncation and compression
m max Allowable maximum column of a matrix in iterations

References

  1. Abou-Kandil, H.; Freiling, G.; Jank, G. On the solution of discretetime markovian jump linear quadratic control problems. Automatica 1995, 31, 765–768. [Google Scholar] [CrossRef]
  2. Liu, Y.; Wang, Z.; Lin, X. Non-zero sum Nash game for discrete-time infinite Markov jump stochastic systems with applications. Axioms 2023, 12, 882. [Google Scholar] [CrossRef]
  3. Do Costa, O.L.V.; Marques, R.P.; Fragoso, M.D. Discrete-Time Markov Jump Linear Systems; Springer: London, UK, 2005. [Google Scholar]
  4. Zhou, B.; Duan, G.-R.; Li, Z.-Y. Gradient based iterative algorithm for solving coupled matrix equations. Syst. Control. Lett. 2009, 58, 327–333. [Google Scholar] [CrossRef]
  5. Wang, Q.; Lam, J.; Wei, Y.; Chen, T. Iterative solutions of coupled discrete Markovian jump Lyapunov equations. Comput. Math. Appl. 2008, 55, 843–850. [Google Scholar] [CrossRef]
  6. Wu, A.-G.; Duan, G.-R. New iterative algorithms for solving coupled Markovian jump Lyapunov equations. IEEE Trans. Autom. Control 2015, 60, 289–294. [Google Scholar]
  7. Dragan, V.; Morozan, T.; Stoica, A.M. Robust Control of Discrete-time Linear Stochastic Systems; Springer: Berlin/Heidelberg, Germany, 2010. [Google Scholar]
  8. Ivanov, I.G. Accelerated LMI solvers for the maximal solution to a set of discrete-time algebraic Riccati equations. Appl. Math. E-Notes 2012, 12, 228–238. [Google Scholar]
  9. Ivanov, I.G. A method to solve the discrete-time coupled algebraic riccati equations. Appl. Math. Comput. 2008, 206, 34–41. [Google Scholar] [CrossRef]
  10. Haghani, F.K.; Soleymani, F. An improved Schulz-type iterative method for matrix inversion with application. Trans. Inst. Meas. Control 2014, 36, 983–991. [Google Scholar] [CrossRef]
  11. Dai, H.; Bai, Z.Z. On eigenvalue bounds and iteration methods for discrete algebraic riccati equations. J. Comput. Math. 2011, 29, 341–366. [Google Scholar] [CrossRef]
  12. Jiang, K.-W.; Li, Z.; Zhang, Y. An inversion-free iterative algorithm with a scalar tuning parameter for coupled Riccati matrix equations arising in LQ optimal control of Markov jump systems. IEEE Trans. Autom. Control 2024, 70, 1913–1920. [Google Scholar] [CrossRef]
  13. Li, Z.; Zhang, Y.; Wu, A.-G. An inversion-free iterative algorithm for riccati matrix equations in discrete-time markov jump systems. IEEE Trans. Autom. Control 2022, 67, 4754–4761. [Google Scholar] [CrossRef]
  14. Feng, T.-T.; Chu, E.K.W. Newton’s method for coupled continuous-time algebraic Riccati equations. J. Appl. Math. Comput. 2024, 70, 1023–1042. [Google Scholar] [CrossRef]
  15. Salama, A.; Gourishankar, V. A computational algorithm for solving a system of coupled algebraic matrix riccati equations. IEEE Trans. Comput. 1974, C-23, 100–102. [Google Scholar] [CrossRef]
  16. Margenov, S.D.; Vulkov, L.G.; Wasniewski, J. (Eds.) Numerical Analysis and Its Applications. In Proceedings of the 4th International Conference (NAA 2008), Lozenetz, Bulgaria, 16–20 June 2008; Revised Selected Papers; Springer: Berlin/Heidelberg, Germany, 2009; Volume 5434. [Google Scholar]
  17. Wang, L.; Zhu, Y.-L. A new inversion-free iterative algorithm for the discrete algebraic Riccati equation. IMA J. Math. Control Inf. 2024, 41, 149–164. [Google Scholar] [CrossRef]
  18. Jerbi, H.; Alshammari, O.; Aoun, S.B.; Kchaou, M.; Simos, T.E.; Mourtas, S.D.; Katsikis, V.N. Hermitian solutions of the quaternion algebraic Riccati equations through zeroing neural networks with application to quadrotor control. Mathematics 2024, 12, 15. [Google Scholar] [CrossRef]
  19. Escorcia, J.M.; Suazo, E. On blow-up and explicit soliton solutions for coupled variable coefficient nonlinear schrödinger equations. Mathematics 2024, 12, 2694. [Google Scholar] [CrossRef]
  20. Yu, B.; Dong, N.; Hu, B.-Q. Operator Smith algorithm for coupled Stein equations from jump control systems. Axioms 2024, 13, 249. [Google Scholar] [CrossRef]
  21. Massei, S.; Palitta, D.; Robol, L. Solving rank structured Sylvester and Lyapunov equations. SIAM J. Matrix Anal. Appl. 2018, 39, 1564–1590. [Google Scholar] [CrossRef]
  22. Massei, S.; Robol, L.; Kressner, D. hm-toolbox: Matlab software for HODLR and HSS matrices. SIAM J. Sci. Comput. 2020, 42, C43–C68. [Google Scholar] [CrossRef]
  23. Higham, N.J. Accuracy and Stability of Numerical Algorithms, 2nd ed.; SIAM: Philadephia, PA, USA, 2002. [Google Scholar]
  24. Yu, B.; Dong, N.; Tang, Q. Factorized squared Smith method for large-scale Stein equations with high-rank terms. Automatica 2023, 154, 111057. [Google Scholar] [CrossRef]
  25. Yu, B.; Fan, H.-Y.; Chu, E.K.-W. Large-scale algebraic Riccati equations with high-rank constant terms. J. Comput. Appl. Math. 2019, 361, 130–143. [Google Scholar] [CrossRef]
  26. Higham, N.J. Functions of Matrices: Theory and Computation; SIAM: Philadelphia, PA, USA, 2008. [Google Scholar]
  27. Korvink, G.; Rudnyi, B. Oberwolfach Benchmark Collection. In Dimension Reduction of Large-Scale Systems; Benner, P., Sorensen, D.C., Mehrmann, V., Eds.; Lecture Notes in Computational Science and Engineering; Springer: Berlin/Heidelberg, Germany, 2005; p. 45. [Google Scholar]
  28. Lang, N. Numerical Methods for Large-Scale Linear Time-Varying Control Systems and related Differential Matrix Equations; Logos-Verlag: Berlin, Germany, 2018. [Google Scholar]
  29. Ober, R.J. Asymptotically Stable All-Pass Transfer Functions: Canonical Form, Parametrization and Realization. IFAC Proc. Vol. 1987, 20, 181–185. [Google Scholar] [CrossRef]
  30. Schmidt, A.; Siebert, K. Design of Adaptive Finite Element Software—The Finite Element Toolbox ALBERTA; Lecture Notes in Computational Science and Engineering; Springer: Berlin/Heidelberg, Germany, 2005; Volume 42. [Google Scholar]
  31. Chahlaoui, Y.; Van Dooren, P. Benchmark examples for model reduction of linear time-invariant dynamical systems, dimension reduction of large-scale systems. In Dimension Reduction of Large-Scale Systems; Springer: Berlin/Heidelberg, Germany, 2005; Volume 45, pp. 379–392. [Google Scholar]
Figure 1. Flowchart of the low-rank structured ONM equipped with truncation and compression.
Figure 1. Flowchart of the low-rank structured ONM equipped with truncation and compression.
Axioms 14 00601 g001
Figure 2. Residual history of operator Smith iteration in solving each subproblem of Example 1.
Figure 2. Residual history of operator Smith iteration in solving each subproblem of Example 1.
Axioms 14 00601 g002
Figure 3. The discretized matrix A and residual history of OSI in solving each subproblem of Example 2.
Figure 3. The discretized matrix A and residual history of OSI in solving each subproblem of Example 2.
Axioms 14 00601 g003
Figure 4. Residual history of OSA in each subproblem of Example 3.
Figure 4. Residual history of OSA in each subproblem of Example 3.
Axioms 14 00601 g004
Table 1. Complexity and memory at each iteration in Algorithm ONM_lr.
Table 1. Complexity and memory at each iteration in Algorithm ONM_lr.
ItemsFlopsMemory
L i , l k + 1 1 F 2 l k + 1 1 ( k ) j = 1 l k + 1 c m j 1 q ( k ) 2 j 1 ( m 2 j 1 + m ) N m 2 l k + 1 1 m l k + 1 1 q ( k ) N
K i , l k + 1 1 F 2 l k + 1 1 j = 1 l k + 1 m ( m j 1 q ( k ) ) 2 ( 1 + m 2 j ) ( j + 1 ) / 2 ( m 2 l k + 1 1 m l k + 1 1 q ( k ) ) 2
L i , l k + 1 1 F 2 l k + 1 1 ( k ) QR 2 ( m 2 l k + 1 1 m l k + 1 1 q ( k ) ) 2 ( N m 2 l k 1 m l k + 1 1 q ( k ) / 3 ) ( m l k + 1 f ( k ) ) 2
Compressed K i , l k + 1 1 F 2 l k + 1 1 4 m l k + 1 f ( k ) ( m 2 l k + 1 1 m l k + 1 1 q ( k ) ) 2 ( m l k + 1 f ( k ) ) 2
L i , l k + 1 Q ( k ) m l k + 1 q ( k ) N
K i , l k + 1 Q ( m l k + 1 q ( k ) ) 2
L i R ( k + 1 ) c m m l k + 1 q ( k ) N [ ( 1 + m ) m l k + 1 q ( k ) + m c ] N
L i R ( k + 1 ) QR 2 [ ( 1 + m ) m l k + 1 q ( k ) + m c ] 2 N [ ( 1 + m ) m l k + 1 q ( k ) + m c ] / 3 ( m r ( k + 1 ) ) 2
Compressed K i R ( k + 1 ) 4 m r ( k + 1 ) [ ( 1 + m ) m l k + 1 q ( k ) + m c ] 2 ( m r ( k + 1 ) ) 2
L i ( k + 1 ) a c m l k + 1 q ( k ) N + 2 m b m l k + 1 q ( k ) ( 2 N + m l k + 1 q ( k ) ) + 14 ( m b ) 2 m l k + 1 q ( k ) / 3 m l k + 1 q ( k ) N
Q ¯ i ( k + 1 ) a ( m l k + 1 q ( k ) + m c ) N
Table 2. Comparison between ONM_lr and SN_HODLR in Example 1.
Table 2. Comparison between ONM_lr and SN_HODLR in Example 1.
ONM_lr SN_HODLR
N It.CPU TimeRel_ResIt.CPU TimeRel_Res
10,0004 6.28 1.83 × 10 13 3 1 , 319.5 1.77 × 10 14
20,0004 6.71 1.86 × 10 13
Table 3. CPU time and residual in Example 1.
Table 3. CPU time and residual in Example 1.
NIt. δ t k t k Rel_ResNC
50,0001 0.370 0.370 1.34 × 10 1 63
2 1.613 1.983 3.26 × 10 2 254
3 2.367 4.350 4.59 × 10 6 327
4 2.697 7.029 1.86 × 10 13 391
70,0001 0.515 0.515 1.34 × 10 1 63
2 1.732 2.246 3.37 × 10 2 250
3 2.382 4.628 4.61 × 10 6 320
4 3.193 7.882 1.85 × 10 13 394
90,0001 0.648 0.648 1.34 × 10 1 63
2 2.204 2.852 3.42 × 10 2 261
3 4.032 6.884 4.82 × 10 6 323
4 4.721 11.606 1.85 × 10 13 388
110,0001 0.713 0.713 1.34 × 10 1 63
2 2.812 3.525 3.40 × 10 2 262
3 4.141 7.667 4.51 × 10 6 309
4 5.021 12.688 1.89 × 10 13 391
Table 4. CPU time and residual in Example 2.
Table 4. CPU time and residual in Example 2.
NIt. δ t k t k Rel_ResNC
16,6261 0.036 0.036 1.74 × 10 4 31
2 0.923 0.959 9.23 × 10 9 68
3 1.731 2.691 2.58 × 10 15 122
Table 5. CPU time and residual in Example 3.
Table 5. CPU time and residual in Example 3.
NIt. δ t k t k Rel_ResNC
20,2091 0.258 0.258 5.65 × 10 4 48
2 2.843 3.101 1.53 × 10 9 88
3 7.472 10.574 1.69 × 10 15 103
79,8411 0.458 0.458 8.72 × 10 4 48
2 5.098 5.556 2.34 × 10 8 88
3 13.912 19.468 4.25 × 10 15 101
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, B.; Liu, Y.; Dong, N. Operator Newton Method for Large-Scale Coupled Riccati Equations Arising from Jump Systems. Axioms 2025, 14, 601. https://doi.org/10.3390/axioms14080601

AMA Style

Yu B, Liu Y, Dong N. Operator Newton Method for Large-Scale Coupled Riccati Equations Arising from Jump Systems. Axioms. 2025; 14(8):601. https://doi.org/10.3390/axioms14080601

Chicago/Turabian Style

Yu, Bo, Yiwen Liu, and Ning Dong. 2025. "Operator Newton Method for Large-Scale Coupled Riccati Equations Arising from Jump Systems" Axioms 14, no. 8: 601. https://doi.org/10.3390/axioms14080601

APA Style

Yu, B., Liu, Y., & Dong, N. (2025). Operator Newton Method for Large-Scale Coupled Riccati Equations Arising from Jump Systems. Axioms, 14(8), 601. https://doi.org/10.3390/axioms14080601

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop