Abstract
In this paper, we investigate -optimal model reduction methods for discrete-time linear time-invariant systems. Similar to the continuous-time case, we will formulate this problem as an optimization problem over a Grassmann manifold. We consider constructing reduced systems by both one-sided and two-sided projections. For one-sided projection, by utilizing the principle of the Grassmann manifold, we propose a gradient flow method and a sequentially quadratic approximation approach to solve the optimization problem. For two-sided projection, we apply the strategies of alternating direction iteration and sequentially quadratic approximation to the minimization problem and develop a numerically efficient method. One main advantage of these methods, based on the formulation of optimization over a Grassmann manifold, is that stability can be preserved in the reduced system. Several numerical examples are provided to illustrate the effectiveness of the methods proposed in this paper.
MSC:
65F99; 93C05
1. Introduction
Model reduction, which is often employed in approximating various very large-scale, complex dynamical systems, has received considerable attention in the past several decades. These systems often arise in various applications, such as fluid dynamics [1], structural dynamics [2], electrical circuits [3], aerodynamics [4], micro-electro-mechanical systems [5], weather prediction [6], and so on. Roughly speaking, the goal of model reduction is to replace a given mathematical model, described via differential equations or difference equations, with a model of the same form but with much smaller state space dimension than the original one, such that the reduced model still describes at least appropriately certain aspects of the original system.
In this paper we consider the model reduction of a discrete-time linear time-invariant (LTI) system described by the following system of difference equations
where the matrices , and represent the state matrix, the input matrix, and the output matrix, respectively. The vectors , , and are the state vector, the input vector, and the output vector of the discrete-time LTI system at time , respectively. For simplicity, we assume that we have a zero initial condition in (1). Discrete-time systems arise naturally when continuous-time systems are discretized by numerical approximation to differentiation. Moreover, they directly appear in some important applications such as economics [7] and transportation network [8]. For a comprehensive treatment of discrete systems, including system characterization, structural properties, stability, optimal control, and applications, we refer to the book [9].
In this paper we aim at constructing a reduced-order model of via projection. Specifically, we hope to find two matrices and with , such that a reduced system is given by
with , , and .
Much research has addressed the construction of the reduced-order model. For continuous-time LTI systems, there exist various model reduction methods. For example, classical model reduction methods, which are based on system Gramians, have been well established, including the balanced truncation method [10], the optimal Hankel norm approximation method [11], and the singular perturbation approximation method [12]. The other class of projection methods are the Krylov subspace [13] and rational Krylov subspace [14] methods, which become increasingly popular for large-scale linear dynamical systems such as those arising from structure dynamics, circuit simulations, and micro-electro-mechanical systems (see [15,16,17,18,19]). Rational Krylov subspace methods have been further improved to develop an iterative rational Krylov subspace algorithm [20], in which the interpolation points are iteratively updated so that the established reduced-order model satisfies the interpolation-based first-order necessary conditions for -optimal model reduction of single-input single-output continuous-time systems. In [21,22], an iterative rational Krylov subspace method (IRKA) is extended to derive a tangential rational interpolation framework for model reduction of multi-input multi-output (MIMO) systems. -optimal model reduction can also be formulated as an optimization problem over a Grassmann manifold. Based on this fact and by making use of one-sided projection, Xu and Zeng [23] proposed a fast gradient flow algorithm and a sequentially quadratic approximation for solving the optimization problem to construct stable reduced systems. The model reduction method based on Grassmann manifold optimization is further improved in [24] by employing two-sided projection. Model order reduction on Grassmann manifolds for other classes of systems are also considered. In [25], model order reduction on Grassmann manifolds has been extended to a special class of linear parameter-varying systems. Xu et al. [26] proposed a preserving–periodic Riemannian descent model reduction iterative method for linear discrete-time periodic systems. In [27], a parametric interpolation parallel MOR method for discrete-time parametric systems is proposed by making use of Grassmann manifolds and discrete Laguerre polynomials. In [28], order reduction based on Grassmann manifolds is extended to bilinear systems. Otto et al. [29] studied model order reduction over the product of two Grassmann manifolds for nonlinear systems. Padovan et al. [30] considered data-driven model reduction for nonlinear systems by solving an optimization problem over the product of two manifolds. In [31], a novel differential geometric framework for model reduction on smooth manifolds is proposed. This general framework can capture and generalize several existing MOR techniques, such as preserving the structures for Lagrangian or Hamiltonian dynamics. Zimmermann [32] reviewed matrix manifolds and outlined the principal approaches to data interpolation and Taylor-like extrapolation on matrix manifolds. For the reduced-order modeling of high-dimensional dynamical systems, Sashittal and Bodony [33] proposed low-rank dynamic mode decomposition by solving a matrix manifold optimization problem with a rank constraint on the solution.
Some researchers have considered -optimal model reduction for discrete-time LTI systems. For example, Van Dooren, Gallivan, and Absil [22] derived two first-order necessary conditions, the discrete-time counterparts of Wilson’s conditions [34] and tangential rational interpolation conditions for -optimal model reduction. In [35], the authors gave the discrete-time counterparts of Hyland–Bernstein Gramian-based first-order necessary -optimality conditions. They showed that Wilson’s conditions, the Hyland–Bernstein conditions, and the tangential interpolation-based conditions are equivalent. Moreover, based on the tangential interpolation-based conditions, a MIMO iterative rational interpolation algorithm is developed for model reduction of discrete-time MIMO LTI systems.
In this paper, we consider constructing reduced systems by both one-sided and two-sided projection. Similar to the continuous-time case, we formulate the -optimal model reduction problems for discrete-time LTI systems as optimization problems over Grassmann manifolds. Based on this formulation, in the one-side projection case, we apply the gradient flow method and the sequentially quadratic approximation method to solve the related optimization problem. For two-sided projection, we combine the techniques of alternating direction iteration and sequentially quadratic approximation to develop an iterative method for the Grassmann manifold optimization problem. Numerical experiments demonstrate the effectiveness of model reduction methods based on Grassmann manifold optimization for discrete-time LTI systems.
Throughout this paper the following notation is used: The sets of all real and complex matrices are denoted by and , respectively. The identity matrix of dimension n is denoted by and the zero matrix by 0. If the dimension of is apparent from the context, we drop the index and simply use I. The actual dimension of 0 will always be apparent from the context. The superscripts T and H denote the transpose and the complex conjugate transpose of a vector or a matrix. The notation denotes the space spanned by the column vectors of the matrix V, and denotes the space spanned by the matrix sequence . For two matrices , the inner product of U and V is defined as , where is the trace of a matrix. The notation denotes the Frobenius matrix norm, which is defined by .
The main contributions of this paper include the following:
- The -optimal model reduction methods based on Grassmann manifold optimization are extended to discrete-time LTI systems.
- For one-sided projection, a gradient flow method and a sequentially quadratic approximation approach are proposed to solve the optimization problem. For two-sided projection, the optimization problem is solved by applying the strategies of alternating direction iteration and sequentially quadratic approximation.
- We present the details of implementation, such as how to efficiently solve sparse–dense discrete-time Sylvester equations.
- The effectiveness of the proposed methods in this paper is demonstrated with two numerical examples.
The remainder of this paper is organized as follows: Section 2 is devoted to a short review of some important results on the -norm of discrete-time LTI systems and Stiefel and Grassmann manifolds. Moreover, we also outline the framework of the gradient flow method for solving optimization problems over Grassmann manifolds. In Section 3, we consider the one-sided projection case. Two iterative methods, one relying on the gradient flow and the other on sequentially quadratic approximation, are proposed. We consider two-sided projection in Section 4. An iterative method based on the combination of alternating direction iteration and sequentially quadratic approximation is given. In Section 5, some issues related to the implementation of these algorithms are investigated, including the initial projection matrix selection and the termination criterion. We also discuss how to solve Stein equations, which appear in the methods proposed in this paper. In Section 6, several numerical examples are presented to illustrate the effectiveness of these model reduction methods. Finally, some conclusions are drawn in Section 7.
2. Preliminaries
In this section we review some important results on the -norm of discrete-time LTI systems and the Grassmann manifold. Moreover, we also outline the framework of the gradient flow method for solving optimization problems over Grassmann manifolds. Much of the material in this section is standard and can be found in [35,36,37,38].
2.1. -Norm of Discrete-Time LTI Systems
For a vector sequence , , its z-transform is defined by
By applying the z-transform to the system , we obtain, in the frequency domain, the following input–output relation:
where and are the z-transformations of the input and output, respectively, and
is called the transfer function of the system .
In this paper, we assume that the discrete-time system is asymptotically stable, i.e., all the eigenvalues of A lie inside the unit circle. In this case, the controllability and observability Gramians of the discrete-time system are well defined and formulated as
These Gramians also satisfy the Stein equations, also called the discrete Lyapunov equations:
For an asymptotically stable discrete-time LTI system , its squared -norm is defined by
Using these systems’ Gramians P and Q, the -norm can be calculated via the following equation:
2.2. Stiefel and Grassmann Manifolds
In this subsection, we briefly review the definitions of Stiefel manifold and Grassmann manifold and some basic results on these two manifolds.
Definition 1.
The real Stiefel manifold with is defined as the set of all orthonormal matrices, i.e.,
Definition 2.
Two matrices in the real Stiefel manifold are defined to be equivalent if their columns span the same r-dimensional subspaces.
From Definition 2, are equivalent if and only if there exists an orthogonal matrix such that . So, we can define the equivalent class of a point to be
where denotes the set of all orthogonal matrices in .
Definition 3.
The real Grassmann manifold with is defined as the set of all r-dimensional real linear subspaces of .
From the above definition, we know that a point in the real Grassmann manifold is an r-dimensional real linear subspace, which can be spanned by the columns of a matrix in . So, there is a one-to-one correspondence between a point in the real Grassmann manifold and an equivalent class of . Thus, the Grassmann manifold can be seen as a quotient manifold, i.e.,
Definition 4.
For a Stiefel manifold , the tangent space at is defined as
Correspondingly, we have a definition of the tangent space of the Grassmann manifold.
Definition 5.
The tangent space at point in the Grassmann manifold is described as
Let be a function of . Then we can define the matrix of all partial derivatives of J with respect to U:
Definition 6.
Let be a function defined in a Stiefel manifold . Then, its gradient at U in the Stiefel manifold is defined to be the tangent vector such that
holds for all tangent vectors .
Definition 7.
Let be a function defined in a Grassmann manifold . Then, its gradient at point is defined to be the tangent vector such that
holds for all tangent vectors .
It has been shown in [36,37] that the gradient of J defined in a Stiefel manifold is expressed as
while the gradient of J defined in Grassmann manifold is formulated as
2.3. The Gradient Flow Method
We briefly overview the framework of this method in the rest of this section.
Given an optimization problem based on Grassmann manifold ,
where J is the objective function.
It is well known [36] that the gradient flow method is one of the most efficient iterative methods for solving optimization problems based on Grassmann manifolds.
Let denote the matrix of all partial derivatives of J with respect to . Then the gradient of J at point U of Grassmann manifold is . Clearly, the solution U of minimization problem (13) must be a point of such that the gradient , i.e., U, satisfies
To find a zero point of the above equation, it is proposed to solve the following gradient flow problem on the Grassmann manifold :
Given an arbitrary square matrix A, the matrix functions and are defined, respectively, as [39]
Note that for any diagonal matrix , it holds that
Given a matrix , let be the singular value decomposition of F, where and . The geodesic on the Grassmann manifold at point with the direction F is defined by
With the help of the geodesic and by choosing an appropriate time step , the gradient flow iteration scheme, which iteratively constructs the approximate solution of (14) along the geodesic at the negative gradient direction , is formulated as
where is the singular value decomposition of the negative gradient of J at .
3. One-Sided Projection via Optimization on Grassmann Manifold
This section is concerned with model order reduction of the discrete-time linear LTI system in (1). One objective of model order reduction is to reduce the error between the original system (1) and the reduced system (2). In this paper we will be measuring the error in the -norm. For the sake of the derivation of necessary conditions for -optimal model reduction over a Grassmann manifold, let us define the transfer function of the error system :
It is clear that the error system has a realization with
i.e., the transfer function of the error system can be also formulated as
Following the knowledge in Section 2, the squared -norm of the error system can be written in terms of its controllability Gramian and observability as
Note that and are also the solutions of the following two Stein equations associated with the error system:
Let us partition the matrices , into
with , , and . Then, the Stein Equation in (20) is equivalent to the following equations:
On the other hand, the Stein Equation in (21) is equivalent to
Based on the expression of the squared -norm of the error system in (19) and the partition of and in (22), the squared -norm of the error system can now be rewritten as
-norm model reduction on a Grassmann manifold for continuous-time systems was considered in [23,40]. In this section, we aim to extend the one-sided projection technique based on Grassmann manifold optimization to the discrete-time case. The one-sided projection-based reduction approach involves finding one projection matrix , . With U, the reduced system is constructed by setting
In the one-sided projection scheme, the objective function is defined by
The objective function is also reformulated as
where Q, , and Y are the solutions to the equations
respectively.
We will be deriving an expression of the matrix of all partial derivatives of J with respect to U. The following lemma is useful in the derivation of our result.
Lemma 1.
Let , , , and . If satisfy
then the following holds:
Proof.
Obviously, it holds that
This completes the proof. □
Theorem 1.
Let , X, , and Y be the solutions of (31), (32), (34), and (35), respectively. Then, the matrix of all partial derivatives of J with respect to U can be expressed as
where R is defined by
Proof.
Let be the single-entry matrix being one in entry and zero otherwise. By differentiating J with respect to , we obtain
Given a stable discrete-time system (1), one-sided projection -optimal model order reduction aims to minimize the squared -norm of the error between the stable full system and the stable reduced-order system; i.e., it seeks to solve the minimization problem
where and X are the solutions of (31) and (32), respectively. Since U is a point of the Stiefel manifold , the minimization problem can be also expressed as an optimization problem based on the Stiefel manifold .
It is known [23] that in the continuous-time case, the minimization problem can be equivalently rewritten as a minimization problem based on the Grassmann manifold if the cost function depends only on the space spanned by the columns of the matrix U, i.e., if for any . For the discrete-time system considered in this paper, due to the same reason, minimization problem (41) can be reformulated as an equivalent optimization problem based on a Grassmann manifold:
Clearly, the solution U of minimization problem (42) must be a point of such that the gradient , i.e., U, satisfies the necessary conditions
where is given by (36).
3.1. Solving the Optimization Problem via the Gradient Flow Approach
In this subsection, we consider the application of the gradient flow method reviewed in Section 2.3 to the optimization problem over a Grassmann manifold (42).
Suppose that for , , with , is the j-th iterate. From Theorem (1), it follows that the matrix of all partial derivatives of J with respect to can be expressed as
where is given by
where , , , and are the solutions of Equations (31), (32), (34), and (35) with , respectively. With , we see from (12) that the gradient of J at point in Grassmann manifold can be formulated as
Then, the gradient flow method computes the -th approximate solution to minimization problem (41) as
where is the singular value decomposition of the negative gradient of J at .
In iteration (46), the step size can be chosen according to the inexact line search with the Armijo rule as follows: Define
where t denotes the step size parameter. For some given , the Armijo step size is , where l is the smallest nonnegative integer such that the adequate reduction condition holds:
Let and be the solutions of (31) and (32) with , and and be the solutions of (31) and (32) with . Then, by (29), we have
This shows that we do not need to solve (30) for P to apply the inexact linear search with the Armijo condition.
We now outline the gradient flow method for model reduction of discrete-time system (1) in Algorithm 1.
| Algorithm 1 Gradient flow method |
|
We note that a detailed complexity analysis of the gradient flow method for model reduction of continuous-time LTI systems is presented in [23]. With a similar analysis, it is easy to show that the computational complexity of Algorithm 1 is , where k is some fixed integer; denotes the number of nonzero elements; N is the number of iterations; is the maximum number of search steps; are the input, output, and reduced dimensions, respectively; and is the complexity of iterative methods like GMRES for solving linear systems with a coefficient matrix . The computational complexity of Algorithms 2 and 3 can be analyzed similarly.
3.2. Solving the Optimization Problem via Sequentially Quadratic Approximation
Sequentially quadratic approximation [41] is one of the most known methods in the optimization field. In [23], Xu and Zeng applied this technique to minimization problem (42), which arises in the model reduction of continuous-time systems. In this subsection, we extend sequentially quadratic approximation for the model reduction of discrete-time systems.
Let , with , be the j-th projection matrix generated in the iteration. In sequentially quadratic approximation, the key step is to approximate the cost , given in (42), by the following quadratic function:
where and are the solutions of Stein Equations (31) and (32) with , respectively. It is clear that the partial derivative matrix of is given by . By setting , we get if is invertible. Now, we can define a possible search direction by projecting the difference to the tangent space of at point ; i.e., is defined by
In order to generate a gradient-related sequence , it is proposed in [41] to select the new search for obtaining according to the following rule: Given constants and , if
then the search direction is ; otherwise, the search direction is the negative gradient direction of at , i.e., . With the SVD of the new search direction, , the -iterate is generated according to iteration scheme (46) with the step size satisfying the Armijo rule in (47).
We summarize the sequentially quadratic approximation method for the -optimal model reduction of discrete-time system (1) in Algorithm 2.
| Algorithm 2 Sequentially quadratic approximation method |
|
| Algorithm 3 Two-sided projection method |
|
4. Two-Sided Projection via Optimization on Grassmann Manifold
Zeng and Lu [24] considered a two-sided projection for -optimal model reduction of continuous-time linear systems. The cost is minimized over a Grassmann manifold with two project matrices. In this subsection, we will extend this two-sided projection approach to discrete-time linear system (1).
The two-sided projection based approach naturally involves a pair of biorthogonal matrices , i.e., . The coefficient matrices of the reduced-order systems can be obtained via projections as follows:
In the two-sided projection framework, the cost function for the error system is given by
and the minimization problem to be solved is given by the problem
As shown in [24], this optimal problem (51) can be equivalently formulated as the following minimization problem over a Grassmann manifold with two variable matrices:
We give the expressions of the matrices of all partial derivatives of J with respect to V and W in the following theorem, the proof of which is similar to that of Theorem 1 and is thus omitted.
Theorem 2.
The matrices of all partial derivatives of J with respect to can be expressed as
With the help of the matrices of all partial derivatives of , from [24] (Theorem 1), the gradients and of J at points and of Grassmann manifold are explicitly written as
respectively, where are given by (61) and (62), respectively.
So, for the two-sided projection, must satisfy the necessary conditions
Solving the Optimization Problem via an Alternating Direction Approach
Now we consider applying the alternating direction algorithm with sequentially quadratic approximation of the cost to solve minimization problem (52).
Suppose that for , and are the known projection matrices. Firstly, we approximate the cost , expressed by (53), in the V-direction by the following quadratic function:
where and are the solutions of the Stein equations
respectively, with and . It is not difficult to see that the partial derivative of the quadratic function can be formulated as
By setting , we can obtain a new matrix if is invertible. However, we cannot use as a projection matrix since does not usually satisfy . In [24], an approach is proposed to construct the -th iterate based on the difference . We outline the scheme as follows:
- Firstly, by projecting onto the tangent space of at point , a search direction is generated by
- Secondly, we define the search matrix , which should satisfy , bywhere t denotes the step length parameter.
- Finally, the -th iterate is constructed by the inexact line search with the Armijo rule. That is, for some given , we find the smallest positive number l so that satisfies the adequate reduction conditionand then set .
After obtaining the projection matrix , the iterate for W is constructed in an analogous way. From (54), the quadratic function , which approximates at point , is defined by
where and satisfy
respectively, with and . Similarly, by setting , we obtain . Following the steps in the construction of , the new projection matrix is established as follows:
- Construct the search direction by
- Define the search matrix by
- For some given , find the smallest positive number l so thatand then set .
In summary, we propose a two-sided projection method based on Grassmann manifold optimization for model reduction of discrete-time system (1), as presented in Algorithm 3.
5. Implementation Issues
In this section, we consider some issues relative to the implementation of the algorithms proposed in this paper, including how to choose the initial projection matrices and the termination criterion and select suitable methods for the Stein equations.
5.1. Initial Projection Matrix Selection
For the algorithms proposed in this paper, the first possible choice is to construct initial projection matrices randomly. For one-sided projection, after generating the random matrix, we need to compute its QR decomposition to obtain an orthogonal initial matrix . For the two-sided projection method, we first construct two random matrices and then set so that the biorthogonal condition holds.
Secondly, we can choose the initial projection matrices so that the subspaces spanned by them are the block Krylov subspace or the rational block Krylov subspace:
Recall that the first-order necessary conditions for model reduction of a discrete-time system state that the optimal reduced system interpolates the full system at the mirror images with respect to the unit circle of the poles of the reduced system; see, for example, ref. [35]. So the reciprocal of the approximate eigenvalue or the Ritz value of A is the suitable choice of shift parameters for the rational block Krylov subspace. In numerical experiments, we will test random selection and the choice of the block Krylov subspace and the rational block Krylov subspace.
5.2. Termination Criterion
As we know, in fact, the algorithms proposed in this paper are developed for solving optimization problems over Grassmann manifolds. So we should stop these algorithms if the norm of the gradient of the cost function J at some approximate solution is small.
For one-sided projection methods such as Algorithms 1 and 2, the cost J is defined based on one matrix variable U and is given in (42). Therefore, the iteration should be naturally terminated if the gradient at point satisfies
where is a prescribed tolerance and , with given by (44).
For two-sided projection, the cost J has two matrix variables . It is known that if is the solution of minimization problem (52), the gradients at W and V should satisfy . So the stopping rule for Algorithm 3 can be defined by
where and are computed by using (63) and (64), respectively.
5.3. Solving Stein Equations
In this subsection, we consider the numerical solution of two classes of Stein equations appearing in our algorithms.
The Stein equation of the type shown in (66), called the discrete-time Lyapunov equation, has a small dimension r. This kind of equations can be solved by the standard direct method provided in [42], which is a direct extension of the well-known Bartels–Stewart algorithm [43] for continuous Lyapunov equations. Note that in order to obtain the real solution, the real Schur decomposition [13] should be used. In [44], low-rank methods were proposed for solving discrete-time projected Lyapunov equations. For a comprehensive review of numerical methods for solving matrix equations, we refer to [45].
We now consider how to solve Stein equations like (67), which are also known as discrete-time Sylvester equations. We point out that this kind of matrix equations cannot be solved by a Bartels–Stewart-like algorithm. The reason is that the matrix A has a large dimension n, so computing the real Schur decomposition of A requires time complexity.
For simplicity of notation, we drop the index j in (67). This kind of equation is now of the form
with , , , , and . Usually, A is sparse, and is dense. So, this class of matrix equations can be called the sparse–dense discrete-time Sylvester equations. We note that in [46], a method based on Schur decomposition of the dense matrix is proposed for solving sparse–dense continuous-time Sylvester equations, which arise in model order reduction of continuous-time LTI systems.
Here, we will use a variant of the real Bartels–Stewart algorithm to obtain the real solution of (70). Let be the real Schur decomposition of with
where is a real number or a real matrix having a pair of complex conjugate eigenvalues. With this decomposition, Equation (70) is transformed to the equivalent equation
where and . We partition and as
where and have the same number of columns of . From (71) and the above partition, we have
with .
If is a number, it follows from (72) that
Now we consider the case where is a real matrix. Let be one eigenvalue of with the corresponding eigenvector . Then, is also an eigenpair of . Thus, we have
Define
Then, we have
i.e.,
where and denote the real and imaginary parts of a complex number or a complex matrix, respectively, and ȷ is the imaginary unit.
So, we obtain a real formula for :
We summarize the framework of this method for solving the discrete-time Sylvester Equation (70) as follows:
- Compute the real Schur decomposition , and set .
- Finally, compute .
We remark that this approach involves complex arithmetic and the storage of complex matrices only in computing and thus avoids complex arithmetic and the storage of complex matrices as much as possible. Moreover, for a block , only one system of linear equations needs to be solved. The main computation in this method and thus in the algorithms proposed in this paper for model reduction is the solution of linear systems with a coefficient matrix . Thus, sparse direct solvers or Krylov subspace methods like GMRES or QMR [13] can be applied, which makes these model reduction methods suitable for large-scale problems.
6. Numerical Experiments
In this section, we present two examples to illustrate the performance of -optimization over Grassmann manifolds for the model order reduction of linear discrete-time systems. We compare Algorithms 1–3 with balanced truncation (BT) and discrete-time IRKA. We note that the step size in Algorithms 1–3 is chosen to satisfy the Armijo rules. So, stability is preserved if the full-order system is stable (see [23]). However, IRKA may not necessarily generate stable reduced systems even if the original system is stable. The numerical experiments are carried out in Matlab 2022a on an Intel i5 processor with 2.3 GHz and 16 GB memory. The relative error is defined by
where H and are the transfer functions of the full system and the reduced system, respectively. In the numerical examples, we make the following parameter choices: , , , , and .
6.1. Example 1
In the first numerical example, the linear discrete-time system is obtained from the single-input single-output ISS-1R module (cf. [47]), which is a continuous-time model of the international space station. We discretize this model by using a semi-explicit Euler and a semi-implicit Euler method. The discrete-time model is a system with dimension . The numerical results are presented in Table 1. Table 1 shows that of all three algorithms proposed in this paper, the two-sided projection method (Algorithm 3) has the least relative errors for the reduced orders . Moreover, Algorithm 3 and IRKA have almost the same relative errors and are slightly better than BT. In this numerical example, the algorithms tested in this paper have preserved the stability of the original system.
Table 1.
relative errors of Example 1.
6.2. Example 2
In the second numerical example, we discretize the continuous-time CD player model (cf. [47]) by using the same scheme as in the first example to generate a discrete-time system. The discrete-time model is a system with dimension and has 2 inputs and 2 outputs. The relative errors are given in Table 2. It is clear from Table 2 that all three algorithms proposed in this paper are effective. For this example, Algorithm 3 and IRKA have also almost the same relative errors and are slightly better than BT. We point out that Algorithm 3 preserves stability in the reduced system, but IRKA does not preserve stability in this numerical example.
Table 2.
relative errors of Example 2.
7. Conclusions
In this paper, we have investigated -optimal model reduction over Grassmann manifolds for discrete-time time-invariant systems. We have presented necessary conditions for optimization problems over Grassmann manifolds. We have considered two kinds of projections: one-sided projection and two-sided projection. For one-sided projection, we apply the gradient flow method and the sequentially quadratic approximation method to solve the related optimization problem. For two-sided projection, we apply the strategies of alternating direction iteration and sequentially quadratic approximation to the minimization problem. Numerical examples are presented to demonstrate the effectiveness of the proposed algorithms.
Author Contributions
Conceptualization, Y.L.; Methodology, L.Z.; Software, L.Z.; Writing—original draft, Y.L.; Formal analysis, L.Z. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the Natural Science Foundation of Hunan Province under grant 2024JJ7203, the key Project of Hunan provincial Education Department under grant 23A0577, and the Applied Characteristic Discipline at Hunan University of Science and Engineering.
Data Availability Statement
The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- Anderson, J. Computational Fluid Dynamics; McGraw-Hill Education: New York, NY, USA, 1995. [Google Scholar]
- Tedesco, J.W.; McDougal, W.G.; Ross, C.A. Structural Dynamics: Theory and Applications; Prentice Hall: Upper Saddle River, NJ, USA, 1999. [Google Scholar]
- Reis, T. Systems Theoretic Aspects of PDAEs and Applications to Eletrical Circuits. Ph.D. Thesis, Technische Universitat Kaiserlautern, Kaiserlautern, Germany, 2006. [Google Scholar]
- Moran, J. An Introduction to Theoretical and Computational Aerodynamics; John Wiley & Sons: New York, NY, USA, 1984. [Google Scholar]
- Allen, J.J. Micro Electro Mechanical System Design; CRC Press: Boca Raton, FL, USA, 2005. [Google Scholar]
- Coiffier, J. Fundamentals of Numerical Weather Prediction; Cambridge University Press: Cambridge, UK, 2012. [Google Scholar]
- Miao, J. Economic Dynamics in Discrete Time; MIT Press: Cambridge, UK, 2014. [Google Scholar]
- Möller, D.P.F. Introduction to Transportation Analysis, Modeling and Simulation; Springer: London, UK, 2014. [Google Scholar]
- Gu, G. Discrete-Time Linear Systems: Theory and Design with Applications; Springer: New York, NY, USA, 2012. [Google Scholar]
- Moore, B.C. Principal component analysis in linear systems: Controllability, observability, and model reduction. IEEE Trans. Autom. Control 1981, 26, 17–32. [Google Scholar] [CrossRef]
- Glover, K. All optimal Hankel-norm approximations of linear multivariable systems and their L∞-errors bounds. Int. Control 1984, 39, 1115–1193. [Google Scholar] [CrossRef]
- Liu, Y.; Anderson, B.D.O. Singular perturbation approximation of balanced systems. Int. J. Control 1989, 50, 1379–1405. [Google Scholar] [CrossRef]
- Golub, G.H.; Loan, C.F.V. Matrix Computations, 3rd ed.; Johns Hoplins University Press: Baltimore, MD, USA, 1996. [Google Scholar]
- Ruhe, A. Rational Krylov algorithms for nonsymmetric eigenvalue problems. II: Matrix pair. Linear Algebra Its Appl. 1994, 197, 282–295. [Google Scholar] [CrossRef]
- Clark, J.V.; Zhou, N.; Bindel, D.; Schenato, L.; Wu, W.; Demmel, J.; Pister, K.S.J. 3d MEMS simulation using modified nodal analysis. Proc. Microscale Syst. Mech. Meas. Symp. 2000, 68–75. [Google Scholar]
- Craig, R.R. Structural Dynamics: An Introduction to Computer Methods; John Wiley & Sons: New York, NY, USA, 1981. [Google Scholar]
- Freund, R.W. Krylov-subspace methods for reduced-order modeling in circuit simulation. J. Comput. Appl. Math. 2000, 123, 395–421. [Google Scholar] [CrossRef]
- Freund, R.W. Model reduction methods based on Krylov susbspace. Acta Numer. 2003, 12, 267–319. [Google Scholar] [CrossRef]
- Grimme, E. Krylov Projection Methods for Model Reduction. Ph.D. Thesis, University of Illinois, Urbana–Champaign, Champaign, IL, USA, 1997. [Google Scholar]
- Gugercin, S.; Antoulas, A.C.; Beattie, C. H2 model reduction for large-scale linear dynamical systems. SIAM J. Matrix Anal. Appl. 2008, 30, 609–638. [Google Scholar] [CrossRef]
- Van Dooren, P.; Gallivan, K.A.; Absil, P.A. H2-optimal model reduction of MIMO systems. Appl. Math. Lett. 2008, 21, 1267–1273. [Google Scholar] [CrossRef]
- H2-optimal model reduction with higher-order poles. SIAM J. Matrix Anal. Appl. 2010, 31, 2738–2753. [CrossRef]
- Xu, Y.; Zeng, T. Fast optimal h2 model reduction algorithms based on Grassmann manifold optimization. Int. J. Numer. Anal. Model. 2013, 10, 972–991. [Google Scholar]
- Zeng, T.; Lu, C. Two-sided Grassmann manifold algorithm for optimal H2 model reduction. Int. J. Numer. Methodsin Eng. 2015, 104, 928–943. [Google Scholar] [CrossRef]
- Benner, P.; Cao, X.; Schilders, W. A bilinear H2 model order reduction approach to linear parameter-varying systems. Adv. Comput. Math. 2019, 45, 2241–2271. [Google Scholar] [CrossRef]
- Xu, K.; Dong, L.; Wang, B.; Li, Z. Preserving-periodic Riemannian descent model reduction of linear discrete-time periodic systems with isometric vector transport on product manifolds. Appl. Math. Lett. 2025, 171, 231–254. [Google Scholar] [CrossRef]
- Xu, K.; Li, Z.; Benner, P. Parametric interpolation model order reduction on Grassmann manifolds by parallelization. IEEE Trans. Circuits Syst. 2025, 72, 198–202. [Google Scholar] [CrossRef]
- Xu, K.; Jiang, Y.; Yang, Z. H2 order-reduction for bilinear systems based on Grassmann manifold. J. Frankl. Inst. 2015, 352, 4467–4479. [Google Scholar] [CrossRef]
- Otto, S.E.; Padovan, A.; Rowley, C.W. Optimizing oblique projections for nonlinear systems using trajectories. SIAM J. Sci. Comput. 2022, 44, A1681–A1702. [Google Scholar] [CrossRef]
- Padovan, A.; Vollmer, B.; Bodony, D.J. Data-driven model reduction via non-intrusive optimization of projection operators and reduced-order dynamics. SIAM J. Appl. Dyn. Syst. 2024, 23, 3052–3076. [Google Scholar] [CrossRef]
- Buchfink, P.; Glas, S.; Haasdonk, B.; Unger, B. Model Reduct. Manifolds: A Differ. Geom. Framework. Phys. D Nonlinear Phenom. 2024, 468, 134299. [Google Scholar] [CrossRef]
- Zimmermann, R. Manifold interpolation. In System and Data-Driven Methods and Algorithms; De Gruyter: Berlin, Germany, 2021; pp. 229–274. [Google Scholar]
- Sashittal, P.; Bodony, D. Low-rank dynamic mode decomposition using Riemannian manifold optimization. In Proceedings of the IEEE Conference on Decision and Control (CDC), Miami Beach, FL, USA, 17–19 December 2018; pp. 2265–2270. [Google Scholar]
- Wilson, D.A. Optimum solution of model reduction problem. Proc. Inst. Electr. Eng. 1970, 117, 1161–1165. [Google Scholar] [CrossRef]
- Bunse-Gerstner, A.; Kubalińska, D.; Vossen, G.; Wilczek, D. h2-Norm Optimal Model Reduction for Large Scale Discrete Dynamical MIMO Systems. J. Comput. Appl. Math. 2010, 233, 1202–1216. [Google Scholar] [CrossRef]
- Absil, P.-A.; Mahony, R.; Sepulchre, R. Optimization Algorithms on Matrix Manifolds; Princeton University Press: Princeton, NJ, USA, 2008. [Google Scholar]
- Edelman, A.; Arias, T.A.; Smith, S.T. The geometry of algorithms with orthogonality constraints. SIAM J. Matrix Anal. Appl. 1998, 20, 303–353. [Google Scholar] [CrossRef]
- Antoulas, A.C. Approximation of Large-Scale Dynamical Systems; SIAM: Philadelphia, PA, USA, 2005. [Google Scholar]
- Highham, N.J. Functions of Matrices: Theory and Computation; SIAM: Philadelphia, PA, USA, 2008. [Google Scholar]
- Yan, W.-Y.; Lam, J. An approximate approach to H2-optimal model reduction. IEEE Trans. Autom. Control 1999, 44, 1341–1357. [Google Scholar]
- Nocedal, J.; Wright, S.J. Numerical Optimization; Springer: New York, NY, USA, 1999. [Google Scholar]
- Barraud, A.Y. A numerical algorithm to solve ATXA − X = Q. IEEE Trans. Autom. Control 1977, 22, 883–885. [Google Scholar] [CrossRef]
- Bartels, R.; Stewart, G. Solution of the equation AX + XB = C. Commun. ACM 1972, 15, 820–826. [Google Scholar] [CrossRef]
- Lin, Y. Low-rank methods for solving discrete-time projected Lyapunov equations. Mathematics 2024, 12, 1166. [Google Scholar] [CrossRef]
- Simoncini, V. Computational methods for linear matrix equations. SIAM Rev. 2016, 58, 377–441. [Google Scholar] [CrossRef]
- Benner, P.; Köhler, M.; Saak, J. Sparse-Dense Sylvester Equations in H2-Model Order Reduction; Technical Report MPIMD/11-11; Max Planck Institute Magdeburg Preprint: Magdeburg, Germany, 2011. [Google Scholar]
- Chahlaoui, Y.; Dooren, P.V. Benchmark Examples for Model Reduction of Linear Timeinvariant Dynamical Systems. In Dimension Reduction of Large-Scale Systems, Lecture Notes in Computational Science and Engineering 45; Springer: New York, NY, USA, 2005; pp. 379–392. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).