1. Introduction
Spectral theory of dynamical systems shifts the focus of investigation of dynamical systems behavior away from trajectories in the state space and towards spectral features of an associated infinite-dimensional linear operator. Of particular interest is the composition operator—in a measure-preserving setting called the Koopman operator [
1,
2,
3,
4,
5]. Its spectral triple—eigenvalues, eigenfunctions and eigenmodes—can be used in a variety of contexts, from model reduction [
5] to stability and control [
6]. In practice, we only have access to finite-dimensional data from observations or outputs of numerical simulations. Thus, it is important to study approximation properties of finite-dimensional numerical algorithms devised to compute spectral objects [
7]. Compactness is the property that imbues infinite-dimensional operators with quasi-finite-dimensional properties. Self-adjointness also helps in proving the approximation results. However, the composition operators under study here are rarely compact or self-adjoint. In addition, in the classical, measure-preserving case, the setting is that of unitary operators (and essentially self-adjoint generators for the continuous-time setting [
8]), but in the general, dissipative case, composition operators are neither.
There are three main approaches to finding spectral objects of the Koopman operator:
- 1.
The first, suggested already in [
9] is based on long time weighted averages over trajectories, rooted in ergodic theory of measure-preserving dynamical systems. An extension of that work that captures properties of continuous spectrum was presented in [
10]. This approach was named Generalized Laplace Analysis (GLA) in [
11], where concepts pertaining to dissipative systems were discussed also in terms of weighted averages along trajectories. In that sense, the ideas in this context provide an extension of ergodic theory for capturing transient (off-attractor) properties of systems. For on-attractor evolution, the properties of the method acting on
functions were studied in [
4]. The off-attractor case was pursued in [
4,
12] where Fourier averages (which are Laplace averages for the case when the eigenvalue considered is on the imaginary axis) were used to compute the eigenfunction whose level sets are isochrons, and [
13] in which analysis for general eigenvalue distributions was pursued in Hardy-type spaces. This study was continued in [
14] to construct dynamics-adapted Hilbert spaces. The advantage of the method is that it does not require the approximation of the operator itself, as it constructs eigenfunctions and eigenmodes directly from the data. In this sense, it is close (and in fact related) to the power method of approximating the spectrum of a matrix from data on iteration of a vector. In fact, the methodology extends the power method to the case when eigenvalues can be of magnitude 1. It requires separate computation to first determine the spectrum of the operator, which is also done without constructing it. This can potentially be hard to do because of issues such as spectral pollution—see remarks at the end of
Section 3; also note the general long-standing problems of spectral pollution and computing the full spectrum of Schrödinger operators on a lattice were recently solved in [
15]. The recent work [
16] enables computation of full spectral measures using the combination of resolvent operator techniques (used for the first time in the Koopman operator context in [
17]) and ResDMD—an extension of Dynamic Mode Decomposition (introduced next) technique that incorporates computation of residues from data snapshots (computation of residues was considered earlier in [
18]).
- 2.
The second approach requires construction of an approximate operator acting on a finite-dimensional function subspace i.e., a finite section—the problem that is also of concern in a more general context of approximating infinite dimensional operators [
7,
19,
20]. The best known such method is the Dynamic Mode Decomposition (DMD), invented in [
21] and connected to the Koopman operator in [
22]. It has a number of extensions (many of which are summarized in [
23]), for example, Exact DMD [
24]; Bayesian/subspace DMD [
25]; Optimized DMD [
26,
27]; Recursive DMD [
28]; Variational DMD [
29]; DMD with control [
30,
31]; sparsity promoting DMD [
32]; DMD for noisy systems [
33,
34,
35]. The original DMD algorithm featured state observables. The Extended Dynamic Mode Decomposition [
36] recognizes that nonlinear functions of state might be necessary to describe a finite-dimensional invariant subset of the Koopman operator and provides an algorithm for finite-section approximation of the Koopman operator. A study of convergence of such approximations is provided in [
37], but the convergence was established only along subsequences, and the rate of convergence was not addressed. Here, we provide the first result on the rate of convergence of the finite section approximation under assumptions on the nature of the underlying dynamics. In addition, spectral convergence along subsequences is proven in [
37] under the assumption of the weak limit of eigenfunction approximations not being zero. This condition is hard to verify in practice. Instead, in
Section 5.2, we prove a result that obviates the weak convergence assumption using some additional information on the underlying dynamics. It was observed already in [
9] that, instead of an arbitrary set of observables forming a basis, one can use observables generated by the dynamics—namely time delays of a single observable filling a Krylov subspace—to study spectral properties of the Koopman operator. In the DMD context, the methods developed in this direction are known under the name Hankel-DMD [
38,
39]. It is worth noticing that the Hankel matrix approach of [
38] is in fact based on the Prony approximation and requests a different sample structure than the Dynamic Mode Decomposition. Computation of residues was considered in [
18] to address the problem of spectral pollution, where discretization introduces spurious eigenvalues. As mentioned before, the recent work [
16] provides another method to resolve the spectral pollution problem, introducing ResDMD—an extension of Dynamic Mode Decomposition that incorporates computation of residues from data snapshots. The relationship between GLA and finite section methods was studied in [
40].
- 3.
The third approach is based on the kernel integral operator combined with the Krylov subspace methodology [
41], enabling approximation of continuous spectrum. While GLA and EDMD techniques have been extended to dissipative systems, the kernel integral operator technique is currently available only for measure-preserving (on-attractor) systems.
In this paper, we continue with the development of ergodic theory-rooted ideas to understanding and numerically computing the spectral triple—eigenvalues, eigenfunctions and modes—for the Koopman operator. After some preliminaries, we start in
Section 3 with discussing properties of algorithms of Generalized Laplace Analysis type in Banach spaces. Such results have previously been obtained in Hardy-type spaces [
13], and here, we introduce a Gel’fand-formula-based technique that allows us to expand to general Banach spaces. We continue in
Section 4 with setting the finite-section approximation of the Koopman operator in the ergodic theory context. An explicit relationship of finite section coefficients to dual basis is established. Under assumptions on the underlying dynamics, we provide the first result on the convergence rate under sample size increase in the finite-section approximation. The error in the finite section approximation is analyzed. In
Section 5, we study finite section approximations of the Koopman operator based on Krylov sequences of time-delays of observables, and prove that under certain conditions, the approximation error decreases as the number of samples is increased, without dependence on the dimension of the problem. Namely, the Krylov subspace (Hankel-DMD) methodology has the advantage of convergence in the number of iterates and does not require a basis exponentially large in the number of dimensions. This solves the problem of the
choice of observables, since the dynamics selects the basis by itself. In
Section 6, we discuss an alternative point of view on the DMD approximations which is not related to finite sections, but samples of continuous functions on finite subsets of the state-space. The concept of weak eigenfunctions is discussed, continuing the analysis in [
37]. We conclude in
Section 7.
2. Preliminaries
For a Lipshitz-continuous (ensuring global existence and uniqueness of solutions) dynamical system
defined on a manifold
(i.e.,
—where we by slight abuse of notation identify a point in a manifold
M with its vector representation
in
), where
is a vector and
is a possibly nonlinear vector-valued smooth function, of the same dimension as its argument
, denote by
the position at time
t of trajectory of (
1) that starts at time 0 at point
. We call the family of functions
the flow.
Denote by
an arbitrary, vector-valued observable
. The value
of the observable
that the system trajectory starting from
at time 0 sees at time
t is
Note that the space of all observables
is a linear vector space. The family of operators
acting on the space of observables parameterized by time
t is defined by
Thus, for a fixed time
,
maps the vector-valued observable
to
. We will call the family of operators
indexed by time
t the Koopman operator of the continuous-time system (
1). This family was defined for the first time in [
1], for Hamiltonian systems. In operator theory, such operators, when defined for general dynamical systems, are often called composition operators, since
acts on observables by composing them with the mapping
[
3]. Discretization of
for times
leads to the
-mapping
with the discrete dynamics
and the associated Koopman operator
U defined by
Let
be a space of observables and
the Koopman operator associated with a map
T (note this means that
if
). Appropriate (dynamics-adapted) spaces are discussed in [
14]. A function
is an eigenfunction of
U associated with eigenvalue
provided
Let
be the spectrum of
U. The operator
U is called scalar [
42] on
provided
where
E is a family of spectral projections forming resolution of the identity, and the integral is over
. Further, the operator
U is called spectral provided
where
S is scalar and
N quasi-nilpotent. Examples of functional spaces in which Koopman operators are scalar and spectral are given in [
14]. Let
be a vector of observables. For a scalar operator
U, the Koopman mode
of
associated with an eigenvalue
of algebraic multiplicity 1 is given by
where
is the unit norm eigenfunction associated with
, and
Note that, denoting by
the projection on the eigenspace associated with the eigenvalue
, we have
since
and
by one of the key properties of the spectral resolution [
42]. Now,
for some constant
, proving that
is well-defined and independent of
.
Remark 1. Note that in the more general case with algebraic multiplicities of eigenvalues larger than 1, an analogous definition of the Koopman mode can be obtained. For example, if algebraic multiplicity and geometric multiplicity are 2 and there are two linearly independent eigenfunctions and associated with the eigenvalue λ of multiplicity 2, and we are computing , then (10) contains an additional term on the RHS, , and similarly for , forming 2 equations. In the case of spectral operators, one works similarly, but the added complexity is in the use of generalized eigenfunctions [14]. We assume that the dynamical system
T has a Milnor attractor
such that for every continuous function
g, for almost every
with respect to an a priori measure
on
M (without loss of generality as we can replace
M with the basin of attraction of
) the limit
exists. This is the case, e.g., for smooth systems on subsets of
with Sinai–Bowen–Ruelle measures, where
is the Lebesgue measure [
43]. For such systems, Hilbert spaces on which the Koopman operator is spectral have been constructed in [
14].
3. Generalized Laplace Analysis
An example of what we call Generalized Laplace Analysis (GLA) is the computation of eigenspace at 0 (namely, invariants) of dynamical systems using time averages: recall
is the time average at initial condition
of the function
h under the dynamics of
. For fixed point attractors
As shown previously, this is valid in a much larger context: limit cycle attractors, toroidal attractors, Milnor attractors, and measure-preserving systems.
We generalize the idea that averages along trajectories produce eigenfunctions, by introducing weights:
where
is a function of time—typically a (possibly complex) exponential, and
is a sampling time interval. If we have a vectorized set of initial conditions
, then we can generate a data matrix
Vectorizing
, we get
H is the
data matrix. For
, we get
, where
is a vector of 1’s with
components. To obtain eigenfunctions using Fourier averages, as developed in [
12], we set
, to obtain
Both of the above examples were for the case when corresponding to eigenvalues , both on the imaginary axis. In the next subsection, we provide a general theorem that deals with eigenvalues distributed arbitrarily in the complex plane.
GLA for Fields of Observables
Many of the problems of interest in applications feature a distributed field of observables. For example, time evolution of temperature in a linear rod described by the coordinate , is , where is the initial condition that belongs to the state space of all possible temperature distributions satisfying the boundary conditions, and t is time. We will set our analysis up having this example in mind—namely, we consider a field of observables , where is in state space, and is an indexing set—and consider the time evolution of such observables starting from an initial condition .
Let be a bounded field of observables , continuous in , where the observables are indexed over elements of a set A, and M is a compact metric space. We will occasionally drop the dependence on the state-space variable and denote and the iterates of f by . Let U be the Koopman operator associated with a map . We assume that U is bounded, and acting in a closed manner on a Banach space of continuous functions C (this does not have to be the space of all continuous functions on M, see the remark after the theorem).
Theorem 1. (Generalized Laplace Analysis).
Let be simple eigenvalues of U such that and there are no other points λ in the spectrum of U with . Let be the eigenfunction of U associated with . Then, the Koopman mode associated with is obtained by computingwhere , is an eigenfunction of U with and is the k-th Koopman mode. Proof. We introduce the operator
Then, for some function
consider
where the last line is obtained by boundedness of
g. Due to the boundedness of
U and continuity of
g, we have
This is obtained as the consequence of the so-called Gel’fand formula that states that for a bounded operator
V on a Banach space
X,
where
is the spectral radius of
V [
44] (note that in our case
). Thus, the first term in (
22) vanishes in the limit. Denoting
where the convergence is again obtained from the Gel’fand formula, utilizing the assumption on convergence of time averages and (
23). Thus, we obtain
and, thus,
is an eigenfunction of
U at eigenvalue
(note that
by the fact that partial sums form a Cauchy sequence). If we have a field of observables
, parameterized by
, we get
since
is an eigenfunction of
U at eigenvalue
, so for every
, it is just a constant (depending on
) multiple of the eigenfunction
of norm 1. If we denote
(note is a bounded projection operator), we can split the space of functions C into the direct sum .
Now, let
. Consider the space of observables
complementary to the subspace
spanned by
. The operator
, the restriction of
U to
has eigenvalues
. Since
does not have a component in
, we can reduce the space of observables to
, on which
satisfies the assumptions of the theorem, and obtain
If we have a field of observables
, then
and, thus,
□
In other words, is the skew-projection of the field of observables on the eigenspace of the Koopman operator associated with the eigenvalue .
Remark 2. The assumptions on eigenvalues in the above theorem would not be satisfied for dynamical systems whose eigenvalues are dense on the unit circle (e.g., a map that, as approaches a unit circle in the complex plane on which the dynamics is given by , where ω is irrational w.r.t. π). However, in such a case, the space of functions can be restricted to the span of functions , and the requirements of the theorem would be satisfied. This amounts to restricting the observables to a set with finite resolution, which is standard in data analysis.
Remark 3. Function spaces in which Koopman operators are spectral are typically special tensor products of on-attractor Hilbert spaces—for example, where μ is the physical invariant measure—and off-attractor spaces of functions that are continuous or possess additional smoothness [14]. Provided we do not restrict the on-attractor part to a finite-dimensional subset like we did in the previous remark, the above theorem would apply to the off-attractor subset (which is an ideal set of functions that vanish a.e. on the attractor). However, the on-attractor Koopman modes can be obtained a.e. using the same procedure as above, and results relying on the Birkhoff’s Ergodic Theorem, valid in , as in [4,5,45,46]. In principle, one can find the full spectrum of the Koopman operator by performing Generalized Laplace Analysis, where Theorem 1 is used on some function
starting from the unit circle, successively subtracting parts of the signal corresponding to eigenvalues with decreasing
. In practice, such computation can be unstable, since at large
t, it involves a multiplication of a very large with a very small number. In addition, the eigenvalues are typically not known a priori. A large class of dynamical systems have eigenvalues on and inside the unit circle (or left half of the complex plane inclusive of the imaginary axis in the continuous time case) [
14]. The eigenvalues on the unit circle can be found using the Fast Fourier Transform (FFT). Once the contributions to the dynamics from those eigenvalues are subtracted, the next largest set of eigenvalues have magnitude less than 1. Thus, the power method would enable finding the magnitude
of the resulting eigenvalue. Scaling the operator (restricted to the space of functions not containing components from eigenspaces corresponding to eigenvalues of magnitude 1) with that magnitude, FFT can be performed again to identify the arguments of the eigenvalues of magnitude
. Alternatively, as shown in the next section, we describe the finite section method, in which the operator is represented in a basis, and a finite-dimensional truncation of the resulting infinite matrix—a finite section—is used to approximate its spectral properties. Under some conditions [
37], increasing the dimension of the finite section and the number of sample points, eigenvalues of the operator can be obtained.
4. The Finite Section Method
The GLA method for approximating eigenfunctions (and thus modes) of the Koopman operator, analyzed in the previous section, was proposed initially in [
4,
5,
9] in the context of on-attractor (measure-preserving) dynamics, and extended to off-attractor dynamics in [
11,
12,
13,
39,
47]. It is predicated on the knowledge of (approximate) eigenvalues—since the eigenvalues need to be known a priori to be able to perform weighted trajectory sums in (
20). There is always the eigenvalue 1 that is known, and the trajectory sums in that case lead to invariants of the dynamics [
45,
46]. Other eigenvalues with modulus 1 can be approximated using signal processing methods (see e.g., [
39]). Importantly, the GLA does not require the knowledge of an approximation to the Koopman operator and is in effect a sampling method which avoids the curse of dimensionality. In contrast, DMD-type methods, invented initially in [
21] without the Koopman operator background, and connected to the Koopman operator setting in [
22] produce a matrix approximation to the Koopman operator. There are many forms of the DMD methodology, but all of them require a choice of a finite set of observables that span a subspace. In this section, we analyze such methods in the context of finite section of the operator and explore connections to the dual basis.
4.1. Finite Section and the Dual Basis
Consider the Koopman operator acting on an observable space
of functions on the state space
M, equipped with the complex inner product
(note that we are using the complex inner product linear in the first argument here; the physics literature typically employs the so-called Dirac notation, where the inner product is linear in its second argument) and let
be an orthonormal basis on
, such that, for any function
, we have
Consider the (not necessarily orthogonal) unconditional basis
. The action of
U on an individual basis function
is given by
where
are now just coefficients of
in the basis. We obtain
and we again have
As in the previous section, associated with any closed linear subspace
of
, there is a projection onto it, denoted
, that we can think of as projection “along” the space
, since, for any
, we have
and, thus, any element of
has projection 0. We denote by
the infinite-dimensional matrix with elements
. Thus, the finite-dimensional section of the matrix
is the so-called compression of
that satisfies
where
is the projection “along”
to the span of the first
n basis functions,
.
The key question now is: how are the eigenvalues of
related to the spectrum of the infinite-dimensional operator
U? This was first addressed in [
37].
Example 1. Consider the translation T on the circle given by Let Then, Thus, from (34) , where for and zero otherwise (the Kronecker delta), and is a diagonal matrix. In this case, the finite section method provides us with the subset of the exact eigenvalues of the Koopman operator. The following example shows how careful we need to be with the finite-section method when the underlying dynamical system has chaotic behavior:
Example 2. Consider the map T on the circle given by This is a mixing map that does not have any eigenvalues of the Koopman operator on except for the (trivial) 1, while its spectrum is the whole unit circle [48]. Let Then, Thus, is given byprovided . In this case, the finite section method fails, as has eigenvalue 0 of multiplicity n. This example illustrates how the condition in [37] that the weak convergence of a subsequence of eigenfunctions of to a function ϕ must be accompanied by the requirement in order that the limit of the associated subsequence of eigenvalues converges to a true eigenvalue of the Koopman operator. In particular, no subsequence of eigenvalues in this case converges to the true eigenvalue of the Koopman operator, since the map is measure preserving, and thus, its eigenvalues are on the unit circle. The example shows the peril of applying the finite section method to find eigenvalues of the Koopman operator when the underlying dynamical system has a continuous spectral part [5] (in this case, Lebesgue [48]) spectrum. Continuous spectrum is effectively dealt with in [10,49] using harmonic analysis and periodic approximation methods, respectively. To apply the finite-section methodology of approximation of the Koopman operator, we need to estimate the coefficients
from data. If we have access to measurements of
N orthogonal functions
on
m points on state space, as indicated in [
37], assuming ergodicity, this becomes possible:
Theorem 2. Let be an orthogonal set of functions in and let T be ergodic on M with respect to an invariant measure μ. Let be a trajectory on M. Then, for almost any Proof. This is a simple consequence of the Birkhoff ergodic theorem ([
50]). Recall that
and the last expression is equal to
by the Birkhoff Ergodic Theorem applied to the function
. □
In the case of non-orthonormal Riesz basis, denote by
the dual basis vectors, such that
where
for any
j, and
if
. For the infinite-dimensional Koopman matrix coefficients, we get
Let us consider the
finite set of independent functions
and the associated dual set
in the span
of
, that satisfy
Note that the functions
are unique, since they are each orthonormal to
vectors in
. Let
and
the orthogonal projection on
(this in effect assumes all the remaining basis functions are orthogonal to
). Then,
since, by self-adjointness of orthogonal projections, and
Now, we have
and thus, since
, the coefficients
are the elements of the finite section
in the basis
. It is again possible to obtain
from data:
Theorem 3. Let be a non-orthogonal set of functions in and let T be ergodic on M with respect to an invariant measure μ. Let be a trajectory on M. Then, for almost any where, for any finite m, are obtained as rows of the matrix , where is the conjugate (Hermitian) transpose of F , and is the column vector . Proof. The fact that
are obtained as rows of the matrix
follows from
where
is the
identity matrix. The rest of the proof is analogous to the proof of Theorem 2. □
Remark 4. The key idea in the above results—Theorems 2 and 3—is that we sample the functions and the dual basis on m points in the state space, and then take the limit . Thus, besides approximating the action of U using the finite section , we also approximate individual functions by their sample on m points. The corollary of the theorems is that the finite sample approximations , obtained by setting the coefficientsconverges to as . This result has been obtained in [51], without the use of the dual basis, relying on the Moore–Penrose pseudoinverse, the connection which we discuss next. We call
F the
data matrix. Note that the matrix
is the so-called Moore–Penrose pseudoinverse of
F. Using matrix notation, from (
59), the approximation of the finite section can be written as
where
,
and
is the column vector
If we now assume that there is an eigenfunction-eigenvalue pair
of
U such that
, then
Thus, the eigenvalue will be in the spectrum of . More generally, it is known that an operator U and a projection commute if and only if is an invariant subspace of U. Thus, the spectrum of the finite-section operator is a subset of the spectrum of U for the case when is an invariant subspace.
If an eigenfunction
of
U is in
it can be obtained from an eigenvector
of the finite section
as
where
satisfies
, since, for such
,
We have introduced above the dot notation, that produces a function in from an N-vector and a set of functions .
Remark 5. The Theorems 2 and 3 are convenient in their use of sampling along trajectory and an invariant measure, thus enabling construction of finite section representations of the Koopman operator from a single trajectory. However, the associated space of functions is restricted since the resulting spectrum is on the unit circle. Choosing a more general measure ν that has support in the basin of attraction is possible. Namely, when we construct the finite section, we then use a sequence of points that weakly converges to the measure ν, and their images under T, . This is the approach in [51]. The potential issue with this approach is the choice of space—typically, will have a very large spectrum, for example, filling the entire unit disk of the complex plane [52]. In contrast, Hilbert spaces adapted to the dynamics of a dissipative systems can be constructed [14], starting from the ideal set of continuous functions that vanish on the attractor, enabling a natural setting for computation of spectral objects for dissipative systems. Koopman mode is the projection of a field of observables on an eigenfunction of
U. Approximations of Koopman modes can also be obtained using a finite section. Let
be a finite section of
. Let
be a vector observable (thus, a field of observables indexed over a discrete set). Then, the Koopman mode
associated with the eigenvalue
of
U is obtained as
where
are the eigenfunction and the dual eigenfunction associated with the eigenvalue
. Let
be eigenvectors of
, and thus, the associated eigenfunctions of the finite section are
where
. Then, we get the dual basis
where
This is easily checked by expanding:
Thus, the approximation
to the Koopman mode
associated with the eigenvalue
of the finite section reads
Now, assume that
,
Thus, the Koopman modes associated with the data vector of observables are obtained as the left eigenvector of the finite section of the Koopman operator .
Assuming that the approximation of the finite section, the
matrix
has distinct eigenvalues
, we write the spectral decomposition
where
is the diagonal eigenvalue matrix and
is the column eigenvector matrix. From
we get that the data can be
reconstructed by first observing
This represents
N equations with
m unknowns for each column of
. Assuming
, it is an underdetermined set of equations that can have many solutions for columns of
. Then,
is the projection of all these solutions on the subspace spanned by the columns of
F. If
, (
79) is overdetermined, and the solution
is the closest—in least squares sense—to
in the span of the columns of
F.
Note that
is the matrix in which rows are the Koopman modes
:
and, thus,
Using (
68), we get
where
is an eigenfunction of the finite section, and
’s are the columns of
A. Note that
. Using (
80), we get
Remark 6. The novelty in this section is the explicit treatment of the finite section approximation in terms of the dual basis that enables error estimates in the next subsection. The finite section is also known under the name Galerkin projection [36]. The relationship between GLA and finite section methods was studied in [40]. 4.2. Convergence of the Finite Sample Approximation to the Finite Section
The time averages in (
59) converge due to the Birkhoff’s Ergodic Theorem [
50]. In the case when a dynamical system is globally stable to an attractor with a physical invariant measure, the rates of convergence depend on the type of asymptotic dynamics that the system is exhibiting. Namely, the Koopman operator
U, when restricted to measure-preserving, on-attractor dynamics, is unitary. Its spectrum can in that case be written as
, where
denotes the point spectrum corresponding to eigenvalues of
U and
the continous spectrum [
53]. The next theorem describes convergence of the finite sample approximation to
when the asymptotic dynamics has only the point spectrum—e.g., when the attractor dynamics is that of a fixed point, limit cycle or ergodic rotation on a higher dimensional torus:
Theorem 4. Let be a dynamical system with an attractor A and an invariant measure supported on the attractor. Let U be the Koopman operator on , with a pure point spectrum that is either a non-dense set on the unit circle, or generated by a set of eigenvalues whose imaginary parts satisfy the Diophantine conditions . Let be for all . Note that the coefficients in the finite section matrices depend on the initial condition of the trajectory that was used to generate the finite section, with the notation . Then, for almost all initial conditions where is the Frobenius norm.
Proof. We suppress the dependence on
in the notation. The entries
of
(see (
62)) converge a.e. w.r.t.
. Since
T is conjugate to a rotation on an Abelian group [
54], which is either discrete or the dynamics is uniformly ergodic (in which case, by assumption, the Diophantine condition is satisfied), for sufficiently smooth
T and
[
55,
56,
57], we have
and the statement follows by setting
. □
Remark 7. The smoothness of and the Diophantine condition are required in order for the solution of the homological equation to exist [55]. Only finite smoothness is required [55], but we have assumed for simplicity here. The above means that converges to spectrally:
Corollary 1. Let be an eigenvalue of with multiplicity h. Then, for arbitrary , for sufficiently large , there is a set of eigenvalues λ of whose multiplicity sums to h such that
Proof. This follows from continuity of eigenvalues [
58] to continuous perturbations (established by theorem 4). □
Remark 8. If are independent, the convergence estimate above deteriorates to . Presence of continuous spectrum without the strong mixing property can lead to convergence estimates with [56]. Remark 9. Spectral convergence in the infinite-dimensional setting is a more difficult question (see [37] in which only convergence along subsequences was established under certain assumptions). Even if the result could be obtained, the practical question is the convergence in m and N. To address it further, we start with the formula for error in the finite section. 4.3. The Error in the Finite Section
It is of interest to find out how big is the error we are making in the finite section approximations discussed above. We have the following result.
Proposition 1. Let be an eigenfunction of the finite section associated with the eigenvalue and eigenvector . Then, Proof. The first term on the right side of (
87) follows from the definition of
. We then need to show
However, the left side is just , and since , , which proves the claim. □
6. Weak Eigenfunctions from Data
In the sections above we presented finite section approximations of the Koopman operator, starting from the idea that bounded infinite-dimensional operators are, given a basis, represented by infinite matrices, and then truncated those. In this section, we will present an alternative point of view that provides additional insights into the relationship between the finite-dimensional approximation and the operator. As a consequence of this approach, we show how the concept of a weak eigenfunction, first discussed in [
37], arises.
We start again with a vector of observables, . Except when we can consider this problem analytically, we know the values of observables only on a finite set of points in state space, . Assume also that we know the value of at . We can think of as a sample of the observable on .
Consider the case
. There are many
matrices
A such that
One of them is the transpose of the companion matrix (
91)
but there are many values that
can assume, since the only requirement on them is
and there are
m unknowns and 1 equation that determines them. However, the
need not depend on
j, since the operator that maps the vectors
to
is not dependent on
j. Clearly, if there are
m observables, then we get
and, thus, we can determine
uniquely.
If the number of observables
N is larger than
m, then
are elements of an
matrix
F (note that this data matrix is precisely the transpose of the one we have used before, in (
122)) and, thus, there are not enough components in
to solve
This system is overdetermined, so in general does not have a solution. The Dynamic Mode Decomposition method then solves for
using the following procedure: let
P be the orthogonal projection onto span of columns of
F. Then,
has a solution, provided F has rank
m:
is an
N-dimensional vector in the span of the columns of
F and thus can be written as a linear combination of those vectors. In fact, we can write
We now discuss the nature of the approximation of the Koopman operator
U by the companion matrix (
91)
where
obtained from Equation (
147).
Let
be an invariant set for
, where
M is a measure space, with measure
. Consider the space
, of continuous functions in
restricted to
. This is an
m-dimensional vector space. The restriction
of the Koopman operator to
, is then a finite-dimensional linear operator that can be represented in a basis by an
matrix. An explicit example is given when
represent successive points on a periodic trajectory, and the resulting matrix representation in the standard basis is the
cyclic permutation matrix
If
is not an invariant set, an
approximation of the reduced Koopman operator can still be provided. Namely, if we know
m independent functions’ restrictions
,
in
and we also know
,
, we can provide a matrix representation of
. However, while in the case where
is an invariant set, the iterate of any function in
can be obtained in terms of the iterate of
m independent functions, for the case when
is not invariant, this is not necessarily so. Namely, the fact that
is not invariant means that functions in
do not necessarily experience linear dynamics under
. However, one can take
N observables
,
, where
, and approximate the nonlinear dynamics using linear regression on
where
—i.e., by finding an
matrix
C that gives the best approximation of the data in the Frobenius norm,
We have the following:
Theorem 7. Let be a measure μ-preserving transformation on a metric space M, and let be a trajectory such that, when , becomes dense in a compact invariant set . Then, for any N-vector of observables , we have Proof. By density of
, for sufficiently large
M,
implies
for some
. By continuity of observables,
for some constant
D. Taking
M sufficiently large makes
. □
Consider an m-dimensional eigenvector
of
, associated with the eigenvalue
. Since the eigenvector satisfies
we have
Thus,
can be considered as an eigenfunction on the finite set
. On the last point of the sample,
, we have
Let us now consider the concept of the weak eigenfunction, or eigendistribution. Let
be some prior measure of interest on
M. Let
be a bounded function that satisfies
. We construct the functional
L on
by defining
Set
and we get
Clearly, this is satisfied if
is a continuous eigenfunction of
U at eigenvalue
. However, Equation (
157) is applicable for cases with much less regularity. Namely, if
is a measure and
the associated linear functional, then we can define the action of
U on
L by
Consider, for example, a set of points
and assume that for every continuous
h there exists the limit
Then, by the Riesz representation theorem, there is a measure
such that
Definition 1. Let a measure μ be such that the associated linear functional L satisfies for some . Then, μ is called a weak eigenfunction of U.
Now, we have
proving the following theorem:
Theorem 8. Consider a set of points , on a trajectory of T, and assume that for every continuous h, there exists the limit Then, the μ associated with by is a weak eigenfunction of U associated with the eigenvalue λ.
From the above, it follows that the left eigenvectors of
are approximations of the associated (possibly weak) Koopman modes, as it is assumed that ł is such an eigenvector,
Then,
is the projection of
on the eigenspace spanned by the eigenvector
. Moreover, since
the statement can be obtained in the limit
by the so-called Generalized Laplace Analysis (GLA) that we described in
Section 3.
Remark 12. The standard interpretation of the Dynamic Mode Decomposition (e.g., on Wikipedia) was in some way a transpose of the one presented here: the observables (interpreted as column vectors) were assumed to be related by a matrix . Instead, in the nonlinear, Koopman operator interpretation, each row is mapped into its image, and this allows interpretation on the space of observables. This is particularly important in the context of evolution equations, for example, fluid flows, where the evolution of the observables’ field—the field of velocity vectors at different spatial points—is not evolving linearly.