Abstract
The identification of partially observed continuous nonlinear systems from noisy and incomplete data series is an actual problem in many branches of science, for example, biology, chemistry, physics, and others. Two stages are needed to reconstruct a partially observed dynamical system. First, one should reconstruct the entire phase space to restore unobserved state variables. For this purpose, the integration or differentiation of the observed data series can be performed. Then, a fast-algebraic method can be used to obtain a nonlinear system in the form of a polynomial dynamical system. In this paper, we extend the algebraic method proposed by Kera and Hasegawa to Laurent polynomials which contain negative powers of variables, unlike ordinary polynomials. We provide a theoretical basis and experimental evidence that the integration of a data series can give more accurate results than the widely used differentiation. With this technique, we reconstruct Lorenz attractor from a one-dimensional data series and B. Muthuswamy’s circuit equations from a three-dimensional data series. 
    1. Introduction
The problem in nonlinear dynamical system reconstruction from data series can be described as follows: For a given data series, a system must be found that produces data series that are close in a certain sense to the initial series. Obtaining system models using a limited set of observations is crucial for the reverse engineering of gene regulatory networks [], analysis of chemical oscillations [], studies on biological circadian rhythms [], reconstruction of neuronal models [], etc.
When systems described by ordinary differential equations (ODEs) are considered, one can specify the problem statement as follows: For a given data series  and its time derivative  a system of ODEs must be reconstructed as
      
      
        
      
      
      
      
    
      where  is a vector function. Usually, some information about  is known apriori, for instance, the possible elementary operations constituting it. These operations usually construct a group or, more specifically, a ring. One such ring is a polynomial ring which includes a set of polynomials in several variables with coefficients belonging to a certain field such as a field of real or complex numbers [,]. Dynamical systems with polynomial derivative functions are often referred to as polynomial dynamical systems (PDSs). In these systems,  is considered as
      
      
        
      
      
      
      
    
      where  is the vector of unknown coefficients by the known nonlinear terms (monomials)  and  denotes element-wise multiplication. The problem usually consists of finding all the entries in , taking into account that the system can be sparse and many of these coefficients can equal zero. The set of monomials  is a-priory known. Nonlinear dynamical models of Type (2) are used quite often, examples can be found in the classical works of E. Lorenz, O. Rossler, and J.C. Sprott []. System reconstruction in terms of PDS has been investigated since the early 1980s by many authors, for example, by J. P. Crutchfield and B. S. McNamara [], who proposed one of the classical definitions of the problem, and J. Cremers and A. Hübler, who used Legendre polynomials as a basis for the derivative function construction []. Later, Taylor-like expansions were introduced by J. L. Breeden et al. []. Many alternatives to PDS models have been proposed, for example, neuronal networks, also including nonlinearities and coefficients found from the experimental data []. A neuronal network can sometimes give more precise results and is more preferable in cases of complex systems. The disadvantage of the neuronal technique is that it yields large-scale and computationally costly models. Neuronal networks and the PDS approach can also be combined, for example, for gene regulatory network identification.
The problem of finding a reliable algorithm for determining coefficients of PDS (2) is difficult, especially in case(s) of noisy and incomplete data. Determining coefficient vector  can be considered as the optimization problem
      
      
        
      
      
      
      
    
      where  is the set of observed derivatives , and  is a set of function values evaluated on . The size of the unknown vector  is equal to  where  is the system dimension and  is a number of available monomials.
The Problem (3) can be treated as a problem of nonlinear regression. Several researchers, for example, H. Iba and D. P. Searson et al. [,] developed specialized versions of genetic algorithms for solving it. The main disadvantage of genetic algorithms is that their run times are, in a general case, much longer in comparison to other approaches. Meanwhile, classical optimization methods such as Newton or quasi-Newton methods are inefficient on this problem because the optimization problem is discontinuous. When one or several coefficients become equal to zero, the nonlinear regression result changes sufficiently.
Fortunately, Problem (3) can also be considered as the linear regression problem. Indeed, consider  as 2-norm, and notice that since the set  is already known, the function values  can be evaluated. Therefore, the least squares method (LSM) can be applied in order to obtain :
      
        
      
      
      
      
    
      where .
The disadvantage of LSM is that Problem (3) is often ill-conditioned since some of the monomials vanish on the given data series and cause the singularity of the matrix . To avoid this, Laubenbacher and Stigler proposed division by Gröbner basis to simplify the initial set of monomials used for regression [], but this approach is not applicable to noisy data series.
Even if monomials do not vanish or the monomial set was somehow simplified to remove all vanishing elementary functions, the matrix  usually has a large condition number which makes Problem (3) difficult to solve. Recently, D. Yogatama and N. Smith considered the application of -regularized LSM to these types of problems, which converge better than the classical version of LSM []. Later, H. Kera and Y. Hasegawa proposed excluding vanishing polynomials using the approximate Buhberger-Müller algorithm and, then, deleting minor terms from regression using a specialized procedure []. The latter algorithm, further referred to as the Kera-Hasegawa (KH) method, underlies the reconstruction procedure developed in this paper.
From a physical point of view, some state variables needed to perform the reconstruction procedure can be unknown. The simplest case is when only system states are known and the derivatives are not, so numerical differentiation should be applied. Howeve, if the system is partially observed, phase space reconstruction must be performed. For example, when a circuit with a memristive element is considered, one of its state variables, magnetic flux or charge in a memristive device, cannot be directly measured in a macroscopic scale []. The problem is more complicated in the domain of physical chemistry where only one variable is usually observed, for example, in Belousov-Zhabotinsky reaction this is Br-ISE electric potential []. Using the well-known Takens’ embedding theorem [] we can reconstruct the discrete-time system phase space using a shifted time series. For continuous-time reconstruction, we can use differentiation or integration. The latter technique is used quite rarely because it can cause the emergence of linear trends in integrated data but we will show how to avoid this and why it is more preferable than differentiation.
The main aim of this paper is to establish an efficient and reliable technique for identifying chaotic circuits, especially, ones with nonlinear elements with unknown parameters. The problem arises from the fact that real circuits even with simple nonlinearities and known parameters need certain parameter optimization for simulation despite the theoretical prediction that the circuit can accurately correspond to the ODE []. When the circuit contains a nonlinear element, the problem becomes more complicated since the relevant model of this element may also be different from the prediction. For example, it has recently been shown that original HP models are irrelevant to the actual resistance switching devices with memory effect []. The relevant model proposed in this work [] contains exponents, hyperbolic trigonometric functions and negative powers.
This paper is organized as follows: First, we describe how PDS can be reconstructed from complete data using the extended Kera-Hasegawa method; second, we explain how these results can be applied to data series lacking some state variables; third, we carry out computational experiments showing the validity of the approach, to be more precise, we reconstruct the Lorenz system from 1-dimensional data series and a 4-dimensional memristive circuit proposed by B. Muthuswamy [] using incomplete 3-dimensional data and we show that the accuracy of numerical integration and differentiation plays an important role in the success of reconstruction, even though the reconstruction method is noise-tolerant; finally, we discuss the limitations of the proposed approach and give a conclusion.
2. Algebraic Method for ODE Reconstruction
In this section, we describe an algebraic approach to the system reconstruction method based on subsequent application of the approximate Buchberger–Möller method, the least-squares method and deleting minor terms method, close to the method proposed by Kera and Hasegawa, but we extend this approach to the Laurent polynomials which can have negative powers of monomials in comparison with ordinary polynomials with nonnegative powers only.
2.1. Polynomial Rings, Ideals, ABM Algorithm
Consider a nonlinear function in Equation (1) as
        
      
        
      
      
      
      
    
        where  is dimension, is the number of elementary monomials , and  are the coefficients of each elementary monomial, and . Each component of Equation (5) represented in the form  is a polynomial. The monomials have general form as
        
      
        
      
      
      
      
    
        where  are the powers of each variable. The polynomial functions
        
      
        
      
      
      
      
    
        form a polynomial ring  over the field , where operations of addition and multiplication are defined as usual. More precisely,  is an Abelian group under addition since it is associative, commutative, has an additive inverse and has zero element ,  is monoid under multiplication since it is associative and has an identity element  which is exactly
        
      
        
      
      
      
      
    
        also this multiplication is distributive. The field over which the ring is defined determines the properties of the coefficients . In this paper, we consider  over a field of real numbers .
When , the functions of Equation (6) form ordinary polynomials, or Taylor-like expressions, e.g., for the two-variable function it can be
        
      
        
      
      
      
      
    
When , the functions for Equation (5) are called Laurent polynomials where powers are conventionally defined within a symmetric interval , but we will not require this property further. The terms in Equation (8) can be ordered in various ways. We will consider a degree-lexicographic order, where we primarily consider the power of the entire monomial defined as
        
      
        
      
      
      
      
    
Second, we consider the powers of variables selected alphabetically organized in ascending order of absolute value of power. This may be exemplified for a set of Laurent monomials of 2 variables and order in  as
        
      
        
      
      
      
      
    
Originally, only positive powers are considered in degree-lexicographic order, so we introduce this equation here by analogy; for discussion, see Section 4.
The order ideal  of the polynomial ring  is a finite set of monomials in  if it is closed under divisors, i.e., satisfying the condition: for each monomial  all monomials  dividing it also belong to . A “border” of the order ideal  in an ordinary polynomial ring is a set of monomials in  defined as
        
      
        
      
      
      
      
    
        where “minus” denotes set subtraction. If the monomial set  is an order ideal in an ordinary polynomial ring, its border is
        
      
        
      
      
      
      
    
In Laurent polynomial ring, a border is defined similarly to Equation (9) but also includes negative powers :
      
        
      
      
      
      
    
The vanishing ideal of a set of data points  is a set of all polynomials vanishing (i.e., equals to zero) on :
      
        
      
      
      
      
    
The vanishing ideal  can be spanned by a finite set of polynomials called the basis of . Let  be an order ideal with a border . Let  be a set of polynomials, and there is an ideal . The set  is an -border prebasis if the polynomials in it have the form
        
      
        
      
      
      
      
    
The set  is an -border basis of  if  generates  and if the residue classes of the terms in  constitute a vector basis of . Several other bases are known, such as Gröbner basis and Macaulay basis. To find a border basis, Buchberger–Möller (BM) algorithms can be used []. Border basis  is called approximate if , where  and  is a small positive number. Modification of BM algorithm, known as the approximate Buchberger–Möller (ABM) algorithm can be used to compute an approximate border basis on a noisy dataset .
We also define the evaluation map , which transforms a given dataset  into a matrix , where
        
      
        
      
      
      
      
    
The ABM algorithm can be described as follows. Suppose, we have a dataset , a tolerance , and set of monomials  in degree-lexicographic order. We should obtain a border basis  and an order ideal . We start with degree  and . Let initially contain all terms of degree 1. Then, the process starts while the order  is not greater than the maximal order defined as  and  is not empty. At the current step, all monomials of degree  have been already processed. We estimate the vector  on the next monomial in  as
        
      
        
      
      
      
      
    
        and, then, remove it from . Let  be the smallest eigenvalue of the matrix 
      
        
      
      
      
      
    
        and if it satisfies the condition
        
      
        
      
      
      
      
    
        then  is appended to  and all the polynomials of order  are added to . A polynomial with coefficients defined by an eigenvector of  corresponding to  is added to .
Application of ABM algorithm to the dataset  and the set of initial monomials  allows obtaining  and ensures that any polynomial
        
      
        
      
      
      
      
    
        where , does not vanish on , even though  can be, strictly speaking, not always an order ideal [].
2.2. KH Algorithm for Polynomial Function Reconstruction
Consider a dataset  which can be represented as observation vectors  for i-th state variable, M is the system dimension. Each data vector has N observation points. Using the order ideal  as a set of basis monomials, one can evaluate them on a given dataset  and then build an interpolation, using derivatives set  which is also given or can be derived from  by the relation :
      
        
      
      
      
      
    
In each dimension, Equation (13) can be reconstructed using the least squares method. Denote  and transform Equation (13) into a linear equation:
      
        
      
      
      
      
    
Equation (14) is usually considered as overdetermined, since the number of estimated points can be arbitrary and, in the presence of noise, a large number of observations are a necessary condition for obtaining a reliable solution. The overdetermined Equation (14) can be solved using the ordinary least squares method, obtaining the minimum
        
      
        
      
      
      
      
    
        by analytical Formula (4). Due to the presence of noise and since it is supposed that the vector  is sparse, -regularization is recommended to obtain better results:
      
        
      
      
      
      
    
        where  is a regularization parameter. Increasing  provides a sparser solution. Problem (16) is also known as a basis pursuit denoising (BPDN) problem, and several special optimization methods have been proposed for solving it, e.g., Nesterov method [,]. If it is assumed that  is dense, -regularization or Tikhonov regularization [] can be applied:
      
        
      
      
      
      
    
        with a known solution in terms of matrices:
      
        
      
      
      
      
    
        where  is an identity matrix [].
Kera and Hasegawa [] proposed removing minor terms from the interpolation polynomial
        
      
        
      
      
      
      
    
        using a special procedure based on an estimation
        
      
        
      
      
      
      
    
For each term in a polynomial  its weight (20) is estimated, and terms with minimal weight are subsequently removed from the interpolation while the error of approximation remains within a given tolerance :
      
        
      
      
      
      
    
Following Kera and Hasegawa, we call this procedure delMinorTerms. It ensures that the solution is sparse enough even if the data is contaminated with noise.
The Kera–Hasegawa (KH) PDS reconstruction method finds a set of interpolating polynomials  in  of type (19), where  is the Cartesian multiplication, satisfying  for a given  and . Firstly, it performs the ABM algorithm to remove monomials from  vanishing on , comprising an order ideal  of . Next, it applies -regularizated LSM to find an interpolation polynomial on  and then runs delMinorTerms to ensure that the interpolated polynomials are sparse while the accuracy criterion Equation (20) is satisfied.
2.3. Phase Space Reconstruction Using Integration Embedding
The KH method can be performed only if  and  are complete, i.e., the dimension M of the reconstructed system corresponds to dimensions of  and . In the general case, we can observe only one or several state variables, and thus we need to reconstruct the entire phase space. The well-known F. Takens’ theorem [] states that having a time-delayed series of a single-variable signal , we can reconstruct M-dimensional phase space due to time delay embedding, and, similarly, N. Packard et al. [] proposed reconstructing M-dimensional phase using  subsequent derivatives
        
      
        
      
      
      
      
    
        due to differential embedding. Recently, J. Lekscha and R. Donner [] have shown that reconstruction is possible in the case of non-uniformly sampled and noisy time series as well, which is an important result for practical use.
Unlike the differential embedding approach, integral embedding did not become popular in scientific literature, despite being more useful in several practical cases.
Lemma 1. 
M-dimensional phase space can be reconstructed due to integral embedding, and a trajectory of the systemreconstructed from time seriescan be written as:
      
        
      
      
      
      
    whereis a residual polynomial whose general form is:
      
        
      
      
      
      
    whereis a natural number.
Proof.  
A system of ODEs in the Cauchy normal form is usually organized so as
        
      
        
      
      
      
      
    
        where the state variable  is related to the variable of the next index through differentiation. It can also be written through the integration
        
      
        
      
      
      
      
    
        where  is a constant depending on the initial condition. Equation (24) can be used for subsequent integration of state variables for reconstructing the trajectory in a full phase space when variables  are known. After K integrations of a time series given by , we obtain
        
      
        
      
      
      
      
    
        from which Equations (22) and (23) follow obviously. □ 
Corollary 1. 
Since the polynomial residual Equation (23) grows very fast ifis large enough, it is undesirable if we consider a computer implementation of the reconstruction algorithm. Otherwise, the accuracy of the data representation may suffer. We propose the following procedure for obtaining. First, we choose zero initial condition to detect and integrateon a given time interval:
      
        
      
      
      
      
    
Then, we estimate its mean value, and supposing, subtract it obtaining the desired solutionAnother way is to de-trend the second integral containingwith a specialized procedure, e.g., detrend function in a MATLAB environment.
The system reconstructed using integration embedding is equivalent to the system reconstructed using differential embedding up to  differentiations. More strictly, this can be formulated as Lemma 2.
Lemma 2. 
Suppose, there is a map,which projectsonto the phase space, can be obtained via subsequent differentiation of the one-dimensional trajectorycorresponding to the state variable. Then, then a dynamical system,exists which projectsonto the phase spacecan be obtained via subsequent integration of the one-dimensional trajectorycorresponding to the state variable.
Proof.  
The following ODE describes a dynamical system  due to the condition that subsequent differentiation of one-dimensional trajectory  generates 
      
        
      
      
      
      
    
Since  the corresponding state variables are equal: . The condition that the projection  of the map  onto the entire phase space can be obtained via subsequent integration of its one-dimensional component  can be expressed in the form of Equation (24). Consequently, the following ODE describes :
      
        
      
      
      
      
    
        where  is a differentiable function. It is related with  in the following way: . □
Corollary 2. 
A dynamical systemwhich can be reconstructed via integration embedding exists if the systemwhich can be reconstructed via differential embedding exists.
The ODE representation of a partially observed system with K observed variables and M-dimensional entirely phase space can be represented in the following way:
      
        
      
      
      
      
    
        where  is a scalar function. Missing derivatives of orders  can be found numerically. Finite-difference-based numerical differentiation is the most common approach, but some other techniques can be used as well, e.g., numerical differentiation based on Legendre polynomials []. If the known state variables are denoted , the system reconstructed via integration reads
        
      
        
      
      
      
      
    
        which is an alternative form of Equation (26) with other variables. □
The number of dimensions  can be selected concerning its physical sense or with one of the several algorithms, e.g., false nearest neighbor (FNN) algorithm [,], or estimating one of the fractal dimensions or entropies [].
The embedded integration approach is more useful in some practical cases due to the following two theorems. Theorem 1 considers a case appearing in many practical applications when the differential embedding approach cannot give a valid reconstruction due to the non-uniqueness of the solution corresponding to the reconstructed trajectory .
Theorem 1. 
A function(t) determined onexists, satisfying the following conditions:
- (1)
- (t) is a function that produces time series u =being a one-dimensional projection of a M-dimensional trajectoryof the dynamical system;
- (2)
- (t) is continuous and satisfies the Lipchitz conditionwhereis the Lipchitz constant;
- (3)
- an M-dimensional trajectorycan be reconstructed from time series u using differentiation which does not satisfy the condition of the uniqueness of the solution;
- (4)
- an M-dimensional trajectorycan be reconstructed from time series u using integration that satisfies the condition of uniqueness of the solution and thus is topologically valid.
Proof.  
  
    
       
      
    
  
  
Consider a continuous function that satisfies the following conditions. First, there exists an infinite set of time points at which the function  and its derivatives equal an arbitrary number :
      
        
      
      
      
      
    
At other time points the function  differs from :
      
        
      
      
      
      
    
Second, for a sufficiently large number :  there exist two time points  and , where the function changes its value relatively to :
      
        
      
      
      
      
    . Then, under this conditions, reconstruction by differentiation yields in a phase space where the point  is a point at which at least two trajectory loops are connected and therefore the uniqueness of the solution is violated: starting from this point and having no apriori information of which branch can be selected we can neither predict the next state of the system nor determine the previous state of the system. Therefore, the trajectory  reconstructed from time series u using differentiation does not satisfy the condition of uniqueness of the solution. Meanwhile, due to Lemma 1 and Lemma 2, it is possible to obtain M-dimensional trajectory  from time series u using integration. Due to Corollary 1 it is possible to select such integration constants  when  and its integrals are not equal to  so it may satisfy the condition of the uniqueness of the solution and thus  is topologically valid. Figure 1 illustrates the phase space geometry. □
 
      
    
    Figure 1.
      Illustration of Theorem 1. The red line corresponds to the double-loop trajectory  obtained by the differential approach, the blue line corresponds to cyclic trajectory  obtained by the integration approach. Here,  and  are phase variables.
  
Time series u =  satisfying conditions 2–4 of the Theorem 1 when are rather common in biology, medicine, genetics and neuroscience.
Theorem 2 is dedicated to noise-amplifying properties of differentiation operators. In the presence of additive white Gaussian noise, when the sampling frequency is high enough and the spectral density of the signal is mostly low-frequency, differentiation tends to decrease the signal-to-noise ratio (SNR), while integration does not.
Theorem 2. 
Denote the Fourier transform of a signalas. Let the power spectrum oflie almost entirely in a frequency interval. Letbe the additive white Gaussian noise (AWGN) contaminating. Then:
- (1)
- after differentiation, the SNR ofwill be smaller than the SNR of;
- (2)
- after integration, the SNR ofwill be at least not smaller than the SNR of.
Proof.  
  
    
       
      
    
  
  
Recall, the signal-to-noise ratio is determined as
        
      
        
      
      
      
      
    
        where  is root-mean-square (RMS) amplitude of the signal with its square equal to
        
      
        
      
      
      
      
    
        where  is time interval, and  is RMS amplitude of the noise.
Denote the sampling frequency as . Denote the Fourier transform of the noise  as . Then, the power spectral density of the noise calculated within frequency interval /2] is
        
      
        
      
      
      
      
    
According to Parseval’s theorem,
        
      
        
      
      
      
      
    
        and the same is true for the signal. Consequently, from Equation (31) and Equations in (32) SNR may be expressed through power spectra as
        
      
        
      
      
      
      
    
Note that AWGN has a uniform power spectral density, i.e.,
        
      
        
      
      
      
      
    
The condition that the power spectrum of  lie almost entirely in a frequency interval  means that its power spectrum  satisfies the condition:
      
        
      
      
      
      
    
        where  is a positive real number close to 1, and  is a frequency up to which almost the whole spectrum is contained, usually, . The gap between  and  is almost occupied by the noise only, see Figure 2a. Consequently, in this case, a formula for SNR can be rewritten as:
      
        
      
      
      
      
    
 
      
    
    Figure 2.
      Illustration of Theorem 2. Blue plot corresponds to the Fourier transform of the signal  and red plot corresponds to the Fourier transform of the noise . (a) Original signal and noise; (b) after differentiation; and (c) after integration.
  
Consider integration and differentiation acting at . Taking Fourier transform, obtain
        
      
        
      
      
      
      
    
        where  is the imaginary unit. The same is true for the noise. Denote , , , . Then, corresponding spectral densities are:
      
        
      
      
      
      
    
From Equation (31) and Equations in (32), SNR of the differentiated signal is:
      
        
      
      
      
      
    
Due to Hölder’s inequality, and the fact that both  and  are nonnegative,
        
      
        
      
      
      
      
    
Since , the  is never greater than the SNR of the original signal under the above made assumptions.
For the integration operator, from Equation (31) and Equations in (32),
        
      
        
      
      
      
      
    
Again, due to Hölder’s inequality,
        
      
        
      
      
      
      
    
Therefore, while the  is not obligatorily greater than the SNR of the original signal, it is not necessarily less than it in contrast to the case of using differentiation. □
Theorem 2 is in good correspondence with the signal processing theory [] from which it is known that differentiation operator relates to a high-pass filter, and amplifies high-frequency noise, while integration operator relates to a low-pass filter.
2.4. Description of the Proposed Reconstruction Technique
The overall proposed technique is described as follows:
First, we consider the time series set  of a partially observed dynamical system and its derivative . Using both integral or differential embedding, preferring the first one due to Theorems 1 and 2, we reconstruct the phase space and the -dimensional phase space trajectory . In this paper we suppose that  is already known. If not, the FNN algorithm or other dimension-finding algorithms can be applied. For the reasons why we do not reject differential embedding, see the Limitations section.
Then, we generate a degree-lexicographically organized set of monomials  of order , where the power of each monomial lies within  and powers of each term in monomial also lies within . From a practical point of view, we assume that  and . The algorithm for generating  is given in pseudocode in Listing 1.
The subroutine generateV returns a special indexing matrix V:
      
        
      
      
      
      
    
The subroutine adds produces a column-wise sum of  as a candidate vector  of monomial terms powers:
      
        
      
      
      
      
    
The monomial  is added to  if it is not already contained in . The order of multiplication is neglected, e.g., the term  is considered similar to the term .
The subroutine shiftForward produces the next state of matrix  1 considering  as a column-type of D-digit number in M-ary numeral system, where each digit has M states defined by the position of the non-zero element in each row, and where the first row is the least significant digit of the number, and  1 denotes summation with the column vector 1 .
Listing 1. A pseudocode for a deglexord routine generating degree-lexicographically ordered Laurent monomials
 
    When the degree-lexicographically ordered set of monomials  is computed, the Kera–Hasegawa method is applied to , , and . The obtained reconstructed PDS corresponds to the original dynamical system if the noise level in the data is sufficiently low and the system is representable in a form of PDS in a given set of coordinates.
3. Experimental Results
In this section, we first illustrate the advantages of using integral embedding using two examples and, then, show how the proposed approach can be used for the reconstruction of Lorenz attractor from a one-dimensional time series and chaotic memristive circuit equations from a three-dimensional time series.
3.1. Illustrations of the Theorems
The following example illustrates the Theorem 1.
Example 1. 
  
    
       
      
    
  
  
Consider the ODE:
      
        
      
      
      
      
    where. The solution of (30) is
      
        
      
      
      
      
    
The functionin (31) satisfies the conditions of the Theorem 1. Its unresolvable point is. Figure 3 shows two variants of 3-dimensional phase spaces, reconstructed via differentiation and integration from time series.
 
      
    
    Figure 3.
      Example of phase space reconstruction. (a) Time series  and its two integrals; (b) comparison of an integration-based approach (blue trajectory) and a differentiation-based approach (red trajectory). The red dot at  denotes a singular point unresolvable in sense of FNN where the uniqueness condition is violated.
  
Example 2 shows how embedded differentiation and integration approaches can be applied to ECG signals. In this example, the conditions of both Theorems are satisfied, i.e., differentiation produces noisy trajectory reconstruction with a singular point, while integration does not have these drawbacks.
Example 2. 
  
    
       
      
    
  
  
Consider the ECG signal from ECG-ID Database [,], as shown in Figure 4a.
 
      
    
    Figure 4.
      Example of the phase space reconstruction. (a) ECG time series ; (b) comparison of the integration-based approach (left) and the differentiation-based approach (right).
  
After applying the embedded differentiation approach, we obtain a trajectory that does not satisfy the uniqueness criterion and is contaminated with noise. After applying embedded integration approaches, we obtain a rather a smooth trajectory with no singular points.
3.2. Reconstruction of Lorenz Attractor
Consider the classical Lorenz attractor
        
      
        
      
      
      
      
    
        with a standard parameter set:  []. Suppose, only one variable  can be observed, so we should transform the Lorenz system to obtain a model in new coordinates:
      
        
      
      
      
      
    
		Equation (37) cannot be reconstructed into ordinary polynomial functions due to the negative powers of variable , but it is possible to reconstruct it as the Laurent polynomial function. The application of the proposed algorithm can be done using both differential and integration approaches.
The deglexord procedure (see Listing 1) generates a set of  monomials or degrees . To ensure that the system of equations is overdetermined, we should take not less than 51 points in . We are never sure that any term will vanish on  and will be removed by the ABM algorithm. An example of the reconstructed system is:
      
        
      
      
      
      
    
To reach a high quality of reconstruction, we generate data using the function ode113 with parameters RelTol  and AbsTol  and constant stepsize and simulation time  in MATLAB 2019b environment (Campus license No. 40502181, ETU “LETI”) and use the 4th order integration formula. Note that Equation (37) can be reconstructed from the data obtained from the original Lorenz system using the time series corresponding to the variable  and its two integrals.
Two projections of the attractor Equation (37) and its reconstructed version are given in Figure 5. The initial conditions were , the parameter .
 
      
    
    Figure 5.
      The trajectory of the modified Lorenz system Equation (37) is projected onto (a) x-y plane and (b) x-z plane. The green dots denote 95 data points, the blue line depicts the original trajectory, and the orange line depicts a trajectory of the reconstructed system.
  
Truncation and round-off errors can be considered as a sort of noise [] which makes it possible to compare differential embedding and integral embedding in terms of sensitivity to the reconstruction error induced by the stepsize. The results are given in Figure 6.
 
      
    
    Figure 6.
      Results of the reconstruction of system Equation (37) using integration and differentiation of orders 2 and 4 with respect to the stepsize. (a) mean error  by all 3 dimensions; (b) overall number of terms in the reconstruction, where  corresponds to the original system (37). Square markers correspond to the differentiation-based approach; crosses correspond to the integration-based approach. We tested 15 stepsizes, 10 experiments were carried out with each stepsize.
  
In Figure 6, we denote the phase space reconstruction method based on differentiation of the 2nd order of accuracy as DiffOrd2 (central differences) and of the 4th order of accuracy as DiffOrd4 (4 order finite differences) []. The method based on of 2nd order integration is denoted as IntOrd2 (trapezoidal rule) and the method of 4th order of accuracy is denoted as IntOrd4 (Simpson’s rule). Two plots summarize the results on mean error  across three state variables and overall number of terms .
We see from both plots that for small stepsizes, the integration based reconstruction method performs well while the differentiation approach suffers from numerical noise following with the proof of Theorem 2. The higher the sampling frequency is, the more noise is added into the data and the algorithm is more likely to fail. However, large stepsizes are more likely to corrupt the integration-based approach, but in this case this happens at when both methods give inaccurate results even on 4th order. The reconstruction techniques based on second-order formulae became inaccurate at , the fourth-order based formulae became inaccurate at . By “inaccurate” we mean that Equation (37) was reconstructed with some missing or excessive terms.
3.3. Reconstruction of Muthuswamy’s Memristive Circuit
Replacing the nonlinear negative resistor (Chua diode) in the well-known Chua’s circuit with the flux-controlled memristor one obtains a circuit able to exhibit the chaotic behavior []. The circuit proposed by B. Muthuswamy includes five elements (Figure 7): a linear passive inductor, two linear passive capacitors, a linear passive resistor and a nonlinear active memristor.
 
      
    
    Figure 7.
      Schematics of the five-element chaotic circuit [].
  
After scaling and transformation, we obtain the following system of ODEs:
      
        
      
      
      
      
    
Parameters for chaotic behavior are ; ;;;;;; We simulated the system by the function ode113 with parameters RelTol  and AbsTol  and constant stepsize  and final time . In this experiment we used data obtained from the original Equation (39) and reconstructed the state variable  by integrating variable . This is motivated by two facts: first, variable  stands for the magnetic flux between the terminals of the memristive device [], and therefore cannot be measured on a macroscopic scale, so it is assumed to be unobservable. Second, integrating  is very natural to the system (39) due to the relation  and therefore, needs no complicated transformations.
The degrees of terms in the monomials were placed within the interval , the tolerance parameter of AMB algorithm was . The reconstructed equations are
        
      
        
      
      
      
      
    
Note, that the Equations in (40) are a rescaled version of the Equations in (39) when  instead of  in the original equation.
Two projections of the attractor of original and reconstructed systems are given in Figure 8. The initial conditions were . We compared three integration methods by changing the tolerance parameter  in the delMinorTerms algorithm in the range . Figure 9 depicts the result of the experiment. Using higher-order integration formulae allows for the achievement of more precise and more compact solutions. This experiment shows the difficulty of choosing : its low values force the algorithm to generate equations with many excessive terms making the least-squares solution more precise. But, as one can see in Figure 9, the accuracy of the proper reconstruction is rather high due to high absolute values of the state variables (see Figure 6) and thus truncation and numerical errors in the reconstructed data for  are large as well.
 
      
    
    Figure 8.
      The trajectory of the system Equations in (40) projected onto (a) x-y plane and (b) x-z plane. The green Dots denote 53 data points, the blue line depicts the original trajectory, and the orange line depicts a trajectory of the reconstructed system scaled to fit the original trajectory. The equations were reconstructed with high accuracy.
  
 
      
    
    Figure 9.
      Results of the reconstruction of system Equations in (39) using integration of orders 1, 2 and 4 (Euler, Trapezoidal and Simpson’s respectively) in accordance to the delMinorTerms tolerance. (a) Mean error  by 4 dimensions; (b) overall number of terms in the reconstruction, where  corresponds to the original system Equations in (39). We tested 15 values of , 10 experiments were carried out with each value.
  
To summarize, the experiment shows that higher-order integration is more preferable and value  should be adjusted with respect to the problem properties.
4. Discussion and Conclusions
Two main findings are highlighted in this paper. First, we show that the integral embedding approach has several advantages over the differential approach often used in cases of incompletely observed systems. Second, we extend the Kera–Hasegawa method to Laurent polynomials and show by example its performance.
Still, some limitations of this method exist. The first, and the most prominent one being, is that not each system can be represented as a PDS, and furthermore, there are some PDSs that do not allow transformation to the form of Equation (22) or Equation (24) or their combinations.
Consider the following system with an unobservable variable :
      
        
      
      
      
      
    
      where  are parameters. The natural way to reconstruct Equations (41) in this case is to transform it into
      
      
        
      
      
      
      
    
      where a new variable  is reconstructed from the observable variable  through differentiation. Equation (42) cannot be represented in terms of polynomial functions and, moreover, needs a special technique to resolve the ambiguity in  sign appearing when  variable is expressed from the second Equation of (41).
The second limitation of the proposed reconstruction method can be exemplified by Muthuswamy’s memristive circuit. In many cases, the algorithm found a PDS different from Equations in (40), but which is also correct on limited data series  and has a stable attractor similar to the attractor of the system Equations in (40):
      
        
      
      
      
      
    
Moreover, the third Equation in (43) is parametric, so infinitesimally many variants of it exist. So, if similar data can be produced by several PDSs, the method may find any one of them.
The third limitation of the proposed reconstruction method was found in both experiments with the Lorenz system and Muthuswamy’s memristive circuit, i.e., it is sensitive to algorithm parameters, such as stepsize, the order of accuracy of the reconstruction method, its type (differentiation or integration), and tolerance of delMinorTerms algorithm . Obviously, sensitivity to the AMB algorithm tolerance  also takes place, but we did not investigate its influence.
The fourth limitation is that the dimension of search grows dramatically when the number of terms or system dimension grows. For example, four-dimensional PDS in complete Laurent polynomials containing degrees in  can be constructed out of 471 unique monomials, five-dimensional PDS out of 1281 unique monomials, and six-dimesional PDS out of 3067 unique monomials. Even if ABM algorithm sufficiently reduces these numbers, it is likely to obtain very poorly conditioned large-scale regression problems, and it is not obvious whether the algorithm remains useful in these cases. This question needs further investigation.
Laurent monomial ordering, delgexord, and BM algorithm for Laurent monomials also need the development of a stricter theoretical basement, as in the case of well-established ordinary monomials []. Moreover, from Equation (4) it is clear that the reconstruction approach based on the LSM can also handle not only PDSs but functions containing other elementary nonlinearities as well (e.g., trigonometric functions). This also needs further development.
Author Contributions
Conceptualization, D.B.; data curation, E.G.N.; formal analysis, E.G.N.; funding acquisition, D.B.; investigation, A.K. and A.T.; methodology, A.K.; project administration, A.K. and D.B.; resources, A.T.; software, A.K. and A.T.; supervision, E.G.N. and D.B.; validation, D.B.; visualization, A.T.; writing—original draft, A.K.; writing—review & editing, E.G.N. and D.B. All authors have read and agreed to the published version of the manuscript.
Funding
The reported study was supported by RFBR, research project no. 19-07-00496.
Acknowledgments
The authors are grateful to Valerii Ostrovskii for preparing Figure 7 and fruitful comments on the introduction.
Conflicts of Interest
The authors declare no conflict of interest.
References
- Laubenbacher, R.; Stigler, B.A. Computational algebra approach to the reverse engineering of gene regulatory networks. J. Theor. Biol. 2004, 229, 523–537. [Google Scholar] [CrossRef] [PubMed]
- Letellier, C.; Maquet, J.; Labro, H.; Sceller, L.L.; Gouesbet, G.; Argoul, F.; Arnéodo, A. Analyzing Chaotic Behavior in a Belousov− Zhabotinskyi Reaction by Using a Global Vector Field Reconstruction. J. Phys. Chem. A 1998, 102, 10265–10273. [Google Scholar] [CrossRef]
- Akman, O.E.; Watterson, S.; Parton, A.; Binns, N.; Millar, A.J.; Ghazal, P. Digital clocks: Simple Boolean models can quantitatively describe circadian systems. J. R. Soc. Interface 2012, 9, 2365. [Google Scholar] [CrossRef] [PubMed]
- Gerhard, F.; Kispersky, T.; Gutierrez, G.J.; Marder, E.; Kramer, M.; Eden, U. Successful reconstruction of a physiological circuit with known connectivity from spiking activity alone. PLoS Comput. Biol. 2013, 9, e1003138. [Google Scholar] [CrossRef]
- Bansal, M.; Della, G.G.; Bernardo, D.D. Inference of gene regulatory networks and compound mode of action from time course gene expression profiles. Bioinformatics 2006, 22, 815. [Google Scholar] [CrossRef]
- Li, J.; Zhang, X.S. An optimization model for gene regulatory network reconstruction with known biological information. Optim. Syst. Biol. 2007, 7, 35. [Google Scholar]
- Sprott, J.C. Some simple chaotic flows. Phys. Rev. E 1994, 50, R647. [Google Scholar] [CrossRef]
- Crutchfield, J.P.; McNamara, B.S. Equations of motion from a data series. Complex Syst. 1987, 1, 417–452. [Google Scholar]
- Cremers, J.; Hübler, A. Construction of differential equations from experimental data. Z. Für Nat. A 1987, 42, 797–802. [Google Scholar]
- Breeden, J.L.; Dinkelacker, F.; Hübler, A. Noise in the modeling and control of dynamical systems. Phys. Rev. A 1990, 42, 5827–5836. [Google Scholar] [CrossRef]
- Aguirre, L.A.; Letellier, C. Modeling nonlinear dynamics and chaos: A review. Math. Probl. Eng. 2009, 2009, 238960. [Google Scholar] [CrossRef]
- Iba, H. Inference of differential equation models by genetic programming. Inf. Sci. 2008, 178, 4453. [Google Scholar] [CrossRef]
- Searson, D.P.; Leahy, D.E.; Willis, M.J. GPTIPS: An open source genetic programming toolbox for multigene symbolic regression. In Proceedings of the International multiconference of engineers and computer scientists, Hong Kong, 17–19 March 2010; Volume 1, pp. 77–80. [Google Scholar]
- Yogatama, D.; Smith, N. Making the most of bag of words: Sentence regularization with alternating direction method of multipliers. In Proceedings of the 31st International Conference on Machine Learning, Beijing, China, 21–26 June 2014; Volume 32, p. 656. [Google Scholar]
- Kera, H.; Hasegawa, Y. Noise-tolerant algebraic method for reconstruction of nonlinear dynamical systems. Nonlinear Dyn. 2016, 85, 675–692. [Google Scholar] [CrossRef]
- Linn, E.; Siemon, A.; Waser, R.; Menzel, S. Applicability of well-established memristive models for simulations of resistive switching devices. IEEE Trans. Circuits Syst. I Regul. Pap. 2014, 61, 2402–2410. [Google Scholar] [CrossRef]
- Abarbanel, H.D.; Brown, R.; Sidorowich, J.J.; Tsimring, L.S. The analysis of observed chaotic data in physical systems. Rev. Mod. Phys. 1993, 65, 1331. [Google Scholar] [CrossRef]
- Karimov, T.; Butusov, D.; Andreev, V.; Karimov, A.; Tutueva, A. Accurate synchronization of digital and analog chaotic systems by parameters re-identification. Electronics 2018, 7, 123. [Google Scholar] [CrossRef]
- Muthuswamy, B. Implementing memristor based chaotic circuit. Int. J. Bifurc. Chaos 2010, 20, 1335–1350. [Google Scholar] [CrossRef]
- Nesterov, Y. A Method of Solving a Convex Programming Problem with Convergence Rate O(1/k2). In Sov. Math. Dokl; USSR: Moscow, Russia, 1983; Volume 27, pp. 372–376. [Google Scholar]
- Shi, Y.; Zhu, X.X.; Yin, W.; Bamler, R. A fast and accurate basis pursuit denoising algorithm with application to super-resolving tomographic SAR. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6148–6158. [Google Scholar] [CrossRef]
- Tikhonov, A.N.; Arsenin, V.Y. Solution of Ill-posed Problems; Winston & Sons: Hoboken, NJ, USA, 1977. [Google Scholar]
- Takens, F. Detecting Strange Attractors in Turbulence. Lecture Notes in Mathematics; Springer Science and Business Media: Berlin/Heidelberg, Germany, 1980; pp. 366–381. [Google Scholar]
- Packard, N.H.; Crutchfield, J.P.; Farmer, J.D.; Shaw, R.S. Geometry from a time series. Phys. Rev. Lett. 1980, 45, 712–716. [Google Scholar] [CrossRef]
- Lekscha, J.; Donner, R.V. Phase space reconstruction for non-uniformly sampled noisy time series. Chaos: Interdiscip. J. Nonlinear Sci. 2018, 28, 085702. [Google Scholar] [CrossRef]
- Kennel, M.B.; Brown, R.; Abarbanel, H.D. Determining embedding dimension for phase-space reconstruction using a geometrical construction. Phys. Rev. A 1992, 45, 3403. [Google Scholar] [CrossRef] [PubMed]
- Rhodes, C.; Morari, M. The false nearest neighbors algorithm: An overview. Comput. Chem. Eng. 1997, 21, S1149–S1154. [Google Scholar] [CrossRef]
- Bradley, E.; Kantz, H. Nonlinear time-series analysis revisited. Chaos: Interdiscip. J. Nonlinear Sci. 2015, 25, 097610. [Google Scholar] [CrossRef] [PubMed]
- Oppenheim, A.V.; Buck, J.R.; Schafer, R.W. Discrete-Time Signal Processing; Prentice Hall: Upper Saddle River, NJ, USA, 2001; Volume 2. [Google Scholar]
- Goldberger, A.L.; Amaral, L.A.N.; Glass, L.; Hausdorff, J.M.; Ivanov, P.Ch.; Mark, R.G.; Mietus, J.E.; Moody, G.B.; Peng, C.-K.; Stanley, H.E. PhysioBank, PhysioToolkit, and PhysioNet: Components of a New Research Resource for Complex Physiologic Signals. Circulation 2003, 101, e215–e220. [Google Scholar] [CrossRef]
- Lugovaya, T.S. Biometric Human Identification Based on Electrocardiogram. Master’s Thesis, Faculty of Computing Technologies and Informatics, Electrotechnical University “LETI”, Saint-Petersburg, Russia, June 2005. [Google Scholar]
- Leonov, G.A.; Pogromsky, A.Y.; Starkov, K.E. The dimension formula for the Lorenz attractor. Phys. Lett. A 2011, 375, 1179–1182. [Google Scholar] [CrossRef]
- Hinamoto, T.; Lu, W.S. Digital Filter Design and Realization. River Publ. 2017, 384. [Google Scholar]
- Jordan, C.; Jordán, K. Calculus of finite differences. Am. Math. Soc. 1965, 33, 15. [Google Scholar] [CrossRef]
- Rahkooy, H.; Zafeirakopoulos, Z. Using resultants for inductive Gröbner bases computation. Acm Comm. Comput. Algebra 2011, 45, 135–136. [Google Scholar] [CrossRef]
© 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
