You are currently viewing a new version of our website. To view the old version click .
Mathematics
  • Article
  • Open Access

7 August 2023

Investigation of the Product of Random Matrices and Related Evolution Models

,
and
1
Atomic Molecular and Optical Physics Research Group, Science and Technology Advanced Institute, Van Lang University, Ho Chi Minh City 700000, Vietnam
2
Faculty of Applied Technology, School of Technology, Van Lang University, Ho Chi Minh City 700000, Vietnam
3
Auriga Inc., 400 TradeCenter, Ste 5900, Woburn, MA 01801, USA
4
A.I. Alikhanyan National Science Laboratory (Yerevan Physics Institute) Foundation, 2 Alikhanian Brothers St., Yerevan 375036, Armenia
This article belongs to the Section E3: Mathematical Biology

Abstract

In this paper, we study the phase structure of the product of D * D order matrices. In each round, we randomly choose a matrix from a finite set of d matrices and multiply it with the product from the previous round. Initially, we derived a functional equation for the case of matrices with real eigenvalues and correlated choice of matrices, which led to the identification of several phases. Subsequently, we explored the case of uncorrelated choice of matrices and derived a simpler functional equation, again identifying multiple phases. In our investigation, we observed a phase with a smooth distribution in steady-state and phases with singularities. For the general case of D-dimensional matrices, we derived a formula for the phase transition point. Additionally, we solved a related evolution model. Moreover, we examined the relaxation dynamics of the considered models. In both the smooth phase and the phase with singularities, the relaxation is exponential. The superiority of relaxation in the smooth phase depends on the specific case.
MSC:
92D15; 60B20

1. Introduction

The investigation of the product of Random Matrices is a significant area in modern statistical physics concerning disordered systems and data science [1,2,3,4,5,6,7,8]. This field of mathematics is well-established and has connections to quantum condensed matter, Hidden Markov models (an essential tool in data science) [9,10,11,12,13], genetic network population dynamics [14], evolution in dynamically changing environments [15,16,17,18,19,20,21], and quantum resetting [22]. Our recent study focused on the phase structure of the product of correlated random matrices with D = 2 [23]. We discovered a transition similar to the localized-delocalized transition previously observed in evolution models [14,24]. The choice of the matrix in the current round can either be correlated, depending on the matrix chosen in the previous round of iteration, or uncorrelated.
During the investigation of the product of random matrices, a key objective is to calculate the maximal Lyapunov index. This can be achieved using the formalism described in [12,13]: by introducing a vector and multiplying it at each round by a randomly chosen matrix type for that round. This process generates an n-th vector at the n-th round. The norm of the n-th vector grows exponentially with n, and the coefficient of this exponential growth corresponds to the maximal Lyapunov index.
In the case of correlated matrix type choice, a functional equation has recently been derived to describe the probability distribution for the vector mentioned earlier. A similar equation has also been derived for the Hidden Markov model. Further investigation of this equation revealed the identification of a phase transition in the steady-state distribution for the rotating reference vector. This transition is directly related to the phase transition found in [14]. In [14], transitions occur between different types of environments in infinitesimal periods of time, resulting in a system of ordinary differential equations (ODEs) instead of a functional equation.
In our recent investigation, we focused on the product of correlated d = 2 matrices of order D = 2, with correlation in the type choice. In Section 2, we identify the phase structure for the case of uncorrelated matrices. Additionally, we derive formulas to determine the phase transition points in the case of high-order matrices. In Section 3, we delve into the related evolution on a fluctuating fitness landscape, including its relaxation dynamics.

2. The Product of Matrices

2.1. The Uncorrelated Matrices

We have two matrices, M 1 A and M 2 B . We have a probability π l for the choice of the matrix type x. We choose some initial vector X , then multiply it by the randomly chosen matrix, obtaining the updated version of the vector X . We define the distribution P ( n , X ) at the n-th round, and d X P ( n , X ) = 1 . We have a vector X with a distribution P ( n , X ) at the n-th round, then at the next step we obtain
p ( n + 1 , X ) = l π i P ( n , M l 1 X ) | det ( M l ) 1 |
For the case of 2 × 2 matrices, we can parameterize the two-dimensional vector via angular parameter α .
( x 1 , x 2 ) = d ( cos ( α ) , sin ( α ) )
where d is a norm of the vector.
If we have a vector with angle α after multiplying it by matrices M l , then the matrix A will move it to the point f 1 ( α ) , where f ( 1 ) is the inverse function of f. It means that matrix A acts as a transformation that takes the vector from the angle α to its inverse according to the function f.
α > f l 1 ( α )
where
f 1 1 ( α ) = arccos ( ( A 11 cos ( α ) + A 21 sin ( α ) ) ( A 11 cos ( α ) + A 21 sin ( α ) ) 2 + ( A 12 cos ( α ) + A 22 sin ( α ) ) 2 ) f 2 1 ( α ) = arccos ( ( B 11 cos ( α ) + B 21 sin ( α ) ) ( B 11 cos ( α ) + B 21 sin ( α ) ) 2 + ( B 12 cos ( α ) + B 22 sin ( α ) ) 2 )
then the functional Equation (1) transforms to
P ( n + 1 , α ) = l | f l ( α ) | P ( f l ( α ) ) π l
Equation (4) is derived for general assumptions.
For the numerical investigation (see the Appendix B for the details of the numerical algorithm) it is convenient to look at the probability distribution with fixed types P l ( n , α ) ,
P l ( n + 1 , α ) = f l ( α ) | P ( n , f l ( α ) ) π l , P ( n , α ) = P 1 ( n , α ) + P 2 ( n , α )
For the real eigenvalues case, we assume a steady state solution P ( α ) = P 1 ( α ) + P 2 ( α ) for Equation (4)
P ( α ) = l | f l ( α ) | P ( f l ( α ) ) π l
First, we need to identify the support [ α 1 , α 2 ] where the solution of Equation (4) is defined. We define α 1 , α 2 as the angles of the eigenvectors that are associated with the maximal eigenvalues:
α 1 = f 1 ( α 1 ) , α 2 = f 2 ( α 2 )
We assume that the angles [ α 1 , α 2 ] are the special points for the functions functions P 1 , P 2 .
For the steady state case of Equation(4), we expand the function f l ( α ) near the point α l
f 1 ( α ) α 1 + k 1 ( α α 1 ) , f 2 ( α ) α 2 + k 2 ( α α 2 )
We look an Ansatz
ρ 1 ( α ) = a 1 | α α 1 | 1 + n 1 , ρ 2 ( α ) = a 2 ( | α α 2 | 1 + n 2
near the points α 1 and α 2 .
Let us define n 1 , n 2 . Looking the solutions with the scaling behavior near the corresponding critical points
P 1 ( α ) ( n 1 + 1 ) ( α α 1 ) n 1 , P 2 ( α ) ( n 2 + 1 ) ( α α 2 ) n 2
When the transition rates tend to 0, then our Ansatz Equation (10) transforms to δ ( α α l )
We obtain from Equation (7)
1 = π 1 ( k 1 ) n 1 , 1 = π 2 ( k 2 ) n 2
Thus
n 1 = log ( π 1 ) log ( k 1 ) , n 2 = log ( π 2 ) log ( k 2 )
The critical cases are when n 1 = 1 .
π 1 k 1 = 1
and
π 2 k 2 = 1
We derived Equations (13) and (14) by assuming a scaling Ansatz as given in Equation (10). Although the scaling Ansatz is not observed in all situations according to our numerical simulations, Equations (13) and (14) are still well-confirmed and hold true.
For simplicity, we use the following representation for the matrices
A = | α 1 > | α 1 > * + λ 1 | π 2 + α 1 > | π 2 + α 1 > * , B = | α 2 > | α 2 > * + λ 2 | π 2 + α 2 > | π 2 + α 2 > *
where | > * means a transposition of the vector. We denote also k 1 = 1 / λ 1 , k 2 = 1 / λ 2
Then we have
f l 1 ( α ) = arctan sin ( α l ) cos ( α α l ) + λ l cos ( α l ) sin ( α α l ) cos ( α l ) cos ( α α l ) λ l sin ( α l ) sin ( α α l )
Figure 1 and Figure 2 illustrate the phase transition. In our investigation, we choose both matrices with equal probabilities.
Figure 1. The distributions p 1 ( x ) (the smooth line) and p 2 ( x ) (the dashed line) versus x α / ( 2 π ) , see Equation (5). α 1 = 0 , α 2 = 0.25 π , see Equation (15). (a) The case without singularities k 1 = 1.5 ,   k 2 = 1.5 , see Equation (15). The right border of the smooth line is f 1 ( x 2 ) / ( 2 π ) . The right border of the dashed line is f 2 ( x 1 ) / ( 2 π ) . The left border of the dashed line is close to the point f 2 ( x 2 ) / ( 2 π ) . (b) A singular distribution. k 1 = 4 ,   k 2 = 4 . We identify the left border of the smooth line with x 0 , and the right border of the dashed line with x 1 . The right border of the smooth line with f 1 ( x 2 ) / ( 2 π ) , the right border of the dashed line with f 2 ( x 1 ) / ( 2 π ) .
Figure 2. The distributions p 1 ( x ) (the smooth line) and p 2 ( x ) (the dashed line) versus x α / ( 2 π ) see Equation (5). α 1 = 0 ,   α 2 = 0.25 π , see Equation (15). (a) k 1 = 1.5 ,   k 2 = 4 .The right border of the smooth line with a singularity is at the point f 1 ( x 2 ) / ( 2 π ) = 0.093 The left border of the smooth line is x 1 , and the right border of the dashed line is x 2 . (b) k 1 = 4 ,   k 2 = 1.5 , The singularity point of the dashed line is f 2 ( x 1 ) / ( 2 π ) = 0.031 . The left border of the smooth line is x 1 , and the right border of the dashed line is x 2 .

2.2. The Relaxation Dynamics

During our investigation, we also studied the relaxation dynamics and confirmed that there is an exponential fast relaxation to the steady state in both phases. In these phases, the square distance to the steady state decreases approximately as e x p [ 2 r t ] . Although we were not able to derive a general expression for the relaxation dynamics in all cases, we observed that when we choose λ 1 = λ 2 = λ , it is reasonable to consider r as the gap between the maximal and next eigenvalues.
r = 1 λ
where the maximal eigenvalue is 1 (the probability balance condition) and r is the gap between the maximal and the next to the maximal eigenvalues, λ m a x [ λ 1 , λ 2 ] see Equation (15). Figure 3 gives the results of numerics for the r.
Figure 3. The relaxation dynamics for the uncorrelated case, the r ( k ) for the model with k 1 = k 2 = k , α 1 = 0 ,   α 2 = π / 4 . The smooth line is our theoretical estimate, r = 1 1 / k the dashed line is the result by numerics.
We also investigated the relaxation dynamics in case of correlated choice of the matrices: at the next round, we have the current choice of the matrix type with a probability q and an alternative choice with a probability 1 q . Figure 4 gives the results of the numerics for the r.
Figure 4. The numerical results for the relaxation dynamics, the r ( k ) for the model with k 0 = k 1 = k ,   α 1 = 0 ,   α 2 = π / 4 . The system holds with probability 0.4 the current landscape and, with the probability 0.6 , moves to the alternative one. The change of distribution is at k = 2.5 .

2.3. The Identification of the Phase Transition in Higher Dimensions

Consider the case when there are d > 2 order matrices. For the sake of simplicity, we take the following representation for the matrices
A = | V 1 > | V 1 > * λ 1 + λ 11 | V 11 > | V 11 | > * + λ 12 | V 12 > | V 12 > * B = | V 2 > | V 2 > * λ 2 + λ 21 | V 21 > | V 21 > * + λ 22 | V 22 > | V 22 > *
The Matrix A has a maximal eigenvalue λ 1 , the next eigenvalues are λ 11 , λ 12 , and λ 1 > λ 11 > λ 12 . The Matrix B has a maximal eigenvalue λ 2 , the next eigenvalues are λ 21 , λ 22 and λ 2 > λ 21 > λ 22 .
If we take a vector | V 1 > + δ | V 12 > , then multiplying by A we obtain ∼ | V 1 > + δ λ 12 λ 1 | V 12 > , thus the distance to the V 1 decreased λ 1 / λ 12 times. We have that there is
λ 1 λ 12
times attraction. Thus for d = 2 case we identify f 1 ( ff ) = λ 1 / λ 12 .
Exact critical point corresponds to the limit
π | λ 1 λ 12 | = 1
When
π | λ 1 λ 12 | > 1
The steady-state distribution is expected to be focused on an eigenvector related to the maximal eigenvalue. We derived Equations (20) and (21) without making any scaling assumption. Our derivation is on a qualitative level, in contrast to the strict derivation of Equations (13) and (14) after using the Ansatz (10). Currently, the state of the system is described by a d-dimensional unit vector v .
In the Figure 5 we take V 1 ( 1 , 0 , 0 ) , V 11 = ( 0 , 1 , 0 ) , V 12 = ( 0 , 0 , 1 ) , V 2 ( 1 , 1 , 1 ) , V 21 ( 51 , 1 , 0 ) , V 22 ( 1 , 1 , 2 ) . Our numeric confirms the existence of phase transition according to Equations (21) and (22). We calculate numerically the distribution via x = ( 1 + v . V 1 ) / 2 . for the case with the A matrix choice, and via x = ( 1 + v . V 2 ) / 2 . for the alternative case.
Figure 5. We choose the matrix A via Equation (15) x 1 = 0 (the related distribution is given by smooth line), then the matrix B via Equation (22) (the related distribution is given by dashed line. (a) k 1 = 1.5 ,   c = 1 ) . We observe a lot of singularities in the distributions. (b) k 1 = 4 ,   c = 1 . We observe a smooth distribution.
In the d = 2 case, we observed non-ergodicity with several separated domains. However, in the current case, we do not observe such non-ergodic behavior with separated domains. This difference in behavior is attributed to the one-dimensional character of the functional equation in the current scenario. In summary, the change in the dimensionality of the functional equation has an impact on the presence or absence of non-ergodicity and separated domains in the system.
Although we are unable to derive a steady-state distribution for α similar to the one described by Equation (5), we can still measure some distribution numerically during the iteration process. By doing so, we can identify the phase transition by observing the singularities of the distribution. Even without an analytical expression for the steady-state distribution, numerical methods can provide valuable insights into the system’s behavior and help identify important transitions and patterns.

2.4. Complex Eigenvalues Case

We assume that the support of the steady state distribution can be larger than [ 0 , 2 π ] . We take the following matrix
B = 1 c c 1
and
f 3 1 ( α ) = arccos ( cos ( α ) + c sin ( α ) 1 + c 2 )
In the case where one matrix has a real eigenvalue α 1 , while another matrix has complex eigenvalues, we observe a transition in both distributions, as described by Equation (13), as shown in Figure 6.
Figure 6. The three-dimensional case. Equation (18). We take 10 random initial vectors then make 100,000 iterations. The smooth line is for p 1 ( x ) , the dashes line is for p 2 ( x ) . λ 1 = 1 ,   λ 2 = 1 . (a) λ 11 = 1 / 3 ,   λ 12 = 1 / 6 . λ 21 = 1 / 3 ,   λ 22 = 1 / 6 . The probability distribution p versus x = ( 1 + v . V 1 ) / 2 . Both distributions correspond to the localization case. (b) λ 11 = 1 / 1.5 ,   λ 12 = 1 / 1.5 . λ 21 = 1 / 3 ,   λ 22 = 1 / 2 . The probability distribution p versus x = ( 1 + v . V 2 ) / 2 . There is no singularity in p 1 ( x ) . .
The presence of complex eigenvalues in one matrix introduces additional dynamics and can lead to interesting transitions in the distributions. The interplay between real and complex eigenvalues influences the behavior of the system and can result in distinct phase transitions.

4. Conclusions

In this current article, our investigation focused on the product of random matrices and its related evolution model. The study aimed to understand the behavior and dynamics of the system under consideration. In this article, we formally define the “localized phase” as a phase characterized by a singularity in the steady-state distribution and the “de-localized phase” as a phase with a smooth distribution. This definition provides a rigorous classification of the phases based on the behavior of the steady-state distributions. Upon analysis, we identify that the de-localized phase exhibits a wide distribution, indicating a broader range of possible outcomes for the system. In our investigation, we numerically studied the product of uncorrelated discrete random matrices and identified a phase transition between localized and de-localized phases. The derived master Equation (3) for this case was found to be simpler than the one for correlated random matrices. By considering matrices with real eigenvalues and complex eigenvalues separately, we were able to obtain analytical formulas for the transition points, including a formula (Equation (19)) for the appearance of singularities in the phase space.
For the case of 2 × 2 matrices, we observed a series of transitions, as shown in Figure 1. However, we did not find any indication that different phases in the statics have different expressions for the Lyapunov indices. Regarding the relaxation dynamics, we investigated both the correlated and uncorrelated cases. We found that there is no transition in the dynamics, but relaxation is faster in the singular phase. Despite obtaining answers to some questions related to the random matrix problem, certain issues remain unsolved. Two significant problems are determining the conditions under which relaxation in the localized phase is faster than in the de-localized phase and identifying cases of explicit non-ergodic behavior in the steady-state distribution. Furthermore, we investigated the related evolution model, which resembles the case of random matrix products when matrices have real eigenvalues. In this model, we verified that the relaxation dynamics are somehow faster in the non-singular phase. Although our research focused on the Master equation, we acknowledge that the existence of the phase transition may have practical implications for related models, such as Hidden Markov Models. In conclusion, our work contributes to a better understanding of the random matrix problem, but some challenges remain unresolved. We hope that future research will address these difficult problems and provide further insights into the practical implications of the phase transition in related models.
For Hidden Markov Models (HMMs), in the functional equations such as Equation (5), the transition probabilities between states depend on the current state of the system. In HMMs, the system’s state is hidden and can only be inferred from the observed output (emission) probabilities associated with each state. The transition probabilities in the functional equations determine the probabilities of transitioning between different hidden states based on the current observed state. The existence of phase transition here could have a serious impact. Indeed, the view that learning is especially successful near the phase transition point is supported by a vast amount of research and experience over the last decades. In various complex systems, such as statistical physics, machine learning, neural networks, and optimization algorithms, it has been observed that the system’s performance improves significantly when operating near a phase transition point [28].
The Kalman filter [29] is another interesting model related to the product of correlated random matrices. The Kalman filter is an essential tool in control theory and estimation, used for state estimation in dynamic systems. It deals with estimating the state of a system based on noisy measurements. In the linear case, the Kalman filter has a well-defined and efficient solution. However, in the nonlinear case, the correct solution remains an unsolved problem [30]. The nonlinear Kalman filter is an active area of research in control theory and estimation. The challenge of solving the nonlinear Kalman filter is due to the complexity introduced by the nonlinearity of the system dynamics and the uncertainty in the process and measurement noise. Nonlinear Kalman filters, such as the Extended Kalman Filter (EKF) and Unscented Kalman Filter (UKF), are commonly used as approximations, but they may have limitations and are not guaranteed to provide the globally optimal solution. Researchers continue to work on developing better algorithms and approaches to handle the nonlinear case effectively and efficiently. We hope that the methods developed here and in [31] could work for the nonlinear Kalman filter case as well.

Author Contributions

Software, V.S.; Investigation, H.M. and D.B.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Russian Science Foundation grant No. 23-11-00116, SCS of Armenia grants No. 21T-1C037, 20TTAT-QTa003 and Faculty Research Funding Program 2022 implemented by the Enterprise Incubator Foundation with the support of PMI Science.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Acknowledgments

D.B.S. thanks I. Gornyi and P. Popper for discussions.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. The Relaxation Dynamics for the Correlated Case

Consider now the correlated choice of the matrix type. If we have a matrix A, then at the next round, we hold it with the probability q 11 = 1 π 1 , while changing it to matrix B with the probability q 12 = 1 π 1 . If we have a matrix B, then at the next round, we hold it with the probability q 22 = 1 π 2 while changing it to matrix B with the probability q 21 = 1 π 2 .
We have the following equation for the probability distributions [23]
ρ ^ 1 ( n + 1 , α ) = [ q 11 ρ ^ 1 ( n , f 1 ( α ) ) + q 21 ρ ^ 2 ( n , f 1 ( α ) ] f 1 ( α ) ρ ^ 2 ( n + 1 , α ) = [ q 12 ρ ^ 1 ( n , f 2 ( α ) ) + q 22 ρ ^ 2 ( n , v + g 2 ( α ) f 2 ( α ) ] f 2 ( α )
f i ( α ) = arccos [ ( A i , 11 1 ( α ) + A i , 21 1 sin ( α ) ) < α | ( A i 1 ) * A i 1 | α > ] , A i , 21 1 cos ( α ) + A i , 22 1 sin ( α ) 0
When A i , 21 1 cos ( α ) + A i , 22 1 sin ( α ) < 0 , we take f i ( α ) π + f i ( α ) .

Appendix B. Numerical Algorithms

We used two numerical algorithms.

Appendix B.1. The First Algorithm

We take grids uniformly with a step ϵ = 2 π / N and define the probability distribution at the grid x i as p i . We replace p ( f l ( x i ) ) by p j , where x j is the coordinate of the closest grid to the point f ( x i ) .

Appendix B.2. The Second Algorithm

Consider our functions f 0 ( x ) , f 1 ( x ) in the interval 0 2 π . Let us we have a singularity point at the point x c . Then we consider the grids, n 1 + 1 points with the step x c / n 1 = a between 0 and x c , n 2 1 points with the step ( 2 π x c ) / n 2 = b between x c and 2 π . We identify the points 0 , 2 π .
We have n = n 1 + n 2 points in the cycle:
x i = ( i 1 ) a , 1 i n 1 + 1 , x i = x c + ( i 1 ) b , n 1 + 2 i n
Next, we identify the probabilities in in 0 , 2 π points.
We define the total probability as
j = 1 n p j w j
where
w j = a ,   1 i n 1 , w j = b ,   1 + n 1 j n 1 + n 2
w 1 = a / 2 + b / 2 ,   w n 1 + 1 = a / 2 + b / 2
Next, for any j we define N i and N i + as the neighbors of the f ( x j ) point. Then we define d = d ( N i , f ( x j ) ) and d + = d ( N i + , f ( x j ) . Next, we define α j , N = d + d + d + and α j , N + = d d + d + . Eventually, we define the transition matrix us
M i j = W α W 1
where W , α are matrices with the matrix elements W i i = w i .

References

  1. Furstenberg, H.; Kesten, H. Products of random matrices. Ann. Math. Statist. 1960, 31, 457–469. [Google Scholar] [CrossRef]
  2. Furstenberg, H. Noncommuting random products. Trans. Am. Math. Soc. 1963, 108, 377–428. [Google Scholar] [CrossRef]
  3. Bougerol, P.; Lacroix, J. Products of Random Matrices with Applications to Schrodinger Operators; Birhauser: Basel, Switzerland, 1985. [Google Scholar]
  4. Chamayou, J.F.; Letac, G. Explicit stationary distributions for compositions of random functions and products of random matrices. J. Theor. Prob. 1991, 4, 3–36. [Google Scholar] [CrossRef]
  5. Crisanti, A.; Paladin, G.; Vulpiani, A. Products of Random Matrices in Statistical Physics; Springer: Berlin/Heidelberg, Germany, 1993. [Google Scholar]
  6. Comtet, A.; Luck, J.M.; Texier, C.; Tourigny, Y. The Lyapunov exponent of products of random 2 × 2 matrices close to the identity. J. Stat. Phys. 2013, 150, 13–65. [Google Scholar] [CrossRef]
  7. Comtet, A.; Texier, C.; Tourigny, Y. Lyapunov exponents, one-dimensional Anderson localization and products of random matrices. J. Phys. A 2013, 46, 254003. [Google Scholar] [CrossRef]
  8. Comtet, A.; Tourigny, Y. Impurity models and products of random matrices. arXiv 2016, arXiv:1601.01822. [Google Scholar]
  9. Ephraim, Y.; Merhav, N. Hidden Markov processes. IEEE Trans. Inform. Theory 2002, 48, 1518–1569. [Google Scholar] [CrossRef]
  10. Mamon, R.S.; Elliott, R.J. (Eds.) Hidden Markov Models in Finance: Further Developments and 147 Applications Volume II; Springer Nature: Berlin/Heidelberg, Germany, 2014. [Google Scholar]
  11. Saakian, D.B. Exact solution of the hidden Markov processes. Phys. Rev. E 2017, 96, 052112. [Google Scholar] [CrossRef]
  12. Zuk, O.; Kanter, I.; Domany, E. The entropy of a binary hidden Markov process. J. Stat. Phys. 2005, 121, 343–360. [Google Scholar] [CrossRef][Green Version]
  13. Allahverdyan, A.E. Entropy of Hidden Markov Processes via Cycle Expansion. J. Stat. Phys. 2008, 133, 535. [Google Scholar] [CrossRef][Green Version]
  14. Hufton, P.G.; Lin, Y.T.; Galla, T.; McKane, A.J. Intrinsic noise in systems with switching environments. Phys. Rev. E 2016, 93, 052119. [Google Scholar] [CrossRef]
  15. Mayer, A.; Mora, T.; Rivoire, O.; Walczak, A.M. Diversity of immune strategies explained by adaptation to pathogen statistics. Proc. Natl. Acad. Sci. USA 2016, 113, 8630. [Google Scholar] [CrossRef]
  16. Kussell, E.; Leibler, S. Phenotypic diversity, population growth, and information in fluctuating environments. Science 2005, 309, 2075–2078. [Google Scholar] [CrossRef]
  17. Skanata, A.; Kussell, E. Evolutionary Phase Transitions in Random Environments. Phys. Rev. Lett. 2016, 117, 038104. [Google Scholar] [CrossRef]
  18. Wienand, K.; Frey, E.; Mobilia, M. Evolution of a Fluctuating Population in a Randomly Switching Environment. Phys. Rev. Lett. 2017, 119, 158301. [Google Scholar] [CrossRef]
  19. Rivoire, O.; Leibler, S. The value of information for populations in varying environments. J. Stat. Phys. 2011, 142, 1124–1166. [Google Scholar] [CrossRef]
  20. Rivoire, O. Informations in models of evolutionary dynamics. J. Stat. Phys. 2016, 162, 1324–1352. [Google Scholar] [CrossRef]
  21. Saakian, D.B. Semianalytical solution of the random-product problem of matrices and discrete-time random evolution. Phys. Rev. E 2018, 98, 062115. [Google Scholar] [CrossRef]
  22. Snizhko, K.; Kumar, P.; Romito, A. Quantum Zeno effect appears in stages. Phys. Rev. Res. 2020, 2, 033512. [Google Scholar] [CrossRef]
  23. Poghosyan, R.; Saakian, D.B. Frontiers in Frontiers, Infinite Series of Singularities in the Correlated Random Matrices Product. Front. Phys. 2021. [Google Scholar] [CrossRef]
  24. Nilsson, M.; Snoad, N. Error Thresholds for Quasispecies on Dynamic Fitness Landscapes. Phys. Rev. Lett. 2000, 84, 191. [Google Scholar] [CrossRef] [PubMed]
  25. Crow, J.F.; Kimura, M. An Introduction to Population Genetics Theory; Harper and Row: New York, NY, USA, 1970. [Google Scholar]
  26. Baake, E.; Wagner, H. Mutation-selection models solved exactly with methods of statistical mechanics. Genet. Res. 2001, 78, 93–117. [Google Scholar] [CrossRef] [PubMed]
  27. Saakian, D.B.; Hu, C.-K. Solvable biological evolution model with a parallel mutation-selection scheme. Phys. Rev. E 2004, 69, 046121. [Google Scholar] [CrossRef]
  28. Xie, R.; Marsili, M. A random energy approach to deep learning. J. Stat. Mech. Theory Exp. 2022, 7, 073404. [Google Scholar] [CrossRef]
  29. Kalman, R.E. A New Approach to Linear Filtering and Prediction Problems. J. Basic Eng. 1960, 82, 35. [Google Scholar] [CrossRef]
  30. Gustafsson, F.; Hendeby, G. Some Relations between Extended and Unscented Kalman Filters. IEEE Trans. Signal Process. 2012, 60, 555. [Google Scholar] [CrossRef]
  31. Galstyan, V.; Saakian, D.B. Quantifying the stochasticity of policy parameters in reinforcement learning problems. Phys. Rev. E 2023, 107, 034112. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.