Path Counting on Tree-like Graphs with a Single Entropic Trap: Critical Behavior and Finite Size Effects

It is known that maximal entropy random walks and partition functions that count long paths on graphs tend to become localized near nodes with a high degree. Here, we revisit the simplest toy model of such a localization: a regular tree of degree p with one special node (“root”) that has a degree different from all the others. We present an in-depth study of the path-counting problem precisely at the localization transition. We study paths that start from the root in both infinite trees and finite, locally tree-like regular random graphs (RRGs). For the infinite tree, we prove that the probability distribution function of the endpoints of the path is a step function. The position of the step moves away from the root at a constant velocity v=(p−2)/p. We find the width and asymptotic shape of the distribution in the vicinity of the shock. For a finite RRG, we show that a critical slowdown takes place, and the trajectory length needed to reach the equilibrium distribution is on the order of N instead of logp−1N away from the transition. We calculate the exact values of the equilibrium distribution and relaxation length, as well as the shapes of slowly relaxing modes.


I. INTRODUCTION
One of the most natural ways to characterize a graph or a network is by its adjacency and Laplace matrices [1,2].The adjacency matrix of an N -node network is defined as N × N matrix A, where the entry A ij = 1 if nodes numbered i and j are connected to each other, and 0 otherwise.The elements of the Laplacian matrix L of a graph are defined as where deg j is the degree of node j.This matrix defines the relaxation of a continuous-time diffusion process on a graph, i.e., the probability distribution to find a particle at one of the nodes, P L (t) = (P L (1, t), P L (2, t), . . ., P L (N, t)) T , evolves according to the following: and, therefore, eigenvalues and eigenvectors of the Laplacian play a crucial role in describing the relaxation of a simple diffusion process on the networks (see, e.g., [3]).
It seems natural to ask whether there is a process whose relaxation is related in a similar way to the eigenvalues and eigenvectors of the adjacency matrix.We define a sequence of vectors Z(t) = (Z 1 (t), Z 2 (t), . . ., Z N (t)) T that are dependent on discrete time t, satisfying Z(t + 1) = AZ(t). ( Appended with the initial condition Z(0) = (1, 1, . . ., 1) T , it can be interpreted as a vector of partition functions counting trajectories of length t on a graph, with Z i (t) counting trajectories ending at the i-th node.Moreover, we define which is the probability distribution to find the end of a randomly chosen trajectory at point i.A similar problem arises (although typically in a continuous space setting) in the classical polymer theory [4][5][6], so the path-counting problem (PC) described above is also often referred to as the counting of "conformations of an ideal polymer" [7,8].Note the difference between the probability distribution (4) and that of an endpoint of a symmetric discrete-time random walk on a graph (2).Indeed, in a symmetric random walk, the probabilities of trajectories are weighted inversely to the product of the degrees of the visited nodes, while in the path-counting problem (PC), all trajectories are weighted equally.Is it possible to reformulate a path-counting problem as a random walk problem, i.e., as a process satisfying with some stochastic transfer matrix T? To the best of our knowledge, the answer is "no", but for finite graphs, a very close approximation known as the maximal entropy random walk (MERW) is possible [9,10].Namely, if we define λ 1 , Φ 1 to be the largest eigenvalue of the adjacency matrix and the corresponding eigenvector, and arXiv:2308.02611v3[cond-mat.stat-mech]21 Sep 2023 choose the elements of the transfer matrix as where ϕ 1i is the i-th element of Φ 1 , then (i) these weights can be understood as step probabilities of some random walk (indeed, i A ji ϕ 1i = λ 1 ϕ 1j (7) so the probability is conserved), and (2) the trajectory weight in such a walk is only dependent on its length, starting point, and endpoint, i.e., all trajectories with given lengths and endpoints are equiprobable.As a result, if one considers sufficiently long MERW trajectories, the density of links at a given site i becomes proportional to the square of the value of the largest eigenvector at this site ϕ 2 1i , in full analogy with the density of the ideal polymer in the external field [4,6].
Thus, on the one hand, MERW and PC problems are fundamentally related.Answers to both can be fully formulated in terms of eigenvectors and eigenvalues of the same adjacency matrix A. On the other hand, the probability distribution P A (t) (probability distribution of the endpoint of a trajectory of a given length) defined above for the PC problem has no direct analog in MERW.We formulate our results below in terms of this probability distribution, so, to avoid confusion, we restrict ourselves to the terminology of the path-counting problem, although the results can easily be reformulated in terms of MERW.
The most striking feature of the PC (and MERW) as opposed to regular random walks is the potential for localization.Indeed, while on any finite connected graph, P L converges to a uniform distribution for sufficiently long trajectories, it is not generally the case for P A .In fact, large-degree nodes work as entropic traps in the PC problem, which may lead to the localization of trajectories.Perhaps the simplest system where such localization can happen is a regular tree with a branching degree p and a single "special" node ("root") with a different branching degree p 0 .This system has been studied extensively in [8] (also see recent generalizations in [11,12]) where it has been shown that for an infinite tree, there exists a critical branching degree: such that for root degrees p 0 that are smaller than p cr , sufficiently long trajectories span the whole graph, and all nodes are visited with approximately equal probabilities; for larger values of p 0 , trajectories are localized in the vicinity of the root, causing the probability of identifying the endpoint of a randomly selected trajectory at a distance x from the root to decrease exponentially with x.
Interestingly, in the analogous MERW problem, which has been studied in [13], the localization transition occurs at a different value of the root degree: This discrepancy can be understood when recognizing that the large-t distributions of trajectory inner points and their ends, typically studied in the MERW and PC setups, are proportional to ϕ 2 1i and ϕ 1i .It is clear from the symmetry of the system that values of ϕ 1i can only depend on distance x from node i to the root, ϕ 1i = ϕ 1x , and the probability of being at distance x from the origin is proportional, up to the normalization constant to ϕ 2 1x (p − 1) x and ϕ 1x (p − 1) x for inner points and endpoints, respectively.As a result, localization for endpoints occurs if ϕ 1x decays with x faster than (p − 1) −x .For inner points, localization occurs if the decay is faster than (p − 1) −x/2 .Notably, the standard Brownian random walk on a tree-like graph can be localized around a single site, as discussed in [14], but for this to happen, one needs to apply a uniform global external field [14].Indeed, a random walk on a tree-like graph can be mapped on a biased random walk on a half-line [15], and if the external field is strong enough to compensate for that bias [16], condensation occurs in a similar way to a random walk on a half-line that is biased in the direction of the origin.Contrary to that, localization in the path-counting problem occurs as a result of a point-like local disorder [8] or, equivalently, as noted in [17,18], as a result of an external field applied only to the endpoint of the trajectory.For Brownian motion, such localization is impossible; it is easy to see that in a potential composed of a localized well at the origin and a constant bias steering away from it, the Brownian walker always escapes.
In this paper, we revisit this system and pay special attention to the behavior at the transition point p 0 = p cr .First, we consider the behavior on an infinite tree.Here, we prove the results presented in [8], asserting that the distribution of the endpoint x of the trajectory commencing at the root asymptotically has the form of a traveling wave where velocity v equals (p − 2)/p and Θ(x) is the Heaviside theta function.We calculate the shape of the wave in the vicinity of the front, i.e., for vt − x ∼ O( √ t).Second, we study the critical behavior of trajectories for p 0 = p cr on a finite locally tree-like graph with a fixed degree of all nodes, except the root (random regular graph, RRG).In particular, we find the limiting distribution P A (x, t → ∞) and show that this limiting distribution is approached extremely slowly, so that the maximal relaxation time t relax ∼ √ N as opposed to t relax ∼ log (p−1) N for p 0 ̸ = p cr .

II. MODEL AND RESEARCH QUESTIONS
Following [8], let us briefly review what is known about path localization on an infinite regular tree with a single special point.We consider an infinite tree graph with coordination number p in all points, except for a single point (root) with a different coordination number p 0 > p.We are interested in the statistics of long paths on such a graph.On an infinite tree, the parity of x + t, where x is the distance of the end of a t-step trajectory to the root, is exactly conserved.For the statistics of the endpoints of long paths, three regimes are possible [8].
If p 0 < p cr , the typical trajectory escapes from the vicinity of the root, with an average distance from the trajectory's endpoint to the root growing as ⟨x(t)⟩ = (p − 2)t/p.The fluctuation of x(t) in this case is normally distributed with the width of distribution proportional to √ t.This behavior is completely analogous to the behavior of a trajectory on the infinite regular tree (corresponding to p 0 = p), and can be described as a biased discrete-time Brownian motion on a half-line.
In turn, if p 0 > p cr , localization occurs, meaning that the endpoint of a long trajectory is localized in the vicinity of the root with ⟨x(t)⟩ converging to a constant for the large t, and the probability of finding the trajectory's endpoint at distance x away from the root diminishing exponentially with x.
Finally, in a critical case, p 0 = p cr , very interesting behavior is numerically observed in [8]: the probability of finding the end of an t-step trajectory starting from the root at a distance x from it is roughly a step function: with some crossover function linking these limiting regimes for x ≈ (p − 2)t/p (see Fig. 1b).However, this latter result has not been proven in [8].In this paper, we focus on the details of the critical behavior, i.e., the path statistics in a tree with branching degrees p in all points but the root, where the branching degree is p 0 = p cr = p(p−1) in order to further elucidate the behavior in this regime.First, in Section III, we address the behavior of the infinite system and prove (11).In Section IV, we discuss how the critical regime looks in the case when a tree is large but finite.We show that it is important to distinguish between a Cayley tree (a regular tree with a finite fraction of boundary nodes of degree one) in the limit of a large number of generations and a random regular graph, i.e., a local tree-like graph where nodes far away from the root are randomly connected to guarantee that each node-except for the root-has degree p.In the first system, there is no critical behavior at p = p cr for sufficiently long trajectories, while for the second system, a very peculiar ultra-slow relaxation occurs.

III. CRITICAL STATE ON AN INFINITE TREE
Introduce a partition function Z t (x), counting the number of paths beginning at the root, consisting of t steps and ending at a distance x from the root.It is easy to see that it satisfies: for x = 0 (12) with initial condition Z 0 (x) = δ x,0 , and we are particularly interested in p 0 = p cr .The introduction of a renormalized partition function W t (x) according to allows us to symmetrize the equations, giving where we substitute the critical value of p according to (8).These equations define, for non-negative integer indices t and x, a two-parametric sequence W , which has a simple combinatorial meaning.Indeed, for p = 1 it simply enumerates trajectories in the upper-right quadrant, consisting of right-up and right-down steps (see Fig. 2), which start at the origin and end at the point with coordinates (t, x).For arbitrary p, it enumerates weighted trajectories: where we introduce C m t (x) -the number of trajectories from (0, 0) to (t, x) which return to the horizontal axis x = 0 exactly m times.Note that the trajectory acquires weight p every time it leaves x = 0, so the additional 1 in the power on the left-hand side accounts for the initial step, and the Kronecker symbol δ x,0 accounts for the scenario where, for x = 0, the trajectory is still on its last return without having left 0. To calculate C m t (x), note that there is a bijection between all paths going from (0, 0) to (t, x) that visit the horizontal axis m times, and all paths from (0, 0) to (t − m, x + m) that never visit the horizontal axis; therefore, This bijection is constructed as follows.Consider first a path from (0, 0) to (t, x) that intersects the horizontal axis m times and exclude from it all the steps from x = 1 to x = 0 (marked in red in Fig. 2); shift the remaining parts of the path upward and to the left so that they form a continuous path.Clearly, the resulting path goes from (0, 0) to (t−m, x+m) without returning to the horizontal axis.Consider now an arbitrary path from (0, 0) to (t − m, x + m) that does not visit the horizontal axis; for each vertical coordinate value from i = 1 to i = m, mark a point (t i , i) on the path where it is attained for the final time.It follows from x + m ≥ m that these points exist and form a strictly increasing sequence t 1 < t 2 < • • • < t m .Now, split the path in these points, shift each of the resulting subpaths by vector (i, −i), and insert downward steps connecting these paths together, as shown in Fig. 2. Clearly, the path resulting from this transformation is a path from (0, 0) to (t, x), visiting the horizontal axis exactly m times.It is easy to see that in both the direct and reverse transformations, different initial paths result in different transformed paths; thus, there exists a one-to-one correspondence between these two path classes, proving (16).In turn, the number of paths not visiting zero C 0 t−m (x + m) can be easily calculated, e.g., by the method of images, with the result being Substituting this result into (13) and (15), we obtain the following: Consider now the asymptotic behavior of this expression.Assume that t, x ≫ 1 and x/t = α.Then, after replacing factorials with their approximate values using the Stirling formula and introducing notation, obtain Substituting this expression into (18), rewrite the partition function in the following approximate form: where and The partition function, thus, takes the form pertinent to the application of the Laplace method.It is easy to see that the stationary point ξ 0 defined by ϕ ′ (ξ) = 0 is given by and belongs to interval 0 ≤ ξ 0 ≤ 1 if β ∈ [1/p, 1] and, thus, α ∈ [0, (p − 2)/p], while for larger α, the maximal value of ϕ(ξ) is reached for ξ = 0. Thus, the Laplace approximation gives the leading asymptotic of Z t (αt) for any fixed α.If α < (p − 2)/p it gives and Z t (αt) = Z 0 (t) is an α-independent constant, while for α > (p − 2)/p Thus, in the limit of large t, the solution has the shape of a traveling wave.The probability distribution of the position of the trajectory's endpoint is constant up to x = vt, where v = (p − 2)/p, and is asymptotically zero for a larger x.To elucidate the shape of the wave near x = vt, note that the Laplace approximation implies replacing the limits of integration in (21) with plus and minus infinity.For a better approximation of x = vt + ∆ √ t, one should replace the integrand in (21) with the corresponding Gaussian function, but retain the correct integration limit, giving Fig. 3 shows that this expression provides a satisfactory approximation to the shape of the wave at ∆.

CRITICAL BEHAVIOR ON FINITE GRAPHS
Thus, we have proven in the previous section that the distribution of the endpoint of a t-step trajectory on an infinite critical tree reaches a limiting shape of the form erfc c(x − vt)/ √ t , moving toward infinity with constant velocity v = (p − 2)/p.In a finite system, this process cannot continue forever: at some point, the expanding wave feels that the system is finite.Clearly, the number of steps at which this occurs is determined by the following: where N is the total number of nodes in the graph.Further evolution depends on how exactly the finite tree is defined.Indeed, the notion of a "finite tree with constant degree p > 2" is ill-defined.A finite tree of N nodes has exactly N − 1 bonds and, therefore, must have an average degree 2(N − 1)/N ≈ 2. This discrepancy is usually resolved in one of two possible ways.One option is to consider a k-generation tree; it has a boundary layer consisting of p 0 (p − 1) k−1 nodes with degree 1.These nodes maintain a final fraction of the system size when k → ∞ and shift the average node degree to 2. Another option is to consider a graph in which all nodes have degree p (except for the special node with degree p 0 ), but which is only locally tree-like.This means that while in such a graph there are no short cycles (i.e., it is locally tree-like), it does contain long cycles with a minimum length on the order of t 0 .We will call this option a random regular graph (RRG).A more detailed discussion of the terminology is presented in the next section.The qualitative trajectory behaviors in both cases are similar to the behaviors on an infinite tree only up to a trajectory length of around t 0 .For larger trajectories, the behavior is radically different.Indeed, on a large but finite tree, whenever the expanding distribution reaches the boundary, it becomes stationary; for any t > t 0 , it remains essentially the same, namely, the probability of the trajectory to the end at distance x from the root is x-independent and roughly (up to some minor deviations near x = 0 and x = k) equals: which means that the probability of ending the trajectory at any given node of the tree is, in fact, proportional to (p − 1) −x (considering that there are roughly p(p − 1) x nodes at a distance x from the root).However, as we show in the next section, both the limiting behavior and the typical length needed to reach it, are radically different in the case of the RRG.Indeed, in the limit of large t, the trajectory's endpoint distribution consists of two zones, namely, the zone close to the root, where the likelihood of finding the trajectory's end decays with x at a rate of (p − 1) −x , and the remaining bulk of the system, where the probability of finding the trajectory end remains position-independent.However, the trajectory length t relax needed to reach this final state is extremely long, t relax ∼ √ N (compare this with t 0 ∼ ln N ).

Random Regular Graph (RRG)
Consider a graph constructed as follows.Take a k-generation tree with degree p at all branching points except the root, and degree p 0 = p cr = p(p − 1) in the root.This tree has p(p − 1) k "leaves" (periphery nodes of degree one).Now, add p(p − 1) k+1 /2 new bonds connecting the leaves in a way that each leave becomes a node of degree p.Such graphs are reminiscent of the random regular graph (RRG) ensemble (which is defined as a maximal entropy ensemble of graphs whose nodes have a constant degree p) in the sense that it is locally tree-like (the probability of forming a short cycle is exponentially small in k), and all nodes, except for the root, have a constant degree p.The most important difference between the graph defined above and the true RRG ensemble is as follows: in the former the shortest cycle, which includes the root, has, by construction, length 2k, while in the latter there exist shorter cycles, which include the root.However, the fraction of such cycles exponentially decreases with 2k − l; therefore, we expect the difference to be negligible.
To check whether the resulting graphs are similar to the true RRG, we calculate (see Fig. 5) the eigenvalue spectrum of an ensemble of such graphs (trees with randomly connected external nodes) for p 0 = p and compare it to the known limiting distribution for the true RRG ensemble [19,20] Clearly, the distribution is very close to the limiting one, except for k delta-functional peaks reflecting the existence of a regular tree sub-graph.However, the relative sizes of these peaks decrease, and for a large k, the distribution converges to the Kesten-McKay result (31).This result is, in fact, to be expected, given that the only assumptions needed to prove (31) (see [20]) are (i) that all nodes have the same degree p, and (ii) the concentration of a cycle of any finite length m converges to zero with the growing size of the network; both are true for the suggested tree closure construction.
For the purposes of this paper, the most interesting question is the behavior at the edge of the spectrum, where the properties of the true RRG ensemble, and our model RRGs are exactly the same.For p 0 = p, there is a single maximal eigenvalue λ 1 = p, reflecting the fact that-on each next step-every trajectory has exactly p possible continuations, and the continuous spectrum with density converges to 0 as so there is a k-independent final gap between the maximal and second-largest eigenvalue.
Let us now calculate the distribution of the endpoints of a t-step trajectory on such a quasi-RRG graph.To do that, we introduce a (k + 1)-dimensional vector Z t , so that the x-th component of this vector equals the number of t-step trajectories starting from the root and ending at distance x.
The recurrent relation for the number of paths can be written in matrix form: with the tri-diagonal (k + 1) Note the (p−1) entry in the lower-left corner of the transfer matrix.It corresponds to the trajectories "jumping" from one node of the k-th generation to the other along the random connection bonds shown in Fig. 4. It is this single entry that distinguishes the transfer matrix of the RRG graph problem from the transfer matrix T 0 of the finite tree one, where this entry equals 0.
FIG. 4: A sketch of a model graph with a fixed degree p = 3, except for the root; up until the k = 3'rd generation from the root with degree p 0 = 5, the graph is a tree, and the outer nodes are connected randomly with each other so that their degree equals p.It is easy to see that all eigenvalues of both T and T 0 are non-degenerate and real.Indeed, it is sufficient to note that Ω −1 TΩ, where Ω is a diagonal matrix with elements is a symmetric tri-diagonal matrix.The decomposition of T allows us to write where index i enumerates the eigenvalues and eigenvectors of T. Assume for convenience that the eigenvalues are enumerated in decreasing order of their absolute value: . . .Clearly, the large t, asymptotic of the partition function, is controlled by the eigenvalues with the largest absolute values and corresponding eigenvectors.
The eigenvalue problem for the transfer matrix T 0 has been studied at length in [8,21].The characteristic polynomials of that problem P k (λ) satisfy And have the explicit form where and where we introduce parametrization Notably, parametrization (40) means that u is real for |λ| ≥ 2 √ p − 1 and imaginary otherwise.To proceed further, replace p 0 = p(p − 1) and note that characteristic polynomials D(λ) = det(λI − T) of the RRG transfer matrix T are easily expressed in terms of P k (λ) Given that we are interested in eigenvalues with the largest absolute value, we concentrate here on the real solutions of the characteristic equation D k (u) = 0.After substituting (37)-( 40) into (41), we rewrite it in the form: where we introduce notation v = √ p − 1.For every k, this equation has three roots, u 1,2,3 , converging to the zeros of the numerator for the large k: Expanding (42) up to second order in ϵ and solving the resulting equations, we obtain the following finite-size correction terms for positive and negative eigenvalues, respectively.After returning to the original variables, we obtain and Figure 6a,b show the asymptotic behaviors of the discrepancies λ i −λ i (k = ∞) as compared to their approximate values given by ( 46) and (47).53) and (54) are shown.Ψ (1) : Red triangles and green line; Ψ (2) : black triangles and blue line; Ψ (3) : light green triangles and purple line, respectively.
We notice that, in the limit of the large number of generations, the gaps between the absolute values of the three leading eigenvalues are exponentially small: where we use notation ∆λ i = |λ i | − |λ i+1 |, indicating that trajectories on the order of (p − 1) k/2 ∼ √ N steps are needed to reach the asymptotic distribution of the endpoints.Note that since there are only three real solutions of (42), the remaining eigenvalues and these are separated from the triplet defined by ( 46) and (47).
To understand the limiting distribution of the endpoints, as well as the modes that demonstrate a remarkably slow decay, let us find the eigenvectors corresponding with these three leading eigenvalues.Let Ψ i be the i-th eigenvector, i.e., a solution to (i) x be its x-th element.Choosing the initial condition Ψ (i) 0 = 1 and solving recursively with respect to x, we obtain where P x−1 (λ) is defined above in (37).Noticing that and expanding P x−1 (λ) in the vicinity of these points, we obtain the following in the leading order: and Ψ (2)  x ≈ (−1) Thus, from the point of view of the senior eigenvector Ψ 1 , the graph consists of two different zones: x ≲ k 2 and x ≳ k 2 .In the first zone (see Fig. 7), the probability distribution over x is flat, while in the second, it grows as (p − 1) x , i.e., proportionally to the number of nodes in the x-generation.Thus, the probability of the trajectory ending at a given node is position-independent in the second zone and proportional to (p − 1) −x (where x is the much richer and consists of two regimes.There is a fast (on the order of t 0 ∼ log p−1 N ) expanding wave regime, similar to the evolution of the infinite system, followed by a much slower relaxation with maximal relaxation time (on the order of t relax ∼ √ N ).This regime is an example of a critical slowdown and, remarkably, it is completely unattainable when studying infinite systems.Indeed, while for any reasonable value of N t 0 may seem quite small, it formally diverges as N → ∞.
The existence of the critical slowdown regime can be qualitatively understood from the fact that, in the thermodynamic limit, at the critical point, the transfer matrix of the problem has multiple eigenvalues with exactly the same absolute value (in the case considered here, there are three of them, reflecting a conservation of the parity of t + x, in the absence of such additional conservation laws, there are typically two).Meanwhile, the maximal eigenvalue of the adjacency matrix of a finite connected graph is non-degenerate [1], which dictates that there must be a finite gap between these eigenvalues in any finite system.It is natural to expect this gap to be quite small in the case of a large graph, that is, to be much smaller than what one expects away from the critical point where the gap remains finite even in the infinite graph.This qualitative consideration explains why one expects the critical slowdown to occur.
One expects that such a slowdown is a general phenomenon, which should be typical in the study of localization transitions on other large but finite graphs.One example is the Anderson localization on the Bethe lattice [22,23] (also see review [24]) and on RRG [25,26], which has been actively studied recently due to its suggested connection to the many-body localization problem [27].Indeed, the exponential divergence of the diffusion constant in the vicinity of the critical point [22] is reminiscent of the critical slowdown discussed here.
Another relevant example is the celebrated quasispecies model [28] of biological evolution.The model can be briefly formulated as follows.Consider a set of possible genomes σ = (σ 1 , σ 2 , . . ., σ d ), consisting of d binary variables σ i = 0, 1, so that there are 2 d possible genomes.We introduce a population function Z(σ, t), reflecting the occurrence of each genome at time t.Then a discrete time evolution of the quasispecies model in the absence of the niche size constraint reads where α σ is the fitness of the genome σ and Π σ,σ ′ is the probability of obtaining σ while copying σ ′ , allowing for possible copying errors.In this system, depending on the probability of copying errors, a localizationdelocalization transition, often called the error threshold or error catastrophe [28,29] (also see the pedagogical presentation in [30]) is possible.In the localized state, the observed genomes are close to the optimal one (i.e., the one with the highest fitness) and form a localized cloud known as a quasispecies, while in the delocalized state, all genomes appear with equal probability.We make numerical simulations of this system for the simplest case, where where σ 0 is the genome with optimal fitness, α > 0 and q ∈ (0, 1) are parameters, and d H = d H (σ, σ ′ ) is the Hamming distance (the number of differences in genome letters) between genomes σ and σ ′ .For simulations, we use d = 60 and q = 0.5 and find that localization transition occurs at α = d − 1.Moreover, Fig. 8 shows that there is strong evidence of critical slowdown at the critical point (Fig. 8b), where the number of steps of the time evolution needed to approach the limiting distribution is 1-2 orders of magnitude larger than in the bulk of localized (Fig. 8a) and delocalized (Fig. 8c) states.We are not able to solve this model analytically but we suggest that the simpler, exactly solvable model presented in this paper can serve as a toy model for this more complex one.

4 FIG. 2 :
FIG.2: Illustration of the bijection between trajectories from (0, 0) to (t, x) visiting 0 m times and trajectories from (0, 0) to (t − m, x + m) not visiting zero, reflected in Equation (16).To get from trajectory shown in the top panel to that shown in the bottom panel, remove the steps highlighted in red, and move the remaining pieces so they are joined together.To get from trajectory shown in the bottom panel to that shown in the top panel, insert (1,−1) steps (red lines) in place of the red dots.Red dots are points where x = 1, 2, . . ., m for the final time, i.e. rightmost intersections of the trajectory and x = 1, 2, . . ., m lines (dotted lines).

FIG. 5 :
FIG.5: Eigenvalue spectrum of the approximate RRG graph created by a procedure shown in Fig.4for p = 3, k = 9 averaged over 100 independent realizations (histogram) vs. the limiting spectrum of the true large RRG graph given by (31) (red line).Note the spectral gap between the largest eigenvalue λ 1 = p = 3 and the edge of the continuous spectrum at λ = 2 √ p − 1 ≈ 2.828.

FIG. 6 :
FIG. 6: (a,b): Largest eigenvalues of the transfer matrix T as functions of k for p = 3 in the logarithmic (main) and linear (inset) scales.(a): Absolute value of the difference between positive eigenvalues λ 1,3 and p; numerical results are shown in blue (λ 1 ) and red (λ 3 ) triangles; asymptotic (46) is shown by magenta and green lines.(b) Absolute value of the difference between the negative eigenvalue λ 2 and −p with an inserted plot of eigenvalues in a normal scale.(c): Components of eigenvectors Ψ (i) , i = 1, 2, 3 as functions of coordinate x for p = 4, k = 30; numerical values and asymptotic given by (53) and (54) are shown.Ψ(1) : Red triangles and green line; Ψ(2) : black triangles and blue line; Ψ(3) : light green triangles and purple line, respectively.