Next Article in Journal
Protecting Infrastructure Networks: Solving the Stackelberg Game with Interval-Valued Intuitionistic Fuzzy Number Payoffs
Previous Article in Journal
Verification and Enforcement of (ϵ, ξ)-Differential Privacy over Finite Steps in Discrete Event Systems
Previous Article in Special Issue
Estimates of the Convergence Rate in the Generalized Rényi Theorem with a Structural Digamma Distribution Using Zeta Metrics
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Limit Distributions of Products of Independent and Identically Distributed Random 2 × 2 Stochastic Matrices: A Treatment with the Reciprocal of the Golden Ratio

by
Santanu Chakraborty
School of Mathematical and Statistical Sciences, University of Texas Rio Grande Valley, 1201 West University Drive, Edinburg, TX 78539-2999, USA
Mathematics 2023, 11(24), 4993; https://doi.org/10.3390/math11244993
Submission received: 27 September 2023 / Revised: 4 December 2023 / Accepted: 6 December 2023 / Published: 18 December 2023

Abstract

:
Consider a sequence ( X n ) n 1 of i.i.d. 2 × 2 stochastic matrices with each X n distributed as μ . This μ is described as follows. Let ( C n , D n ) T denote the first column of X n and for a given real r with 0 < r < 1 , let r 1 C n and r 1 D n each be Bernoulli distributions with parameters p 1 and p 2 , respectively, and 0 < p 1 , p 2 < 1 . Clearly, the weak limit of the sequence μ n , namely λ , is known to exist, whose support is contained in the set of all 2 × 2 rank one stochastic matrices. In a previous paper, we considered 0 < r 1 2 and obtained λ explicitly. We showed that λ is supported countably on many points, each with positive λ -mass. Of course, the case 0 < r 1 2 is tractable, but the case r > 1 2 is very challenging. Considering the extreme nontriviality of this case, we stick to a very special such r, namely, r = 5 1 2 (the reciprocal of the golden ratio), briefly mention the challenges in this nontrivial case, and completely identify λ for a very special situation.

1. Introduction

As the title of the paper suggests, the reader can understand that this paper deals with a situation where one considers products of independent and identically distributed random 2 × 2 stochastic matrices and their limiting behavior. In other words, here we are considering a probability measure μ on a collection of 2 × 2 stochastic matrices and studying the limiting behavior of the convolution sequence μ n . To a reader new to this area, the author would like to refer the reader to the book by Hognas and Mukherjea [1]. This book starts from the very basic concepts, such as the definition of a semigroup, topological semigroups, semigroups of matrices, etc., in chapter 1 and then moves forward to more complex concepts, such as probability measures of semigroups, convolution products of probabilities and convergence, random walks on semigroups, random walks on semigroups of nonnegative matrices (and in particular stochastic matrices), etc. The current author collaborated on a few papers in this area [2,3,4,5,6].
For complete understanding of this article, we will go over a few details about convergence of convolution products of probability measures on semigroups of matrices. If B denotes the collection of Borel subsets of a set S, then P ( S ) can be the set of all regular probability measures μ on B . Then, denoting the collection of continuous functions on S as C ( S ) , for μ , ν P ( S ) , and f C ( S ) , one defines the following iterated integral:
I ( f ) = f ( x y ) μ ( d x ) ν ( d y )
By the Riesz representation theorem, there exists a unique regular probability measure λ such that for any function f C ( S ) with compact support, we have
I ( f ) = f d λ
Then, λ is called the convolution of the probability measures μ and ν . There is a proposition in [1] that shows that for μ , ν P ( S ) , and B B ,
μ ν ( B ) = μ ( B x 1 ) ν ( d x ) = ν ( x 1 B ) μ ( d x )
Having defined the convolution product of regular probability measures on semigroups, one can consider a sequence of regular probability measures μ 1 , μ 2 , μ 3 , , construct a sequence of convolution products of these regular probability measures μ 1 , μ 1 μ 2 , μ 1 μ 2 μ 3 , , and talk about conditions when such convolution sequences will converge. Then, one can specialize to the independent identically distributed situation where for each i, we have, μ i = μ for i = 1 , 2 , 3 , . Then, the convolution sequence looks like μ n for n = 1 , 2 , 3 , . In all these situations, [1] assumes that S is a locally compact, second countable Hausdorff topological semigroup.
Then, if someone further specializes to the situation when S is a semigroup of nonnegative matrices or say, stochatic matrices of a fixed order d, then one considers the usual matrix topology. There have been quite a few papers that study the conditions when the convolution sequence μ n converges. Mukherjea [7] first gave conditions when such a sequence converges for i.i.d. 2 × 2 stochastic matrices. Then, subsequently such conditions for higher order stochastic matrices were obtained [5,6]. But none of these papers performed detailed study on the nature of the corresponding limiting measures. But motivated by a paper by Chamayou and Letac [8], we have investigated the nature of the limiting measure λ for a very special μ on 2 × 2 i.i.d. stochastic matrices.
Before proceeding further, let us denote the probability measure on stochastic matrices of a fixed order d by μ and its support by S ( μ ) . So, S ( μ ) is a subcollection of stochastic matrices of a fixed order d. Thus, for any convolution product μ n , we will denote its support by S ( μ n ) and the support of the limiting measure λ (if it exists) by S ( λ ) .
If we denote the closure of an arbirary set E by E ¯ , then
S ( μ n ) = A 1 A 2 A n | for   each i , A i S ( μ ) , 1 i n ¯
where n is a positive integer and
S = n = 1 S ( μ n ) ¯
Also, denote P to be the set of d × d strictly positive stochastic matrices in S .
Chamayou and Letac [8] proved that if ( X n ) n 1 is a sequence of d × d i.i.d. stochastic matrices such that P ( min i , j ( X 1 ) i j = 0 ) < 1 , then Y = lim n X n X n 1 X 1 exists almost surely and P ( Y has   rank 1 ) = 1 ; furthermore, if for any Borel B of d × d stochastic matrices (with usual R d 2 -topology), we denote μ ( B ) = P ( X 1 B ) and λ ( B ) = P ( Y B ) , and then λ is the unique solution of the convolution equation λ μ = λ .
Then, in [2], we noted that this wonderful result of Chamayou and Letac also holds under the (slightly weaker) condition that μ m ( P ) > 0 for some positive integer m (as opposed to just 1, instead of m, taken in [8]) where μ m is the distribution of the product X m X 1 and P is the set of d × d strictly positive stochastic matrices. The reason is as follows: the Chamayou and Letac result shows that under the weaker condition, the subsequence Y n m = X n m X n m 1 X 1 converges almost surely to some d × d rank one stochastic matrix, Y 0 , and consequently, any subsequence X n k X n k 1 X 1 with n k > s k m (for some s k ) will also converge almost surely to a d × d stochastc matrix V Y 0 ( = Y 0 , as Y 0 has rank one), where V is a limit point of the product subsequence X n k X n k 1 X s k m + 1 . This establishes our observation.
Next we mention below some situations when S ( λ ) consists of all rank one matrices:
Situation 1: If ( X i ) i 1 , as before, is i.i.d. d × d stochastic matrices such that for some positive integer m 1 ,
μ m ( P ) > 0
then the sequence μ n , where μ ( B ) = P ( X 1 B ) for Borel sets B of d × d stochastic matrices, converges weakly to a probability measure λ and S ( λ ) consists of all rank one stochastic matrices in S such that λ ( P ) > 0 .
Situation 2: When λ is the weak limit of ( μ n ) n 1 and S contains a rank one matrix, then the support of λ , S ( λ ) consists of all rank one stochastic matrices in S . This is an algebraic fact for the support of an idempotent probability measure (note that λ = λ λ ; see [1]).
In the same paper, Chamayou and Letac (see also [9]) tried to identify λ in the case when the rows of X 1 above are independent, and for 1 i d , the i-th row of X 1 has Dirichlet distribution with positive parameters α i 1 , α i 2 , , α i d , and they were successful in the case when j = 1 d α i j = j = 1 d α j i , 1 i d . Indeed, there are only very few (other than those given in [8,9,10]) examples in the literature even for 2 × 2 stochastic matrices when the limit distribution λ has been identified completely in the above context. Our paper [2] is an example.
In [2], we considered 2 × 2 i.i.d. stochastic matrices ( X n ) n 1 with X n = C n 1 C n D n 1 D n , each X n is distributed as μ and r 1 C n and r 1 D n are each Bernoulli distributions (with possibly different parameters p 1 and p 2 , 0 < p 1 , p 2 < 1 ) for a real r satisfying 0 < r 1 . Our goal was to identify λ , the distribution of lim n X n X n 1 X 1 . Clearly, there are exactly four matrices in the support of μ , each with positive mass. It is well known that that μ n converges weakly to a limiting measure λ and the support of λ consists of rank one matrices. In particular, if r equals 1, the support of λ has exactly two matrices, namely, 0 1 0 1 and 1 0 1 0 . In [2], a complete solution is given to the problem for 0 < r 1 2 and also for r = 1 .
The situation 1 2 < r < 1 is much more challenging. Before explaining where the challenge lies, let us make the following convention:
From now on, we will often denote the matrix x 1 x x 1 x by simply x when there is no fear of confusion. Thus, for the limiting measure λ , λ ( x ) will mean λ x 1 x x 1 x and if we write that the support of λ , S ( λ ) is contained in [ 0 , 1 ] , then this means the following:
S ( λ ) x 1 x x 1 x : 0 x 1
Now, we are going to explain why the case 1 2 < r < 1 is more challenging. Although we find it quite easy to observe that λ ( 0 ) and λ ( r ) have the same expressions as in the previous case, it is indeed hard to exhibit a point in ( 0 , r ) with positive λ -mass.
However, there is a special situation when things are more tractable, namely, r = 5 1 2 (the reciprocal of the golden ratio). We denote this special r as r g . Notice that r g satisfies the equation r g 2 + r g 1 = 0 . Using this equation extensively, we completely solve for λ in this particular situation. It can be seen that although this is just one case, the proof is highly nontrivial. According to the author, the reason why r g works for us is because of the fact that λ ( 1 r g ) could be found out easily and so this technique of proof worked.
It may be mentioned here that there have been numerous studies in the literature involving the golden ratio. One very recent study invloving golden ratio is in the context of machine learning [11].
As in the case of 0 < r 1 2 , here also λ is discrete with masses at countably many points. Our main theorem appears in Section 4.
One gets a feeling that for any other r satisfying 1 2 < r < 1 , finding the value of λ ( 1 r ) itself will be a challenge, making it quite nontrivial. Thus, for a general 1 2 < r < 1 , a different technique of proof might be needed to obtain a complete solution.
In the next section (Section 2), we describe our set up, state the results proved in [2] for 0 < r 1 2 , and briefly discuss the more challenging situation 1 2 < r < 1 . In Section 3, we focus on r = r g = 5 1 2 (reciprocal of the golden ratio) and prove two important propostions. We prove our main theorem and a series of lemmas leading to it in Section 4. We have some concluding remarks and comments in Section 5.

2. Preliminaries

In our case, we are considering the case of a probability measure μ on 2 × 2 stochastic matrices. S ( μ ) denotes its support, which is a subcollection of 2 × 2 stochastic matrices. S ( μ n ) denotes the support of μ n where μ n is the convolution sequence. As pointed out in [7], μ n converges if and only if S ( μ ) is not a singleton:
S ( μ ) 0 1 1 0
And in case there is a strictly positive matrix in S ( μ ) , then the support S ( λ ) of the limiting measure λ consists of rank one matrices. Our special case satisfies that condition:
We consider 2 × 2 i.i.d. stochastic matrices ( X n ) n 1 with X n = C n 1 C n D n 1 D n , such that each X n is distributed as μ . Also, assume that for a given r with 0 < r 1 , both r 1 C n and r 1 D n are Bernoulli distributions with parameters p 1 and p 2 respectively.
Then, the support of μ has exactly four matrices as S ( μ ) is given by:
S ( μ ) = 0 1 0 1 , 0 1 r 1 r , r 1 r 0 1 , r 1 r r 1 r
Let the μ -masses at these points be denoted by p 00 , p 01 , p 10 , p 11 respectively so that p 00 + p 01 = 1 p 1 , p 00 + p 10 = 1 p 2 , p 10 + p 11 = p 1 and p 01 + p 11 = p 2 .
Let λ be the distribution of lim n X n X n 1 X 1 .
In case r equals 1, one can easily observe that λ is a Bernoulli distribution with parameters entirely dependent on the probability mass function of μ , namely,
λ ( 0 ) = p 00 + p 01 1 p 10 + p 01
This follows by solving for λ ( 0 ) and λ ( 1 ) in the convolution equation λ μ = λ .
For 0 < r < 1 , the support of μ n , S μ n and consequently S is contained in the set
x 1 x y 1 y : 0 x r , 0 y r
This can be proved using induction on n. One assumes up to some positive integer l and proves for l + 1 by noticing that when one multiplies a matrix in S ( μ l ) by a matrix in S ( μ ) , the entiries in the product matrix satisfies the condition that each entry in the first column is between 0 and r because each entry in the first column of the matrices from S ( μ l ) and S ( μ ) is so.
Also, since the relation λ μ = λ holds, the support of λ , namely, S ( λ ) consists of all rank one matrices in S . As a result, S ( λ ) x : 0 x r , where x stands for x 1 x x 1 x . Moreover, exploiting the identity λ μ = λ , we have
λ ( 0 ) = p 00 1 p 10 , λ r = p 11 + λ ( 0 ) p 01 = p 11 1 p 10 + p 00 p 01 1 p 10
and for other points x with 0 < x < r with positive λ -masses, we have
λ ( x ) = λ ( r 1 x ) p 10 + λ ( 1 r 1 x ) p 01
Next, we state the results proved in [2] for 0 < r 1 2 :

2.1. Case: 0 < r 1 2

First of all, we introduce some notations. For each i 1 , define
A i = j = 1 k ( 1 ) j 1 r i j : 1 i 1 < i 2 < i 3 < < i k = i , k i , A = i = 1 A i
We have two propositions for taking care of the cases 0 < r < 1 2 and r = 1 2 :
Proposition 1. 
For 0 < r < 1 2 , we have the following:
(i) 
For every positive integer i 1 , | A i | = 2 i 1 and each point in A i has positive λ-mass. These are the only points of degree i in the support of λ with positive λ-mass.
(ii) 
Each such point has λ-measure equal to λ r p 10 i 1 k p 01 k . For every i > 1 , λ A i = λ r p 10 + p 01 i 1 .
(iii) 
λ ( 0 ) + i = 1 λ A i = λ ( 0 ) + λ ( r ) · i = 1 p 10 + p 01 i 1 = 1 .
Proposition 2. 
For r = 1 2 , we have the following:
(i) 
The only points that have positive λ-masses are the dyadic rationals in [ 0 , 1 2 ] . Thus, for every i, there are exactly 2 i 2 dyadic rationals of the form k 2 i with k 2 i 1 and k odd with positive λ-mass. A i consists of exactly these points. Also, | A i | = 2 i 2 .
(ii) 
A typical point in A i has λ-measure equal to λ 1 2 p 10 + p 01 p 10 i 1 k p 01 k 1 for some positive integer k. For every i > 1 , λ A i = λ 1 2 p 10 + p 01 i 1 .
(iii) 
The sum of the λ-masses of all dyadic rationals in 0 , 1 2 along with the λ-mass at zero equals 1. Equivalently, λ ( 0 ) + i = 1 λ A i = λ ( 0 ) + λ 1 2 · i = 1 p 10 + p 01 i 1 = 1
The case 1 2 < r < 1 turns out to be quite nontrivial. We briefly introduce that case below:

2.2. Case: 1 2 < r < 1

The case 1 2 < r < 1 is distinctly different from the case r < 1 2 because now we have 1 r < r . Since for each r, λ has masses at 0 and r, it is not absolutely continuous for any r. Now, suppose we continue with the same notation of A introduced in the case 0 < r 1 2 . Thus, A = i = 1 A i where, for every positive integer i,
A i = j = 1 k ( 1 ) j 1 r i j : 1 i 1 < i 2 < i 3 < < i k = i , k i
It then easily follows that each of these points in A also has positive mass even in the case 1 2 < r < 1 . However, it is indeed a challenge to calculate λ -masses at these points.
Also, since 1 r ( 0 , r ) , it is natural to have points of the form 1 + j = 1 k ( 1 ) j r i j , 1 i 1 < i 2 < i 3 < < i k = i , k i for any positive integer i in the interval ( 0 , r ) (to see this, notice that r i 1 > j = 2 k ( 1 ) j r i j ). Accordingly, define A = i = 1 A i , where
A i = 1 + j = 1 k ( 1 ) j r i j : 1 i 1 < i 2 < i 3 < < i k = i , k i
Recall that, for 0 < r 1 2 , each point in A has positive λ -mass and each point in A is outside ( 0 , r ) and has zero λ -mass.
For 1 2 < r < 1 , of course, each polynomial in A is in ( 0 , r ) . But, although some polynomials in A are numerically less than r, it is not easy to see which of these points have positive λ -masses. Clearly, some polynomials in A i are outside ( 0 , r ) and have zero λ -measure if i is large enough. For example, for a fixed r, it is possible to get a positive integer m > 1 such that 1 r m r > 1 r m 1 . Next, consider i 1 = l with l m for a polynomial 1 + 1 k ( 1 ) j r i j in A i with 1 i 1 < i 2 < i 3 < < i k = i , k i . Then, this polynomial is greater than or equal to 1 r m + 2 k ( 1 ) j r i j , which is obviously greater than r and has λ -measure zero. But, it is a possibility that some points in A could have positive λ -masses too.
Recall the very special r, r = r g , the reciprocal of the golden ratio. We know r g satisfies the equation r g 2 + r g 1 = 0 and 1 r g actually equals r g 2 , whose λ -measure can be found out easily. The next two sections deal with this special case.

3. r = r g : Main Results

In this section and also in the next section, we deal with r = r g = 5 1 2 unless stated otherwise. This is a very special case of 1 2 < r < 1 . Note that r g is the reciprocal of the golden ratio and is the positive solution of the equation r 2 + r 1 = 0 . To avoid dealing with too many radical signs and complicating matters, we will continue to use r g in these two sections for this particluar choice of r.
Remark 1. 
A polynomial 1 + 1 k ( 1 ) j r g i j in A i with 1 i 1 < i 2 < i 3 < < i k = i , k i and i 1 2 has zero λ-measure.
This is because, 1 r g 2 = r g implies that such a polynomial is greater than r g in magnitude. However, for i 1 = 1 , such a polynomial may have positive λ -measure as well.
In order to notice this, first observe that, λ ( 1 r g ) > 0 . This is because, using (2), we have
λ ( 1 r g ) = λ ( r g 2 ) = λ ( r g ) p 10 + λ ( 1 r g ) p 01
implying that λ ( 1 r g ) = p 10 1 p 01 λ ( r g ) where λ ( r g ) is already known.
Next, consider a nontrivial example, say, the polynomial 1 r g + r g 2 r g 3 . Using (2) repeatedly and Remark 2, we find that its λ - measure equals
λ ( r g 2 ) p 10 2 p 01 + λ ( 1 r g 2 ) p 10 p 01 2 = λ ( 1 r g ) p 10 2 p 01 + λ ( r g ) p 10 p 01 2
implying that the polynomial under consideration has non-zero λ -measure. Since we know λ ( r g ) and λ ( 1 r g ) , it is possible to find out λ ( 1 r g + r g 2 r g 3 ) explicitly.
But, this is only a particular example. Can we make a general observation? Yes. Look at the following result.
Proposition 3. 
Any polynomial in A either has λ-measure 0 or can be written as a polynomial in A.
Proof. 
To fix ideas, we assume that our polynomial in A is 1 + j = 1 k ( 1 ) j r g i j with 1 i 1 < i 2 < i 3 < < i k = i and k i . Because of Remark 3.0, we can assume that i 1 = 1 . Then, we consider the following cases:
Case 1:  i j = j for j = 2 , 3 , , k .
Then, the given polynomial equals 1 r g + r g 2 + ( 1 ) k r g k
Subcase 1: k is even, say, k = 2 m . Then, the above polynomial equals 1 r g + r g 2 + r g 2 m . Notice that r g j r g j + 1 = r g j + 1 r g j + 3 for j = 0 , 1 , 2 , . Thus, the given polynomial equals r g r g 3 + r g 3 r g 5 + + r g 2 m 1 r g 2 m + 1 + r g 2 m which equals r g r g 2 m + 1 + r g 2 m > r g . So, it has λ -measure 0.
Subcase 2: k is odd, k = 2 m + 1 . Then, the above polynomial equals 1 r g + r g 2 + r g 2 m r g 2 m + 1 . Once again recall that r g j r g j + 1 = r g j + 1 r g j + 3 for j = 0 , 1 , 2 , . So, the given polynoimal equals r g r g 3 + r g 3 r g 5 + + r g 2 m 1 r g 2 m + 1 + r g 2 m + 1 r g 2 m + 3 = r g r g 2 m + 3 . And it is a polynomial in A.
Case 2: There exists an l such that i l > l and i j = j for j < l . Then, the given polynomial equals 1 r g + r g 2 + ( 1 ) l 1 r g l 1 + j = l k ( 1 ) j r g i j .
Subcase 1: l is even, say, l = 2 m . Then, the polynomial equals 1 r g + r g 2 r g 2 m 1 + j = 2 m k ( 1 ) j r g i j . Again, we use r g j r g j + 1 = r g j + 1 r g j + 3 for j = 0 , 1 , 2 , so that the given polynomial equals r g r g 3 + r g 3 r g 5 + + r g 2 m 1 r g 2 m + 1 + j = 2 m k ( 1 ) j r g i j . If i 2 m = 2 m + 1 , then this polynomial equals r g r g 3 + r g 3 r g 5 + + r g 2 m 1 r g 2 m + 1 + r g 2 m + 1 + j = 2 m + 1 k ( 1 ) j r g i j , which equals r g + j = 2 m + 1 k ( 1 ) j r g i j . This is, of course, a polynomial in A. On the other hand, if i 2 m > 2 m + 1 , then the above polynomial equals r g r g 2 m + 1 + j = 2 m k ( 1 ) j r g i j . Once again, it is a polynomial in A.
Subcase 2: l is odd, say, l = 2 m + 1 . Then, the given polynomial equals 1 r g + r g 2 r g 2 m 1 + r g 2 m + j = 2 m + 1 k ( 1 ) j r g i j . Applying once again r g j r g j + 1 = r g j + 1 r g j + 3 for j = 0 , 1 , 2 , , this polynomial equals r g r g 3 + r g 3 r g 5 + + r g 2 m 1 r g 2 m + 1 + r g 2 m + j = 2 m + 1 k ( 1 ) j r g i j . This simplifies to r g r g 2 m + 1 + r g 2 m + j = 2 m + 1 k ( 1 ) j r g i j = r g + r g 2 m + 2 + j = 2 m + 1 k ( 1 ) j r g i j . If i 2 m + 1 = 2 m + 2 , then the above equals r g + j = 2 m + 2 k ( 1 ) j r g i j > r g . So, it has λ -measure zero. If i 2 m + 1 > 2 m + 2 , then the given polynomial equals r g + r g 2 m + 2 + j = 2 m + 1 k ( 1 ) j r g i j which is same as r g + r g 2 m + 2 r g i 2 m + 1 + j = 2 m + 2 k ( 1 ) j r g i j > r g . So, it has λ -measure equal to zero. □
Remark 2. 
Because of the above proposition, it is good enough to consider only polynomials in A. We will rather consider the same polynomials as in the case 0 < r < 1 2 and will try to work out their λ-measures.
We have seen in Section 2 that the number of elements in A n equals 2 n 1 . But, because of the relationship 1 r g = r g 2 in the current situation, there will be redundancy and all polynomials are not distinct. So, we will see that we need to consider at most 2 n 2 elements from A n for each n 3 :
Proposition 4. 
There are at most 2 n 2 distinct elements in A n for each n 3 .
Proof. 
Once again, the identity 1 r g = r g 2 has a big role to play. For n = 1 , 2 , 3 or 4, it is trivial to observe. For general n, first notice that r g n = r g n 2 r g n 1 , and so r g n can be considered to be in A n 1 . More generally, define
Q n = r g n j = 1 k ( 1 ) j 1 r g i j : 1 i 1 < < i k 1 < i k ; i k 1 < n , i k = n ; k < n ; n i k 1 2
and
R n = j = 1 k ( 1 ) j 1 r g i j : 1 i 1 < < i k 1 < i k ; i k 1 = n 1 , i k = n , k n
Let Q = n = 3 Q n and R = n = 3 R n . Then, observe that each polynomial in Q is numerically equal to a polynomial in R of less degree.
We see this as follows:
Consider an n > 2 . Take a polynomial in Q n . If it is r g n , we have already provided the argument, that is, r g n = r g n 2 r g n 1 R n 1 . Otherwise, consider a typical element from Q n , say, r g i 1 r g i 2 + + ( 1 ) k 1 r g i k 1 + ( 1 ) k r g n with 1 i 1 < i 2 < < n and for some k < n . If n i k 1 = 2 , then r g i k 1 r g n = r g n 2 r g n = r g n 1 = r g n 3 r g n 2 . As a result, the given polynomial equals r g i 1 r g i 2 + + + ( 1 ) k 1 r g n 3 + ( 1 ) k r g n 2 . So, it is a polynomial in R of less degree ( n 2 ). On the other hand, if n i k 1 > 2 , then r g i k 1 r g n = r g i k 1 r g n 2 + r g n 1 . So, the given polynomial equals r g i 1 r g i 2 + + ( 1 ) k 1 r g i k 1 + ( 1 ) k r g n 2 + ( 1 ) k + 1 r g n 1 . Once again, this is a polynomial in R of less degree ( n 1 ).
It is clear that for each n, A n = Q n R n and hence A = Q R . So, because of this observation, the only polynomials in A that can be considered for λ -mass calculation are the ones in R. Also, it follows that for n 3 , R n has at most 2 n 2 distinct polynomials. Consequently, A n also has at the most 2 n 2 distinct elements and the proposition follows. □
Remark 3. 
Thus, for each n, we have fewer polynomials of degree n compared to the situation 0 < r < 1 2 .
Now, it is time we prove our main theorem. We prove it in the next section.

4. r = r g : Proof of the Main Theorem

Here is our main theorem:
Theorem 1. 
Consider r = r g = 5 1 2 . Then
λ ( 0 ) + λ ( r g ) + λ ( r g 2 ) + λ ( r g r g 2 ) + λ ( R ) = 1
where R = n = 3 R n with
R n = j = 1 k ( 1 ) j 1 r g i j : 1 i 1 < i 2 < < i k 2 < i k 1 < i k ; i k 1 = n 1 , i k = n , k n
First, notice that, using (2), it follows that λ ( r g 2 ) = p 10 1 p 01 λ ( r g ) and λ ( r g r g 2 ) = p 10 2 1 p 01 + p 01 λ ( r g ) . Thus, in order to prove the theorem, it is enough to prove:
λ ( R ) = p 10 + p 01 1 p 10 p 01 λ ( r g r g 2 )
because then,
λ ( r g r g 2 ) + λ ( R ) = 1 + p 10 + p 01 1 p 10 p 01 λ ( r g r g 2 )
As a result,
λ ( 0 ) + λ ( r g ) + λ ( r g 2 ) + λ ( r g r g 2 ) + λ ( R ) = λ ( 0 ) + 1 1 p 10 p 01 λ ( r g )
But, recall from Section 3:
λ ( 0 ) = p 00 1 p 10 , λ r g = p 11 + λ ( 0 ) p 01 = p 11 1 p 10 + p 00 p 01 1 p 10
This implies that
λ ( 0 ) + 1 1 p 10 p 01 λ ( r g ) = 1
This is the reason that it is good enough to prove (3). For this, we proceed as follows.
First of all, notice that R 3 = { r g 2 r g 3 , r g r g 2 + r g 3 } , R 4 = { r g 3 r g 4 , r g 2 r g 3 + r g 4 , r g r g 3 + r g 4 , r g r g 2 + r g 3 r g 4 } etc. and in general
R n = { r g n 1 r g n , r g n 2 r g n 1 + r g n , , r g r g n 1 + r g n , , r g r g 2 + + r g n 1 r g n }
Next, we introduce some notations for any 0 < r < 1 .
Define g : R R and f j : R R for every positive integer j as follows: g ( p ) = r p and f j ( p ) = r j p . Thus, R 3 = { f 2 r 3 , f 1 f 2 r 3 } , R 4 = { f 3 r 4 , f 2 f 3 r 4 , f 1 f 3 r 4 , f 1 f 2 f 3 r 4 } etc., and in general,
R n = { f n 1 r n , f n 2 f n 1 r n , , f 1 f n 1 r n , , f 1 f 2 f n 1 r n }
We further define operators F j for j 2 on R as follows: F 2 = { f 2 , f 1 f 2 } , F 3 = { f 3 , f 2 f 3 , f 1 f 3 , f 1 f 2 f 3 } etc., and in general,
F n 1 = { f n 1 , f n 2 f n 1 , , f 1 f n 1 , , f 1 f 2 f n 1 }
Thus, R j = F j 1 r j for j = 3 , 4 , and
F j 1 ( p ) = { f j 1 ( p ) , f j 2 f j 1 ( p ) , , f 1 f j 1 ( p ) , , f 1 f 2 f j 1 ( p ) }
In general, one would anticipate | F 2 ( p ) | = 2 , | F 3 ( p ) | = 4 , , | F j 1 ( p ) | = 2 j 2 . But, for r = r g , equality is replaced by ≤ for some ps.
Now, in order to prove (3), we will use a series of Lemmas 1–5. Lemma 1 identifies that connsecutive R i s have nonempty overlaps for i 3 , Lemma 2 evaluates the cardinality of the consecutive overlaps, Lemma 3 evaluates the cardinality of the consecutive differences, Lemma 4 calculates the λ - measures of these differences, and, finally, Lemma 5 puts them together to evaluate the λ -measure of R thereby proving (3). Thus, once Lemmas 1–5 are proved, (3) is proved and the proof of the theorem is complete.
Lemma 1. 
Consecutive R i s ( R i and R i + 1 ) have nonempty intersections for i 3 . In fact, R 4 R 3 = but R j + 1 R j for j > 3
Proof. 
It is trivial to observe that R 4 R 3 = ϕ . Now, notice that r g 2 r g 4 + r g 5 , r g r g 2 + r g 4 r g 5 R 5 R 4 because r g 2 r g 4 + r g 5 = r g r g 2 + r g 5 = r g r g 2 + r g 3 r g 4 R 4 and automatically, r g r g 2 + r g 4 r g 5 = r g ( r g r g 2 + r g 3 r g 4 ) = r g 2 r g 3 + r g 4 R 4 . Thus, R 5 R 4 = F 2 ( r g 4 r g 5 ) and | R 5 R 4 | = 2 . In general, R j + 1 R j F j 2 ( r g j r g j + 1 ) F j 4 ( r g j r g j + 1 ) for j 6 . In fact, we can show that for positive integers k 3
R 2 k 1 R 2 k 2 = F 2 k 4 ( r g 2 k 2 r g 2 k 1 ) F 2 k 6 ( r g 2 k 2 r g 2 k 1 ) F 2 ( r g 2 k 2 r g 2 k 1 )
R 2 k R 2 k 1 = F 2 k 3 ( r g 2 k 1 r g 2 k ) F 2 k 5 ( r g 2 k 1 r g 2 k ) F 3 ( r g 2 k 1 r g 2 k )
So, Lemma 1 is proved. □
Lemma 2. 
For i 4 , | R i R i + 1 | s are evaluated upper bounds for | R i + 1 R i | are determined as follows:
For k 3 , we have,
| R 2 k 1 R 2 k 2 | = 2 3 ( 2 2 k 4 1 ) , | R 2 k R 2 k 1 | = 4 3 ( 2 2 k 4 1 )
so that
| R 2 k 1 R 2 k 2 | 2 2 k 2 + 2 3 , | R 2 k R 2 k 1 | 2 2 k 1 + 4 3
Proof. 
From Lemma 1, it follows that | R 5 R 4 | = 2 implying | R 5 R 4 | 2 3 2 = 6 , | R 6 R 5 | = 4 implying | R 6 R 5 | 2 4 4 = 12 .
In general, notice that for k 4 ,
| F 2 k 2 l ( r g 2 k 2 l + 2 r g 2 k 2 l + 3 ) | = 2 2 k 2 l 1
and
| F 2 k 2 l + 1 ( r g 2 k 2 l + 3 r g 2 k 2 l + 4 ) | = 2 2 k 2 l
for 2 l k 1 . Also,
| R 2 k 1 R 2 k 2 | = 2 2 k 5 + 2 2 k 7 + + 2 = 2 3 2 2 k 4 1
implying that | R 2 k 1 R 2 k 2 | 2 2 k 3 2 3 2 2 k 4 1 = 2 2 k 2 + 2 3 and
| R 2 k R 2 k 1 | = 2 2 k 4 + 2 2 k 6 + + 2 2 = 4 3 2 2 k 4 1
implying that | R 2 k R 2 k 1 | 2 2 k 2 4 3 2 2 k 4 1 = 2 2 k 1 + 4 3 .
Thus, Lemma 2 is proved. □
Lemma 3. 
For i 4 , | R i + 1 R i | s are evaluated exactly by getting rid of the redundancies:
More explicitly, for j 6 , not all elements in R j are distinct. In fact, for k 3 , R 2 k has 2 2 k 4 2 and R 2 k + 1 has 2 2 k 3 2 pairs of elements which are numerically equal so that
| R 2 k R 2 k 1 | = 5 · 2 2 k 4 + 10 3 , | R 2 k + 1 R 2 k | = 5 · 2 2 k 3 + 8 3
Proof. 
From now on, we refer to duplicates as those pairs of polynomials or elements in R which have different algebraic expressions, but becaue of our choice of r, they are numerically equal. In order to exactly evaluate | R i + 1 R i | for i 4 , we need to identify such pairs.
Thus, R 6 R 5 has two pairs of duplicates, namely, r g 2 r g 5 + r g 6 & r g r g 2 + r g 4 r g 5 + r g 6 ; r g r g 2 + r g 5 r g 6 and r g 2 r g 4 + r g 5 r g 6 because
r g r g 2 + r g 4 r g 5 + r g 6 = r g 2 r g 4 + r g 4 r g 5 + r g 6 = r g 2 r g 5 + r g 6
r g r g 2 + r g 5 r g 6 = r g 2 r g 4 + r g 5 r g 6
In general, R 2 k R 2 k 1 has 2 + 4 + + 2 2 k 5 pairs of duplicates implying that | R 2 k R 2 k 1 | = 2 2 k 1 + 4 3 ( 2 2 k 4 2 ) = 5 · 2 2 k 4 + 10 3 . Here, each pair in the union are disjoint sets.
Also, R 2 k + 1 R 2 k has 2 + 4 + + 2 2 k 5 pairs of duplicates implying that | R 2 k + 1 R 2 k | = 2 2 k + 2 3 ( 2 2 k 3 2 ) = 5 · 2 2 k 3 + 8 3 . Again, each pair in the union are disjoint sets.
Thus, Lemma 3 is proved. □
Lemma 4. 
λ-measures of R i + 1 R i for i 3 are calculated as:
First of all, λ ( R 3 ) = λ ( r g r g 2 ) p 10 + p 01 and for k 2 ,
λ R 2 k R 2 k 1 = λ R 2 k 1 R 2 k 2 p 10 + p 01 + λ ( r g r g 2 ) p 10 2 k 3 p 01 p 10 + p 01
which equals
λ R 2 k R 2 k 1 = λ r g r g 2 p 10 + p 01 2 k 2 1 + l = 0 k 2 p 10 2 l + 1 p 01 p 10 + p 01 2 l 1 l = 1 k 3 p 10 2 l + 1 p 01 p 10 + p 01 2 l 2
where for k = 2 , the last sum in the above equation is absent. Also, we have,
λ R 2 k + 1 R 2 k = λ R 2 k R 2 k 1 p 10 + p 01 λ r g r g 2 p 10 2 k 3 p 01 p 10 + p 01
which equals
λ R 2 k + 1 R 2 k = λ r g r g 2 p 10 + p 01 2 k 1 1 + l = 0 k 2 p 10 2 l + 1 p 01 p 10 + p 01 2 l 1 p 10 + p 01 2 l 2
where for k = 2 , R 2 k R 2 k 1 = R 4 R 3 = R 4 and R 2 k 1 R 2 k 2 = R 3 R 2 = R 3 .
Proof. 
Recall that R 3 = F 2 ( r g 3 ) = { r g 2 r g 3 , r g r g 2 + r g 3 } . Then, using (2) and Proposition 3, we have
λ ( R 3 ) = λ ( r g 2 r g 3 ) + λ ( r g r g 2 + r g 3 ) = λ ( r g r g 2 ) ( p 10 + p 01 )
Next, we have R 4 = F 3 r g 4 = { r g 3 r g 4 , r g 2 r g 3 + r g 4 , r g r g 3 + r g 4 , r g r g 2 + r g 3 r g 4 } . We find λ -measures of these points by making use of (2) and Remark 2. Thus, we notice that
λ ( 1 r g + r g 2 r g 3 ) = λ ( r g r g 3 + r g 4 ) = λ ( r g r g 2 ) p 10 p 01
Putting all these together, λ ( R 4 R 3 ) equals
λ ( r g r g 2 ) ( p 10 + p 01 ) 2 + λ ( 1 r g + r g 2 r g 3 ) ( p 10 + p 01 ) = λ ( r g r g 2 ) ( p 10 + p 01 ) 2 + λ ( r g r g 2 ) p 10 p 01 ( p 10 + p 01 )
In other words,
λ R 4 R 3 = λ R 3 p 10 + p 01 + λ ( r g r g 2 ) p 10 p 01 ( p 10 + p 01 )
It is to be noted that R 4 R 3 = , and so R 4 R 3 = R 4 which implies λ R 4 R 3 = λ R 4 .
Before proceeding further, we notice that F j r g j + 1 = g F j 1 ( r g j ) f 1 g F j 1 ( r g j ) for j 4 and R j + 1 R j equals g R j R j 1 f 1 g R j R j 1 for j 5 .
However, at the next stage, we have already noticed that R 5 R 4 , and so R 5 R 4 R 5 . In fact,
R 5 = F 4 r g 5
and
R 5 R 4 = F 4 r g 5 F 2 r g 4 r g 5
Now, notice that
F 4 r g 5 = g R 4 R 3 f 1 g R 4 R 3
So,
R 5 R 4 = g R 4 R 3 f 1 g R 4 R 3 F 2 r g 4 r g 5
This is the same as
g R 4 R 3 g r g r g 3 + r g 4 f 1 g R 4 R 3 f 1 g r g r g 3 + r g 4
Since g R 4 R 3 g r g r g 3 + r g 4 and f 1 g R 4 R 3 f 1 g r g r g 3 + r g 4 do not have overlaps, we deduce that
λ R 5 R 4 = λ R 4 R 3 p 10 + p 01 λ r g r g 2 p 10 p 01 p 10 + p 01
which equals
λ R 5 R 4 = λ r g r g 2 p 10 + p 01 3 + λ ( r g r g 2 ) p 10 p 01 ( p 10 + p 01 ) 2 λ ( r g r g 2 ) p 10 p 01 ( p 10 + p 01 )
Thus, from Equations (6) and (8), we observe that Lemma 4 is proved for k = 2 . For general k, one can use induction on k and carefully sort out the issues with the duplicates to complete the proof of the lemma. □
Lemma 5. 
Finally, we calculate λ-measure of R:
λ R = j = 3 λ R j R j 1 = j = 1 λ r g r g 2 p 10 + p 01 j = λ r g r g 2 · p 10 + p 01 1 p 10 p 01
where we put R 2 = .
Proof. 
Using (4) and (5) for k 2 , Lemma 5 follows trivially and the proof of the theorem is complete. □

5. Concluding Remarks

In the present context, it is interesting to recall an older problem, first introduced in [7]. It is as follows: consider the very simple situation of a μ that is supported on exactly two 2 × 2 stochastic matrices, namely, a 1 1 a 1 b 1 1 b 1 and a 2 1 a 2 b 2 1 b 2 with a i > b i for i = 1 , 2 . Let the μ -masses at these two points be p and 1 p , respectively, where 0 < p < 1 . Let λ be the weak limit of the convolution sequence μ n . What is the nature of λ ? If we denote a 1 b 1 = s and a 2 b 2 = t , then, in [12], some partial solution to this problem was mentioned. In the special case scenario when s = t and p = 1 2 , it was observed in [13] that it is precisely the case of Bernoulli convolutions. In fact, the following proposition is stated in [13]:
Proposition 5. 
Let μ be a probability measure giving equal mass to the matrices a 1 1 a 1 b 1 1 b 1 and a 2 1 a 2 b 2 1 b 2 with a i > b i for i = 1 , 2 . Let, say, a 1 b 1 = a 2 b 2 = t . Then, the limiting measure λ of the convolution sequence μ n is absolutely continuous (where the limt λ is identified as a probability on [ 0 , 1 ] ) iff the law of n = 0 t n ϵ n is absolutely continuous where ϵ n ’s are i.i.d. + 1 and 1 with equal probabilities.
Although the century old problem of Bernoulli convolutions was finally solved in [14], there had been a lot of previous studies at various times in different directions in spite of it being apparently a simple problem with μ concentrated on two points only. Thus, it is quite possible that under our current set up of μ being concentrated on four matrices with 1 2 < r < 1 , the problem may be at least as challenging as the Bernoulli convolution problem.
We bring in the context of Bernoulli convolutions here to make readers aware that for a nontrivial 1 2 < r < 1 , one needs to explore a number of ideas to proceed towards a complete solution for our problem.

Funding

This research received no external funding.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Acknowledgments

I acknowledge my affiliation, the University of Texas Rio Grande Valley, for allowing me to use the university office, university library, and university computers for conducting my research.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Hognas, G.; Mukherjea, A. Probability Measures on Semigroups, 2nd ed.; Springer: New York, NY, USA, 2011. [Google Scholar]
  2. Chakraborty, S.; Mukherjea, A. Limit distributions of random walks on stochastic matrices. Proc. Indian Acad. Sci. (Math Sci.) 2014, 124, 603–612. [Google Scholar] [CrossRef]
  3. Chakraborty, S. Cyclicity and weak convergence for convolution of measures on non-negative matrices. Sankhya Indian J. Stat. 2007, 69, 304–313. [Google Scholar]
  4. Chakraborty, S.; Mukherjea, A. Completely simple semigroups of real d×d matrices and real random walks. Contemp. Math. 2010, 516, 99–108. [Google Scholar]
  5. Chakraborty, S.; Rao, B.V. Convolution powers of probabilities on stochastic matrices of order 3. Sankhya A 1998, 60, 151–170. [Google Scholar]
  6. Chakraborty, S.; Rao, B.V. Convolution powers of probabilities on stochastic matrices. J. Theor. Probab. 2001, 14, 599–603. [Google Scholar] [CrossRef]
  7. Mukherjea, A. Limit Theorems: Stochastic matrices, ergodic markov chains and measures on semigroups. In Probability Analysis and Related Topics; Bharucha Reid, A.T., Ed.; Academic Press: New York, NY, USA, 1979; Volume 2, pp. 143–203. [Google Scholar]
  8. Chamayou, J.F.; Letac, G. A transient random walk on stochastic matrices with Dirichlet distributions. Ann. Prob. 1994, 22, 424–430. [Google Scholar] [CrossRef]
  9. Van Assche, W. Products of 2 × 2 stochastic matrices with random entries. J. Appl. Prob. 1986, 23, 1019–1024. [Google Scholar] [CrossRef]
  10. Chassaing, P.; Letac, G.; Mora, M. Brocot Sequences and Random Walks in SL(2,R); Springer: Berlin/Heidelberg, Germany, 1985; Volume 1034, pp. 37–50. [Google Scholar]
  11. Jaeger, S. The Golden Ratio in Machine Learning. In Proceedings of the IEEE Applied Imagery Pattern Recognition Workshop (AIPR), Washington, DC, USA, 12–14 October 2021; pp. 1–7. [Google Scholar]
  12. Mukherjea, A.; Tserpes, N.A. Measures on Topological Semigroups; Lecture Notes in Mathematics; Springer: Berlin/Heidelberg, Germany, 1976; Volume 547. [Google Scholar]
  13. Chakraborty, S.; Rao, B.V. Bernoulli Convolutions. In Mathematical Models for Bioengineering and Probabilistic Systems; Mishra, J.C., Ed.; Narosa Publishing House: New Delhi, India, 2005; pp. 380–404. [Google Scholar]
  14. Solomyak, B. On the random series ∑±λn (an Erdos problem). Ann. Math. 1995, 142, 611–625. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chakraborty, S. Limit Distributions of Products of Independent and Identically Distributed Random 2 × 2 Stochastic Matrices: A Treatment with the Reciprocal of the Golden Ratio. Mathematics 2023, 11, 4993. https://doi.org/10.3390/math11244993

AMA Style

Chakraborty S. Limit Distributions of Products of Independent and Identically Distributed Random 2 × 2 Stochastic Matrices: A Treatment with the Reciprocal of the Golden Ratio. Mathematics. 2023; 11(24):4993. https://doi.org/10.3390/math11244993

Chicago/Turabian Style

Chakraborty, Santanu. 2023. "Limit Distributions of Products of Independent and Identically Distributed Random 2 × 2 Stochastic Matrices: A Treatment with the Reciprocal of the Golden Ratio" Mathematics 11, no. 24: 4993. https://doi.org/10.3390/math11244993

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop