Abstract
Consider a sequence of i.i.d. stochastic matrices with each distributed as . This is described as follows. Let denote the first column of and for a given real r with , let and each be Bernoulli distributions with parameters and , respectively, and . Clearly, the weak limit of the sequence , namely , is known to exist, whose support is contained in the set of all rank one stochastic matrices. In a previous paper, we considered and obtained explicitly. We showed that is supported countably on many points, each with positive -mass. Of course, the case is tractable, but the case is very challenging. Considering the extreme nontriviality of this case, we stick to a very special such r, namely, (the reciprocal of the golden ratio), briefly mention the challenges in this nontrivial case, and completely identify for a very special situation.
MSC:
60B15
1. Introduction
As the title of the paper suggests, the reader can understand that this paper deals with a situation where one considers products of independent and identically distributed random stochastic matrices and their limiting behavior. In other words, here we are considering a probability measure on a collection of stochastic matrices and studying the limiting behavior of the convolution sequence . To a reader new to this area, the author would like to refer the reader to the book by Hognas and Mukherjea [1]. This book starts from the very basic concepts, such as the definition of a semigroup, topological semigroups, semigroups of matrices, etc., in chapter 1 and then moves forward to more complex concepts, such as probability measures of semigroups, convolution products of probabilities and convergence, random walks on semigroups, random walks on semigroups of nonnegative matrices (and in particular stochastic matrices), etc. The current author collaborated on a few papers in this area [2,3,4,5,6].
For complete understanding of this article, we will go over a few details about convergence of convolution products of probability measures on semigroups of matrices. If denotes the collection of Borel subsets of a set S, then can be the set of all regular probability measures on . Then, denoting the collection of continuous functions on S as , for , and , one defines the following iterated integral:
By the Riesz representation theorem, there exists a unique regular probability measure such that for any function with compact support, we have
Then, is called the convolution of the probability measures and . There is a proposition in [1] that shows that for , and ,
Having defined the convolution product of regular probability measures on semigroups, one can consider a sequence of regular probability measures , construct a sequence of convolution products of these regular probability measures , and talk about conditions when such convolution sequences will converge. Then, one can specialize to the independent identically distributed situation where for each i, we have, for . Then, the convolution sequence looks like for . In all these situations, [1] assumes that S is a locally compact, second countable Hausdorff topological semigroup.
Then, if someone further specializes to the situation when S is a semigroup of nonnegative matrices or say, stochatic matrices of a fixed order d, then one considers the usual matrix topology. There have been quite a few papers that study the conditions when the convolution sequence converges. Mukherjea [7] first gave conditions when such a sequence converges for i.i.d. stochastic matrices. Then, subsequently such conditions for higher order stochastic matrices were obtained [5,6]. But none of these papers performed detailed study on the nature of the corresponding limiting measures. But motivated by a paper by Chamayou and Letac [8], we have investigated the nature of the limiting measure for a very special on i.i.d. stochastic matrices.
Before proceeding further, let us denote the probability measure on stochastic matrices of a fixed order d by and its support by . So, is a subcollection of stochastic matrices of a fixed order d. Thus, for any convolution product , we will denote its support by and the support of the limiting measure (if it exists) by .
If we denote the closure of an arbirary set E by , then
where n is a positive integer and
Also, denote to be the set of strictly positive stochastic matrices in .
Chamayou and Letac [8] proved that if is a sequence of i.i.d. stochastic matrices such that , then exists almost surely and ; furthermore, if for any Borel B of stochastic matrices (with usual -topology), we denote and , and then is the unique solution of the convolution equation .
Then, in [2], we noted that this wonderful result of Chamayou and Letac also holds under the (slightly weaker) condition that for some positive integer m (as opposed to just 1, instead of m, taken in [8]) where is the distribution of the product and is the set of strictly positive stochastic matrices. The reason is as follows: the Chamayou and Letac result shows that under the weaker condition, the subsequence converges almost surely to some rank one stochastic matrix, , and consequently, any subsequence with (for some ) will also converge almost surely to a stochastc matrix , as has rank one), where V is a limit point of the product subsequence . This establishes our observation.
Next we mention below some situations when consists of all rank one matrices:
Situation 1: If , as before, is i.i.d. stochastic matrices such that for some positive integer ,
then the sequence , where for Borel sets B of stochastic matrices, converges weakly to a probability measure and consists of all rank one stochastic matrices in such that .
Situation 2: When is the weak limit of and contains a rank one matrix, then the support of , consists of all rank one stochastic matrices in . This is an algebraic fact for the support of an idempotent probability measure (note that ; see [1]).
In the same paper, Chamayou and Letac (see also [9]) tried to identify in the case when the rows of above are independent, and for , the i-th row of has Dirichlet distribution with positive parameters , and they were successful in the case when . Indeed, there are only very few (other than those given in [8,9,10]) examples in the literature even for stochastic matrices when the limit distribution has been identified completely in the above context. Our paper [2] is an example.
In [2], we considered i.i.d. stochastic matrices with , each is distributed as and and are each Bernoulli distributions (with possibly different parameters ) for a real r satisfying . Our goal was to identify , the distribution of . Clearly, there are exactly four matrices in the support of , each with positive mass. It is well known that that converges weakly to a limiting measure and the support of consists of rank one matrices. In particular, if r equals 1, the support of has exactly two matrices, namely, and . In [2], a complete solution is given to the problem for and also for .
The situation is much more challenging. Before explaining where the challenge lies, let us make the following convention:
From now on, we will often denote the matrix by simply x when there is no fear of confusion. Thus, for the limiting measure , will mean and if we write that the support of , is contained in , then this means the following:
Now, we are going to explain why the case is more challenging. Although we find it quite easy to observe that and have the same expressions as in the previous case, it is indeed hard to exhibit a point in with positive -mass.
However, there is a special situation when things are more tractable, namely, (the reciprocal of the golden ratio). We denote this special r as . Notice that satisfies the equation . Using this equation extensively, we completely solve for in this particular situation. It can be seen that although this is just one case, the proof is highly nontrivial. According to the author, the reason why works for us is because of the fact that could be found out easily and so this technique of proof worked.
It may be mentioned here that there have been numerous studies in the literature involving the golden ratio. One very recent study invloving golden ratio is in the context of machine learning [11].
As in the case of , here also is discrete with masses at countably many points. Our main theorem appears in Section 4.
One gets a feeling that for any other r satisfying , finding the value of itself will be a challenge, making it quite nontrivial. Thus, for a general , a different technique of proof might be needed to obtain a complete solution.
In the next section (Section 2), we describe our set up, state the results proved in [2] for , and briefly discuss the more challenging situation . In Section 3, we focus on (reciprocal of the golden ratio) and prove two important propostions. We prove our main theorem and a series of lemmas leading to it in Section 4. We have some concluding remarks and comments in Section 5.
2. Preliminaries
In our case, we are considering the case of a probability measure on stochastic matrices. denotes its support, which is a subcollection of stochastic matrices. denotes the support of where is the convolution sequence. As pointed out in [7], converges if and only if is not a singleton:
And in case there is a strictly positive matrix in , then the support of the limiting measure consists of rank one matrices. Our special case satisfies that condition:
We consider i.i.d. stochastic matrices with , such that each is distributed as . Also, assume that for a given r with , both and are Bernoulli distributions with parameters and respectively.
Then, the support of has exactly four matrices as is given by:
Let the -masses at these points be denoted by respectively so that , , and .
Let be the distribution of .
In case r equals 1, one can easily observe that is a Bernoulli distribution with parameters entirely dependent on the probability mass function of , namely,
This follows by solving for and in the convolution equation .
For , the support of , and consequently is contained in the set
This can be proved using induction on n. One assumes up to some positive integer l and proves for by noticing that when one multiplies a matrix in by a matrix in , the entiries in the product matrix satisfies the condition that each entry in the first column is between 0 and r because each entry in the first column of the matrices from and is so.
Also, since the relation holds, the support of , namely, consists of all rank one matrices in . As a result, , where x stands for . Moreover, exploiting the identity , we have
and for other points x with with positive -masses, we have
Next, we state the results proved in [2] for :
2.1. Case:
First of all, we introduce some notations. For each , define
We have two propositions for taking care of the cases and :
Proposition 1.
For , we have the following:
- (i)
- For every positive integer , and each point in has positive λ-mass. These are the only points of degree i in the support of λ with positive λ-mass.
- (ii)
- Each such point has λ-measure equal to . For every , .
- (iii)
- .
Proposition 2.
For , we have the following:
- (i)
- The only points that have positive λ-masses are the dyadic rationals in . Thus, for every i, there are exactly dyadic rationals of the form with and k odd with positive λ-mass. consists of exactly these points. Also, .
- (ii)
- A typical point in has λ-measure equal to for some positive integer k. For every , .
- (iii)
- The sum of the λ-masses of all dyadic rationals in along with the λ-mass at zero equals 1. Equivalently,
The case turns out to be quite nontrivial. We briefly introduce that case below:
2.2. Case:
The case is distinctly different from the case because now we have . Since for each r, has masses at 0 and r, it is not absolutely continuous for any r. Now, suppose we continue with the same notation of A introduced in the case . Thus, where, for every positive integer i,
It then easily follows that each of these points in A also has positive mass even in the case . However, it is indeed a challenge to calculate -masses at these points.
Also, since , it is natural to have points of the form for any positive integer i in the interval (to see this, notice that ). Accordingly, define , where
Recall that, for , each point in A has positive -mass and each point in is outside and has zero -mass.
For , of course, each polynomial in A is in . But, although some polynomials in are numerically less than r, it is not easy to see which of these points have positive -masses. Clearly, some polynomials in are outside and have zero -measure if i is large enough. For example, for a fixed r, it is possible to get a positive integer such that . Next, consider with for a polynomial in with . Then, this polynomial is greater than or equal to , which is obviously greater than r and has -measure zero. But, it is a possibility that some points in could have positive -masses too.
Recall the very special r, , the reciprocal of the golden ratio. We know satisfies the equation and actually equals , whose -measure can be found out easily. The next two sections deal with this special case.
3. : Main Results
In this section and also in the next section, we deal with unless stated otherwise. This is a very special case of . Note that is the reciprocal of the golden ratio and is the positive solution of the equation . To avoid dealing with too many radical signs and complicating matters, we will continue to use in these two sections for this particluar choice of r.
Remark 1.
A polynomial in with and has zero λ-measure.
This is because, implies that such a polynomial is greater than in magnitude. However, for , such a polynomial may have positive -measure as well.
In order to notice this, first observe that, . This is because, using (2), we have
implying that where is already known.
Next, consider a nontrivial example, say, the polynomial . Using (2) repeatedly and Remark 2, we find that its - measure equals
implying that the polynomial under consideration has non-zero -measure. Since we know and , it is possible to find out explicitly.
But, this is only a particular example. Can we make a general observation? Yes. Look at the following result.
Proposition 3.
Any polynomial in either has λ-measure 0 or can be written as a polynomial in A.
Proof.
To fix ideas, we assume that our polynomial in is with and . Because of Remark 3.0, we can assume that . Then, we consider the following cases:
Case 1: for .
Then, the given polynomial equals
Subcase 1: k is even, say, . Then, the above polynomial equals . Notice that for . Thus, the given polynomial equals which equals . So, it has -measure 0.
Subcase 2: k is odd, . Then, the above polynomial equals . Once again recall that for . So, the given polynoimal equals . And it is a polynomial in A.
Case 2: There exists an l such that and for . Then, the given polynomial equals .
Subcase 1: l is even, say, . Then, the polynomial equals . Again, we use for so that the given polynomial equals . If , then this polynomial equals , which equals . This is, of course, a polynomial in A. On the other hand, if , then the above polynomial equals . Once again, it is a polynomial in A.
Subcase 2: l is odd, say, . Then, the given polynomial equals . Applying once again for , this polynomial equals . This simplifies to . If , then the above equals . So, it has -measure zero. If , then the given polynomial equals which is same as . So, it has -measure equal to zero. □
Remark 2.
Because of the above proposition, it is good enough to consider only polynomials in A. We will rather consider the same polynomials as in the case and will try to work out their λ-measures.
We have seen in Section 2 that the number of elements in equals . But, because of the relationship in the current situation, there will be redundancy and all polynomials are not distinct. So, we will see that we need to consider at most elements from for each :
Proposition 4.
There are at most distinct elements in for each .
Proof.
Once again, the identity has a big role to play. For or 4, it is trivial to observe. For general n, first notice that , and so can be considered to be in . More generally, define
and
Let and . Then, observe that each polynomial in Q is numerically equal to a polynomial in R of less degree.
We see this as follows:
Consider an . Take a polynomial in . If it is , we have already provided the argument, that is, . Otherwise, consider a typical element from , say, with and for some . If , then . As a result, the given polynomial equals . So, it is a polynomial in R of less degree (). On the other hand, if , then . So, the given polynomial equals . Once again, this is a polynomial in R of less degree ().
It is clear that for each n, and hence . So, because of this observation, the only polynomials in A that can be considered for -mass calculation are the ones in R. Also, it follows that for , has at most distinct polynomials. Consequently, also has at the most distinct elements and the proposition follows. □
Remark 3.
Thus, for each n, we have fewer polynomials of degree n compared to the situation .
Now, it is time we prove our main theorem. We prove it in the next section.
4. : Proof of the Main Theorem
Here is our main theorem:
Theorem 1.
Consider . Then
where with
First, notice that, using (2), it follows that and . Thus, in order to prove the theorem, it is enough to prove:
because then,
As a result,
But, recall from Section 3:
This implies that
This is the reason that it is good enough to prove (3). For this, we proceed as follows.
First of all, notice that , etc. and in general
Next, we introduce some notations for any .
Define and for every positive integer j as follows: and . Thus, , etc., and in general,
We further define operators for on R as follows: , etc., and in general,
Thus, for and
In general, one would anticipate . But, for , equality is replaced by ≤ for some ps.
Now, in order to prove (3), we will use a series of Lemmas 1–5. Lemma 1 identifies that connsecutive s have nonempty overlaps for , Lemma 2 evaluates the cardinality of the consecutive overlaps, Lemma 3 evaluates the cardinality of the consecutive differences, Lemma 4 calculates the - measures of these differences, and, finally, Lemma 5 puts them together to evaluate the -measure of R thereby proving (3). Thus, once Lemmas 1–5 are proved, (3) is proved and the proof of the theorem is complete.
Lemma 1.
Consecutive s ( and ) have nonempty intersections for . In fact, but for
Proof.
It is trivial to observe that . Now, notice that because and automatically, . Thus, and . In general, for . In fact, we can show that for positive integers
So, Lemma 1 is proved. □
Lemma 2.
For , s are evaluated upper bounds for are determined as follows:
For , we have,
so that
Proof.
From Lemma 1, it follows that implying , implying .
In general, notice that for ,
and
for . Also,
implying that and
implying that .
Thus, Lemma 2 is proved. □
Lemma 3.
For , s are evaluated exactly by getting rid of the redundancies:
More explicitly, for , not all elements in are distinct. In fact, for , has and has pairs of elements which are numerically equal so that
Proof.
From now on, we refer to duplicates as those pairs of polynomials or elements in R which have different algebraic expressions, but becaue of our choice of r, they are numerically equal. In order to exactly evaluate for , we need to identify such pairs.
Thus, has two pairs of duplicates, namely, & ; and because
In general, has pairs of duplicates implying that . Here, each pair in the union are disjoint sets.
Also, has pairs of duplicates implying that . Again, each pair in the union are disjoint sets.
Thus, Lemma 3 is proved. □
Lemma 4.
λ-measures of for are calculated as:
First of all, and for ,
which equals
where for , the last sum in the above equation is absent. Also, we have,
which equals
where for , and .
Proof.
Recall that . Then, using (2) and Proposition 3, we have
Next, we have . We find -measures of these points by making use of (2) and Remark 2. Thus, we notice that
Putting all these together, equals
In other words,
It is to be noted that , and so which implies .
Before proceeding further, we notice that for and equals for .
However, at the next stage, we have already noticed that , and so . In fact,
and
Now, notice that
So,
This is the same as
Since and do not have overlaps, we deduce that
which equals
Lemma 5.
Finally, we calculate λ-measure of R:
where we put .
5. Concluding Remarks
In the present context, it is interesting to recall an older problem, first introduced in [7]. It is as follows: consider the very simple situation of a that is supported on exactly two stochastic matrices, namely, and with for . Let the -masses at these two points be p and , respectively, where . Let be the weak limit of the convolution sequence . What is the nature of ? If we denote and , then, in [12], some partial solution to this problem was mentioned. In the special case scenario when and , it was observed in [13] that it is precisely the case of Bernoulli convolutions. In fact, the following proposition is stated in [13]:
Proposition 5.
Let μ be a probability measure giving equal mass to the matrices and with for . Let, say, . Then, the limiting measure λ of the convolution sequence is absolutely continuous (where the limt λ is identified as a probability on ) iff the law of is absolutely continuous where ’s are i.i.d. and with equal probabilities.
Although the century old problem of Bernoulli convolutions was finally solved in [14], there had been a lot of previous studies at various times in different directions in spite of it being apparently a simple problem with concentrated on two points only. Thus, it is quite possible that under our current set up of being concentrated on four matrices with , the problem may be at least as challenging as the Bernoulli convolution problem.
We bring in the context of Bernoulli convolutions here to make readers aware that for a nontrivial , one needs to explore a number of ideas to proceed towards a complete solution for our problem.
Funding
This research received no external funding.
Data Availability Statement
No new data were created or analyzed in this study. Data sharing is not applicable to this article.
Acknowledgments
I acknowledge my affiliation, the University of Texas Rio Grande Valley, for allowing me to use the university office, university library, and university computers for conducting my research.
Conflicts of Interest
The author declares no conflict of interest.
References
- Hognas, G.; Mukherjea, A. Probability Measures on Semigroups, 2nd ed.; Springer: New York, NY, USA, 2011. [Google Scholar]
- Chakraborty, S.; Mukherjea, A. Limit distributions of random walks on stochastic matrices. Proc. Indian Acad. Sci. (Math Sci.) 2014, 124, 603–612. [Google Scholar] [CrossRef]
- Chakraborty, S. Cyclicity and weak convergence for convolution of measures on non-negative matrices. Sankhya Indian J. Stat. 2007, 69, 304–313. [Google Scholar]
- Chakraborty, S.; Mukherjea, A. Completely simple semigroups of real d×d matrices and real random walks. Contemp. Math. 2010, 516, 99–108. [Google Scholar]
- Chakraborty, S.; Rao, B.V. Convolution powers of probabilities on stochastic matrices of order 3. Sankhya A 1998, 60, 151–170. [Google Scholar]
- Chakraborty, S.; Rao, B.V. Convolution powers of probabilities on stochastic matrices. J. Theor. Probab. 2001, 14, 599–603. [Google Scholar] [CrossRef]
- Mukherjea, A. Limit Theorems: Stochastic matrices, ergodic markov chains and measures on semigroups. In Probability Analysis and Related Topics; Bharucha Reid, A.T., Ed.; Academic Press: New York, NY, USA, 1979; Volume 2, pp. 143–203. [Google Scholar]
- Chamayou, J.F.; Letac, G. A transient random walk on stochastic matrices with Dirichlet distributions. Ann. Prob. 1994, 22, 424–430. [Google Scholar] [CrossRef]
- Van Assche, W. Products of 2 × 2 stochastic matrices with random entries. J. Appl. Prob. 1986, 23, 1019–1024. [Google Scholar] [CrossRef]
- Chassaing, P.; Letac, G.; Mora, M. Brocot Sequences and Random Walks in SL(2,R); Springer: Berlin/Heidelberg, Germany, 1985; Volume 1034, pp. 37–50. [Google Scholar]
- Jaeger, S. The Golden Ratio in Machine Learning. In Proceedings of the IEEE Applied Imagery Pattern Recognition Workshop (AIPR), Washington, DC, USA, 12–14 October 2021; pp. 1–7. [Google Scholar]
- Mukherjea, A.; Tserpes, N.A. Measures on Topological Semigroups; Lecture Notes in Mathematics; Springer: Berlin/Heidelberg, Germany, 1976; Volume 547. [Google Scholar]
- Chakraborty, S.; Rao, B.V. Bernoulli Convolutions. In Mathematical Models for Bioengineering and Probabilistic Systems; Mishra, J.C., Ed.; Narosa Publishing House: New Delhi, India, 2005; pp. 380–404. [Google Scholar]
- Solomyak, B. On the random series ∑±λn (an Erdos problem). Ann. Math. 1995, 142, 611–625. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).