1. Introduction
The evolution of many-particle systems (where the word “particle” must be read in the widest sense, as any individual belonging to a gas, or to a herd of predators or prey, or to any human collectivity, like a nation, social class, or condominium), in view of its wide range of interpretations in different scientific fields, is one of the most interesting and widely studied topics of that branch of applied mathematics dealing with models of natural, political, social, and economic phenomena. The mathematical formalization of such an evolution has been tackled from many different viewpoints and with many different languages and techniques, e.g., that based on Bayesian Networks, see [
1,
2,
3], or that based on the use of Fokker–Planck equation [
4,
5,
6]. Among these, kinetic-theoretic models (see
Section 6 for more details) seem to be the most versatile, expressive, and effective tools to describe many-particle systems and their evolution by means of suitable Equations (we can quote [
7,
8,
9,
10,
11,
12,
13,
14,
15,
16,
17,
18] as just as a small sample of the research addressed to this field). Since the so-called transition matrices (see
Section 2) characterizing Markov Chains play a fundamental rôle in kinetic-theoretic models, as they describe in stochastic terms the results of mutual interactions between the particles of any system. It should be clearly understood that, as pointed out in formal terms in [
19,
20], the equations of the Kinetic Theory for Active Particles are firmly based on the use of Markov Chains; however, rather surprisingly, in the literature about the kinetic-theoretic models, Markov Chains, at least to the best of our knowledge, have never been explicitly related to the evolution equations before [
19,
20]. In [
20], we tried to show that the explicit reference to Markov Chains actually represents a good perspective to describe the behavior of a many-particle system when it interacts with the external world (in this connection, see [
21] where suitable «forcing» terms are introduced in the equations). In [
20], we tried to show that, when we consider the evolution of a many-particle system as an
n-tuple of joined stochastic processes, one of which is a vector Markov Chain [
19], then the interactions with the external world can be described by assuming that the interaction rates and the transition matrices undergo stochastic variations at each step [
19].
The main aim of [
20] was, however, the interpretation rather than the formal and technical analysis of the evolution equation in the presence of random varying interaction rates and transition matrices, As a consequence, the accurate description of the equations obtained in such a case is simply outlined in the final sections. The aim of the present paper is to develop that discussion and analyze the technical differences it introduces in the description of the evolution of the system. We also try to give some interpretations of the meaning of such differences.
The rest of the paper is organized as follows.
Section 6 is devoted to recalling some basic features of kinetic-theoretic models. Definitions and notions about Markov Chains as stochastic processes, with particular regard to discrete time-discrete Markov Chains are given in
Section 2, while in
Section 3 we introduce (two-dimensional) vector Markov Chains and describe some of their basic properties in the discrete time-discrete case.
Section 4 is devoted to introducing what we have called “continuous semi-Markov coupled random processes” with particular concern with time-continuous processes, and to conclude that the equations of the Kinetic Theory, in their most general form, can be seen as a special form of the basic vector equation connecting the absolute probability distributions at different steps of a Markov Chain with each other and with the transition probabilities.
Section 5 is devoted to introducing the Markov Chains with stochastically varying transition matrices in the time-discrete case, in order to give a heuristic basis to their properties, while in
Section 7 we show how this kind of Markov Chain modifies the evolution equations for a many-particle system interacting with the external world. Finally, in
Section 8, we discuss some problems left to be tackled in the modified scheme and outline some possible research perspectives.
2. Some Background on Markov Chains
We now turn to giving the basic definitions of Markov Chains and a short account of their basic properties.
To start with, we recall that a random process [
22] is a sequence
of random variables (readers not acquainted with the notion of “random variable” can usefully consult [
23], p. 37) with a common range
S of possible values (the
states), called the
state space of the process. Both the set
of indexes and the state space
S can be either a discrete or a continuous set. If
is discrete (
), then the process is said to be
time-discrete; if
is continuous (
), then the process is said to be
time-continuous. Analogously, if
S is discrete (
, with
), then the process is said to be
discrete, while, if
S is continuous (i.e., any real interval), then the process is said to be
continuous. From now to the end of this section, we shall first consider discrete (either time-discrete or time-continuous) processes. In both cases, for the sake of simplicity and without loss of generality (as regards Markov Chains, this simply amounts to avoiding matrices with infinitely many rows and columns), the state space
S will be assumed to be finite.
Now, consider a discrete and time-discrete random process
with
,
. As usual, we denote by a vector
(the
state vector of the process at time h) the probability distribution on
S according to the random variable
. In other words,
with
. A discrete time-discrete Markov Chain [
22,
24,
25]
, with
, is a random process such that the Markov condition
holds (where, as customary, we have used the symbol
to denote the conditional probability of an event
A under the assumption that an event
B has taken place [
23]). This means that, for any
, the values of the
h-th random variable of the process depend only on the values of the
-th, not on the values of any previous variable. Now, setting
(with
),
is the
transition probability from state
to state
at the
h-th step, and the matrix
is called the
transition matrix at
h-th step. It is obvious and well known that
and that, in virtue of the law of alternatives [
23],
For the sake of completeness, we also recall that, for any couple
of non-negative integers,
and in particular, when the Markov Chain is stationary, i.e., a transition matrix
exists such that
for any
,
where the power at right-hand side must be interpreted in the sense of the row-by-column product of matrices (for further details, see, e.g., [
22,
24,
25,
26]).
3. Joint and Marginal Transition Probabilities
As already carried out in [
20], and almost tracing the steps of the treatment presented there, our next step will be to now consider a special kind of Markov Chain, never explicitly considered before in the literature about random processes (as far as we are aware, a state space endowed with a structure is considered only in [
27]), namely, vector Markov Chains (see [
19]). Like in [
20], for the sake of simplicity, we examine the case in which
(where
, so that
; as usual in the literature about discrete Markov Chains, the states in
S are assumed to be arbitrarily ordered, so that we can set
, and the order chosen is quite immaterial. Since
is a couple of elements of
, we write
for any
, with
, and the process we want to consider is a sequence
, which will be written in vector notation as
. So, the transition probabilities are
(where we have set
and
). Accordingly, the transition probabilities are labeled by 4 indexes, and we can set
For any
, the four-dimensional transition matrix
will be called the
joint transition matrix. Each of its elements expresses the probability that the chain passes from a state
to a state
at
h-th time. These are the
joint transition probabilities. Obviously, together with these joint probabilities, we can consider also a number (in this case,
) of different
marginal probabilities. More precisely, for any
, we can consider the probabilities
and
. The former is the probability that, if
, then
takes one of the values
with
; the latter is the probability that, if
, then
takes one of the values
with
. Thus, we have two three-dimensional matrices, each with
entries. We shall denote by
the entries of each matrix. By definition, we have
Recall now that
, so we can agree to set
and write the first of relations (
9) in the form
(where we have replaced the indexes
and
by
h and
k, respectively, and the indexes
and
by
r and
s, respectively). Next, we introduce the joint state vector
, where
is the probability of the state
at time
h, and the marginal state vector
is the probability of the set of states
, so that
and, in virtue of the law of alternatives
Hence, replacing this last relation in relation (
11), and taking into account relation (
10), we obtain
This last equation, by subtracting the probability
at both sides, yields
Moreover, for any
,
so that
and we can rewrite Equation (
12) in the final form
Finally, if the random variables
and
are independent for any
, and the chain is stationary, then we find
The first term at right-hand side will be called the
gain term of the subset of states
, since it accounts for all the possible transitions from other states to a state of
, while the last term is called the
loss term of
, since it accounts for all the possible transitions from a state of
to any state
outside .
Therefore, we are on the way to show that the statistical equations governing the evolution of many-particle systems (see
Section 6) are simply equations describing the evolution of suitably defined Markov Chains. In the next Section this conclusion will be completely achieved.
4. Continuous Semi-Markov Coupled Random Processes
We now introduce what we shall call a
continuous semi-Markov coupled random process, that is a random process
, where
is a continuous parameter interpreted as the time variable and
can be either finite or infinite. Here,
is a two-dimensional Markov Chain like the one described in
Section 3, and
is a
-dimensional vector Bernoullian process. More precisely,
, with ;
For any and any , is the classical Bernoulli variable, with range ;
For any pair of triples and in , the random variables and are independent.
The meaning of the last condition requires at least a short discussion. When interpreted in the framework of the description of the behavior of many particle systems, it means that the probability that two particles interact at a given time , when they are in two given states and , is not influenced by the assumption that they interacted at a different time nor on the assumption that an interaction between them has occurred when they were in a different couple of states. Of course, at first glance such a condition seems to introduce an undue restriction, but it is classically acknowledged to be quite plausible, since the probability of interaction is linked to states, not to particles, and the condition simply expresses the almost obvious fact that the way in which the states and have been achieved should not be taken into account.
As already observed in [
20], the state vectors obviously depend on time, but the transition probabilities and the probabilities associated with the Bernoulli variables
may or may not depend on time. Obviously, on a strictly empirical ground, a positive probability distribution
on
such that
for any
t cannot be assigned; all what can be done is to assign a continuous “probability density”
such that, for any
and any sufficiently small
,
is (approximately) the probability to find in any interval
some points such that
. Roughly speaking, and referring to the interpretation of probability as relative frequency, we can state that
is the length of the set
, so that
is the ratio of that length to the length
. So, in terms of relative frequency, each
must be interpreted as the
instantaneous interaction rate in the states
and
at time
t. We preliminarily consider the case in which the interaction rate
is constant with respect to time for any
, and write
. So, for any assigned couple
, the Bernoulli variables
and
are identically distributed for any
and, for simplicity, we may write
for any
t.
The continuous Markov Chain and the Bernoulli random process are coupled in the sense that
For any
, the transition matrix
is assumed to depend on
(and only on
), in such a way that
where
(independent of time) is the transition probability from the state
to the state
under the assumption
, and
is at any time the transition probability from the state
to the state
under the assumption
;
We assume that
where
and
are the well-known Kronecker symbols, that is,
when
and
when
, and
when
and
when
. This means that the Chain remains in each state
when
(notice that the Markov Chain is stationary if and only if the matrix
is constant with respect to
t. This remark will prove meaningful in the final sections.)
Accordingly, it is readily seen that, if for any
, the random variables
and
are assumed to be independent, then system (
14) may be written as follows:
Next, by dividing both sides by
and letting
, we obtain
if the transition matrix
is stationary.
A comparison between system (
19) and system (
23), which will be made in
Section 6, now shows at once that the latter can be viewed simply as a particular interpretation of the basic Chapman–Kolmogorov vector equation for the marginal state vectors of a two-dimensional vector Markov Chain. Furthermore, the random variables of the chain are pairs of states of the particles of the system. It could also be easily proved that the dimension of the vector Markov Chain describing the evolution of the system is strictly connected to the choice to allow and consider only pairwise interactions between the particles of the system. If one would decide to allow and to consider multiple interactions (e.g., simultaneous interactions between
particles), then the Markov Chain describing the evolution should be
p-dimensional.
This will also allow us to state that all the conditions leading to the deduction of system (
23) in the mathematical physical framework can be reduced to the sole following axiom:
The random process of p-tuples of states of the particles of the system is a p-dimensional vector Markov Chain conditioned by simultaneous interactions of p particles.
This conclusion allows to acknowledge that all assumptions about the influence of the external world on the evolution of the system can and must involve only the form of the transition matrices and their possible dependence on time or external parameters. In the remaining part of the paper, we shall treat the case in which the transition matrices are in turn random variables depending on other random variables describing relevant features of the external world.
5. Markov Chains with Stochastic Transition Matrices: A Heuristic Description
In this Section, we leave aside for the moment any reference to Kinetic Theory and to the mathematical description of the evolution of a many-particle system, in order to discuss in quite general terms the Markov Chains for which the transition matrices are never assigned but simply treated as random variables, depending on the outcomes of some additional experiments. In particular, we want to offer an example of a sequence of extractions from an urn for which this is the case.
Confining ourselves to the case of stationary Markov Chains, consider a sequence of extractions from an urn
U containing
b black balls and
w white balls, and set
. The extractions are performed according to the following rule: for
, before extracting the
h-th ball from the urn, we replace in the urn the ball we have drawn at the
-th extraction. So, it is readily seen that the transition matrix at
h-th step is
for any
, at least if nothing modifies the state space (i.e., the urn) at any step. However, now, at each step, we actually intervene on the urn
U, and our intervention will be decided by an extraction from another urn
containing only three balls, labeled in any way we want (for instance, one can be white, one can be black, and the third red). If we draw from
the white ball, then we add only a white ball in
U; if we draw the black ball, then we add only a black ball; if we draw the red ball, then we leave
U unchanged; in any case, we replace the ball in the urn
. Accordingly, the transition matrix from the outcomes of the first extraction from
U to those of the second one will be:
if we have drawn from
the white ball;
if the ball drawn from
is black; finally,
will simply be the matrix (
20) if we have drawn from
the red ball. In the considered case, the matrices
,
and
have the same probability
, but if
contained for instance three white balls, four black balls, and only two red balls, then the probability of
was
, that of
was
and that of
was
.
This example aims to introduce in as general terms as possible the notion of Markov Chains coupled with another random process. As a matter of fact, we considered a random process , which is a time-discrete Markov Chain with state space (where “0” stands for “black” and “1” stands for “white”) and another time-discrete independent random process with state space (where “0” stands for “black”, “1” stands for “white”, and “2” stands for “red”). The form of the transition matrix of process at each step i is ruled by the outcomes of process at all the steps from the first one to the i-th one, so that we shall give the following two definitions:
Notice, however, that the variables of the ruling process are assumed to be not only independent but also identically distributed, while independence is assumed only for the sake of simplicity, and does not affect the generality of our description, the identical distribution of the random variables () assures the independence from h of all conditional transition probabilities, at least when each transition matrix is ruled only by and is independent of for any . Of course, the above elementary example does not satisfy such condition, but it is possible to find different examples—though less simple—for which it is actually met.
In order to go a little deeper into the behavior of a Markov Chain in the presence of a ruling process, let us denote by the given Markov Chain and by , by the state space of and by that of , by the absolute probability distribution on D at i-th step (the i-th state vector of the chain) and by the i-th state vector of the ruling process . Moreover, assuming once and for all that the transition matrix of at each step is ruled only by the value of at the same step, we shall use the symbols and to denote the state vector at k-th step and the transition matrix of at the same step conditional to the assumption that .
Now, given the initial state vector
of
, we obviously have
so that
This shows that—under the assumption that the random variables
are independent and identically distributed—a chain conditioned to the repetition of the same “external” event is
stationary. More generally, for any fixed
k, consider any element
of
. One has
so that
(note that the symbol
denotes the state vector at
k-th step depending on all the external events from the first step to “just before” the
k-th step).
These relations are of the greatest importance for applications to the behavior of many-particle systems subjected to random external influences, not only to understand its dependence on the sequence of all past and present external events but also to see that a complete description of such behavior cannot ignore a comparison between different predictions corresponding to different sequences of such events.
6. Kinetic-Theoretic Models: A Recall
As laid out in
Section 4, the equations expressing the link between the state vectors of a coupled semi-Markov continuous process via the transition matrices, when written in the form of differential equations, simply become the equations describing in statistical terms the evolution of a many-particle system according to the Kinetic Theory for Active Particles. In order to clarify this point, we devote this Section to recalling some basic features of the kinetic-theoretic scheme, so that, in the subsequent Section, once the rôle of Markov Chains in the equations of the model is pointed out, we shall be able to study the way in which the dependence on random events modify the equations themselves.
A
complex system is a set
S of a very large number of objects called
individuals or
active particles, that, as already laid out in the Introduction and shown in the papers cited therein, can represent not only particles of a mechanical systems but also cells, living individuals, or human beings. The
state of each individual in
S is defined, according to the context in which we aim to develop our description, as a scalar variable
u or a vector variable
which, much more for historical reasons than for reasons suggested by the context, is called
activity or—more simply—
a state variable. This variable may describe the kinetic state of a material particle, or the activity of a cell in its interactions with other cells of a different tissue, or the health state of a living individual, or the level of wealth and the social condition of a human being in a collectivity. According to the context, the activity may be assumed to take its values in a
discrete domain (typically, a finite or countable subset of
or
) or in a
continuous domain (typically, a bounded or unbounded real interval or a bounded or unbounded domain of
). In any case, the domain
(or
) in which the state variable is allowed to take its values is called the
state space of the system [
21].
As in Boltzmann’s Kinetic Theory of Gases [
7], the viewpoint from which the mathematical framework is developed is statistical, i.e., we are interested to describe the evolution in time of the system as a whole rather than the way in which the states of single particles vary with time. Since we are aware that any precise description of the state of each particle of
S must be given up, for both theoretical and technical reasons, we decide to study the relative frequency (or probability) distribution over the state space at each instant. In other words, the state variable is conceived as a
random variable, and the goal of the study becomes the forecast of the evolution of its probability density function as well as of some of its parameters (expected value, moment of order two, standard deviation, and so on) [
21].
In many cases of interest, S is considered to be split into a family of subsystems, called its functional subsystems. The introduction of such subsystems aims to describe the interactions between different subclasses of individuals of S, characterized by different reactions to interactions, and offers several applications (1) in biology, for instance to describe the fight between tumor cells and immune system, or the competition between different species to model Darwinian selection; (2) in social sciences, to model interactions between social classes; and (3) in economics, to model the fluxes of wealth, e.g., between the class of financial managers, that of entrepreneurs, and that of salaried workers.
In some cases, the same variable can be used to identify the state of the members of all the functional subsystems, while in other cases, different state variables must be used for different subsystems.
In view of the aims of the present paper, we shall, however, disregard any possible decomposition of the system into different subsystems. Such a decomposition would simply result in a quite useless complication of notation and multiplication of similar equations. Accordingly, in this Section, we shall describe in some details only the case of a system
S consisting in one subsystem equal to the whole
S. The state space associated with the system will be denoted by the symbol
For any
t in a time-interval
, the state of
S at time
t will be identified by a probability distribution (also called
a state vector)
, …,
on
D, and, for any
, we can define the function
. According to this definition, we have
Now, the evolution of the system
S is viewed as a time-continuous stochastic process, and the time derivative of each probability function
is expressed, according to the law of alternatives, in terms of transition probabilities. More precisely, for any
, the symbol
will denote the probability that a particle of
S falls from the state
to the state
after an interaction with another particle of
S which is in the state
: accordingly,
Thus, using the law of alternatives, we write the following system of differential equations
where, for any
,
is the so-called
encounter rate of particles of
S in the states
and
, that is the number of pairwise interactions between particles of
S, that are in the states
and
, per time unit, so that for any sufficiently small
, the product
is the probability that one such interaction occur in the time interval
provided the individuals involved are in the states
and
(accordingly,
is in turn a
conditional probability and
is the ratio of a conditional probability to time).
The terms on the right-hand side of Equation (
23) express, respectively, the increase in probability of state
as the probability that some «candidate» particles of
S in the state
interact with some
field particles of
S in the state
and, as a consequence, fall in the state
with a positive probability, and the decrease in probability of state
as the probability that some «test» particles of
S in the state
interact with some
field particles of
S in the state
and, as a consequence,
leave the state
with a positive probability.
7. The Continuous Case and KTAP Systems with Stochastic Transition Matrices
Now, in order to deepen the reasoning presented in [
20] (Section 7), we consider again a stochastic process
, where
,
and
; just for the sake of simplicity,
is assumed to be two-dimensional and
is assumed to be scalar for any
;
is instead a
matrix-valued random variable for any
. For the sake of completeness and clarity, we collect again our assumptions on these processes:
We denote by and , respectively, the ranges of and for any . The range of is simply the set for any and any ;
is a Markov Chain, and for any the variable depends on all the variables but is independent on any variable for ;
For any each variable depends on all the variables only via the dependence of its transition matrices () on the corresponding variables ;
For any the variables and are independent;
For any pair of triples and in , with , the random variables and are independent;
For any pair of times in , the random variables and are identically distributed;
Each variable is independent of all variables , with ;
The variables are independent and identically distributed.
This stated, before going on to the technical development of the consequences of our definitions and assumptions, we need to discuss and explain their meaning in some details. To this end, we start by outlining an example of a stochastic process of the type under examination, and try to show how it suggests the above conditions to be quite reasonable.
It is to be noted that a complete description of any example of a time-continuous process of such kind would require a long and rather complicated treatment, which would go far beyond the limits of this paper and requires a separate paper. So, we shall confine ourselves to a concise outline. Furthermore, in the same way as for the case of time-discrete processes, we shall not refer to the case of interacting subjects, because this would force us to consider at least two-dimensional vector variables, and we want to avoid formal complications.
This stated, consider the evolution in time of the bank account of any subject we may choose. Such evolution is readily seen to be well described as a continuous “random walk”, once the probability distribution on the set of possible amounts of payments and incomes and the rates have been assigned; the probability of any possible amount of money in the account at each instant t does not depend on its history, but—roughly speaking—only on the amount at instant . Now, as the “ruling” urn, we choose a set of possible abrupt external events producing unpredictable new expenses or incomes (climate changes requiring unusual heating or cooling expenses, unexpected taxes, thefts, fires, earthquakes, unexpected inheritances, lottery winnings, unexpected salary advancements), all labeled with the amounts of money they could produce, so that the range D and the range S are the same set (with different probability distribution functions, of course). These events are random, with rates (probabilities) that are usually very low, and must be estimated by means of a statistical analysis of the history of similar events, and are very likely to be independent of each other, as it is difficult to assume that any one of the above events take into account the preceding ones to decide whether they will take place or not. This gives a full explanation of assumption 8. In the same way, the assumptions on the variables seem to be very plausible, as these variables simply express accounting information, which is essentially linked to the life and habits of the subject under examination, and we could say that they define the system.
We are well aware that, while the condition on , as expressing the abruptness and the randomness of external events, can be assumed in almost any application of the model, the conditions on the interaction rates would require to be modified according to the meaning of the word “interaction” in each context, and could turn out to be quite unlikely for more complicated examples, but this point cannot be examined here, and must be postponed to future analysis.
Now, turning again to the formal development proposed above, in order to study the form of System (
19) when the transition matrices of the Markov Chain
depend on the values of random variables
, we shall first agree to set
and to denote by the symbol
the probability that the couple of states
falls in any state whose first component is
when not only a particle in the state
interacts with a particle in the state
but also the random variable
takes the value
. Furthermore, we agree to denote by
the joint probability that an interaction between a particle in the state
and a particle in the state
occurs
and in any time interval of length
, so that
will be the
instantaneous probability density of these joint events (which we have implicitly assumed to be constant).
In addition, we must also introduce—for any —the symbol , to denote any unspecified element of . For reasons that will become clear later, we also agree to denote by the joint probability of events , , …, . Note that, in view of the independence of from the variables of the process , the product is the joint probability density of events , , …, and interactions between a particle in the state with a particle in the state . According to the classical treatment of the kinetic model, we will assume that and are independent of time for any i, j, , , …, .
This stated, in the time-continuous model with transition matrices that vary randomly over time depending on the values of the variables
, System (
19) can be rewritten in many ways, according to the information we want to take into account and the information we want to obtain.
First of all, if we want to link the instantaneous variation of the distribution of the states over the particles of the system to assumptions about possible past
r-tuples of external events (where
r is assigned arbitrarily in advance), we may simply consider the system
where
we have dropped out the index of the component of the state vectors because the variable is always the same;
and
;
the conditional symbol “
” should be read “under the assumption that
r variables of the process
in the time interval
have taken the values
,
, …,
”;
we have set
Note that, as a matter of fact, System (
24) is
not only one system, but a whole set of
systems of equations, one for each possible choice of assumption
in
. As we shall discuss in detail in the final Section of this paper, this would allow us to compare the possible behaviors of the many-particle system under examination corresponding to different possible
histories of external influences if we could treat the distributions at right-hand side like
known functions of
. Since this is not the case, bearing in mind the independence from time of
for any choice of
, we may multiply each system of the set by the corresponding
, then add side by side all the
systems of the set, to obtain the unique system
where we have taken into account the law of total probabilities
It is obvious that, while the last system (
25) is much simpler to manage, the set of systems expressed by (
24) gives much more information in connection with the sequence of external events preceding the one identified by the value of
. This point requires a careful discussion, which will be proposed in the concluding Section of the paper.
8. Conclusions and Perspectives
As the reader will have undoubtedly noticed, the main aim of the present paper was to give a detailed and precise account, in the light of the basic language of Markov Chains, of the case in which the transition matrices appearing in the equations governing the evolution of a many-particle system in turn depend on stochastic events taking place in the external world and influencing the behavior of the system. This case was treated only as a hint in [
20], which focused on possible applications of the Kinetic Theory for Active Particles to social and economic problems, and has been discussed in more general terms here.
In order to synthesize the main significant aspects of the discussion presented here, we want to direct the attention of the reader to the fact that the dependence of the form of transition matrices of a Markov Chain from random events at each step produces a “tree” of possible ordinary, deterministic chains, whose number of branches, in the time-discrete case, grows exponentially with the number of steps. More precisely, at the r-th step, if the number of possible events influencing the form of the transition matrix is n, the number of chains is , even when the form of r-th transition matrix is modified only by the r-th random external event.
In the time-continuous case (which is the most interesting one in view of applications to physical or biological or economic or social problems), the situation is far more complicated. In principle, at any time , we could have any number of different sequences of transition matrices, with different probabilities, and we should formulate suitable assumptions on the sequences to be taken into account. However, since the form of the system hides the dependence of the distributions from , we are forced to use the law of alternatives, and to ignore any assumption on the sequence of external events preceding the one influencing the transition matrix at time t.
Conversely, as a consequence of this remark, the solution of system (
25) in an arbitrary time interval
is somehow misleading, as it suggests the idea that a particular external event took place at time
and influenced the whole behavior of the system in the whole interval
. So, we lose any reference to a sequence of random external events. The only correct way to apply the notion of “random variable transition matrix” is the following: we assume that a sequence of times
exists such that some events
,
, …,
happened, influencing the transition matrix in such a way that in the time interval
, the system governing the evolution of the system is
for
,
for
and
for
, where
and
. Since the distributions on the right-hand side of each system in the interval
are known, as obtained by solving the analogous system in the interval
, we obtain a
possible evolution of the system in the whole interval
. More precisely, we obtain a solution for any assigned
r-tuple
, and each solution has the same probability of the sequence
, that is the probability
. In view of identical distribution and independence of random variables
, one has
where
for any
t.
Now, besides the complete irrelevance of any precise identification of times
,
, …,
, we must remark that
is arbitrary, so that the number of possible solutions in the given time interval
is potentially infinite. More precisely, we may repeat the same procedure for any
, to obtain
solutions
defined by the conditions
where
.
We cannot but acknowledge that this is a purely abstract treatment of the problem of describing the behavior of a many-particle system influenced by random external events. Furthermore, we cannot hide that its abstract character, joined with the formal complication and the increased number of variables and parameters, can make the application of the model to real phenomena remarkably difficult; however, we are encouraged by the remark that the difficulty is in nature the same as in the classical kinetic model in which the transition matrices are assumed to be constant, and the difference is only technical. The subsequent steps, besides the classical existence, uniqueness, and stability problems for each of the partial solutions listed above, and the explicit computation of solutions (at least by means of numerical simulations in suitable particular cases), must be essentially of a statistical type, and mainly oriented by applications (however, this is true also in the classical model). For each application, we need to establish whether a
exists such that the rate of
r-tuples of external events is zero for any
. At the same time, one should also study the
expected behavior of the system, described by the relation
where
is the probability density (rate) of a sequence of
r external events. It is quite clear that the terms corresponding to large values of
r are smaller the larger
r is. However, the application to real systems on the one hand, suggests, and on the other hand, requires that only a finite number of external events actually occur in any given time interval of assigned length. The analysis of these points is the natural development of the present proposal, and should be the object of future papers.