Previous Article in Journal
Quasi-Discrete Time Crystals in the Quasiperiodically Driven Lipkin–Meshkov–Glick Model
Previous Article in Special Issue
Permutation Entropy and Its Niche in Hydrology: A Review
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Ordinal Random Processes

Institute of Mathematics, University of Greifswald, 17487 Greifswald, Germany
Entropy 2025, 27(6), 610; https://doi.org/10.3390/e27060610 (registering DOI)
Submission received: 5 May 2025 / Revised: 2 June 2025 / Accepted: 5 June 2025 / Published: 7 June 2025
(This article belongs to the Special Issue Ordinal Patterns-Based Tools and Their Applications)

Abstract

:
Ordinal patterns have proven to be a valuable tool in many fields. Here, we address the need for theoretical models. A paradigmatic example shows that a model for frequencies of ordinal patterns can be determined without any numerical values. We specify the important concept of stationary order and the fundamental problems to be solved in order to establish a genuine statistical methodology for ordinal time series.
MSC:
37A35; 62M10; 60G07

1. Introduction

1.1. The Main Question

Time series analysis has two parts. First, there are methods for describing data series, like means, correlations, and Fourier spectra. Second, there are models, like Brownian motion or ARMA models, which describe mechanisms that could generate similar sequences of data points. Most models postulate that the next state X t + 1 is a function f ( X t , X t 1 , ) of previous states, with some random influence. A fundamental example is a simple random walk, which states that X t + 1 is either X t + 1 or X t 1 , depending on the result of a coin toss.
In recent years, the study of ordinal patterns in time series has proven successful for many practical applications. However, there is a lack of appropriate models. Established numerical models provide little information on the ordinal structure. This paper proposes directly constructing proper ordinal models. We raise the following question:
Can we define mechanisms that explain the order relations between data points in a time series without using any numerical values?
We shall give an affirmative answer, providing an example of a coin-tossing order. It is similar to a simple random walk. Coin throws are used to define the order relation between all data points while the values, X t , remain undefined. This is a completely new type of model. Due to its simplicity, it can be easily understood. To model real-world data, the example is too simple. It must be modified and extended. Toward this end, we provide a mathematical framework in the last section.

1.2. Contents of the Paper

In this introductory section, we briefly introduce the basic concepts. For more details and background, we refer the reader to [1] [Section II] and [2,3,4,5,6,7]. One new point involves the visualization of the distribution of ordinal patterns by refining ordinal histograms in Section 1.5. The main topic of this paper is explained in Section 1.7. Order-related properties of processes, particularly order self-similarity, are introduced in Section 1.8.
Section 2 deals with a paradigmatic example of an ordinal process that has no numerical values. The construction is algorithmic and intuitive, and the pattern probabilities for length 3 match those of Brownian motion. Unfortunately, the example is not self-similar and is not relevant to applications. We conjecture that there is a self-similar modification.
Section 3 provides a more technical outline of the possible theory of ordinal processes. We construct stationary ordinal processes without using numerical values. Order self-similarity remains an open problem.

1.3. Permutations as Patterns in Time Series

A permutation π of length m is a one-to-one mapping from the set { 1 , 2 , , m } onto itself. We write π k for π ( k ) and π = π 1 π 2 π m . Thus, π = 231 means π ( 1 ) = 2 , π ( 2 ) = 3 , π ( 3 ) = 1 . However, we do not need mapping properties of permutations, like composition. We just consider the graph of π as a geometric pattern (an ordinal pattern).
With x = x 1 , x 2 , , x T denote a time series of length T . This can also be viewed as a mapping that assigns to the time points t = 1 , , T the values x t = x ( t ) . For m time points t 1 < t 2 < < t m between 1 and T, we consider the pattern of corresponding values. We define
( x t 1 , x t 2 , , x t m )   shows   pattern   π   when   x t i < x t j   if   and   only   if   π i < π j .
In other words, the correspondence x t j π j , j = 1 , , m is strictly monotonous. This is very intuitive. The pattern π can be determined from the time series by calculating ranks; π k is the number of time points t j for which x j x k .

1.4. Pattern Frequencies in Time Series

Instead of arbitrary t j , we study equally spaced time points t < t + d < t + 2 d < < t + ( m 1 ) d . We call m the length and d the delay or lag of the pattern. We always take m 6 and mainly consider m = 3 and m = 4 . However, we vary d as much as possible.
Ordinal pattern analysis is a statistical method. We fix a length, m, and determine relative frequencies of all patterns, π , of length m , for various delays, d . We divide the number of occurrences of π by the number of time points where π could occur.
p π ( d ) = 1 T ( m 1 ) d # { t | 1 t T ( m 1 ) d , ( x t , x t + d , , x t + ( m 1 ) d )   shows   pattern   π } .
These frequencies are the basic parameters of ordinal time series analysis. They can be combined to calculate permutation entropy, persistence, and other important parameters [1,2,3,4,5,6,7].

1.5. Visualization of Pattern Frequencies by Ordinal Histograms

This paper is not about parameters like entropy. It studies models for all pattern frequencies together. For this reason, we introduce a graphical method that summarizes the pattern frequencies for a fixed d in a kind of density function. The permutations π are assigned subintervals I π of [ 0 , 1 ] as follows: I 12 = [ 0 , 1 2 ] , I 21 = [ 1 2 , 1 ] . We draw bars over I π , each with area p π ( d ) . For m = 3 , we note that p 12 = p 123 + p 132 + p 231 (neglecting a possible tiny error due to the last value for p 12 ( d ) in (2)). We choose I 123 = [ 0 , 1 6 ] , I 132 = [ 1 6 , 1 3 ] , I 231 = [ 1 3 , 1 2 ] so that their union is I 12 . Then we draw bars with area p π ( d ) . The same is done for π = 213 , 312 , 321 . In this way, the histogram for m = 3 is a refinement of the histogram for m = 2 , in the sense that the area above I 12 = [ 0 , 1 2 ] remains the same. In theory, successive refinements can be obtained for m = 4 , 5 , but in practice we rarely have reliable estimates for m > 4 . This arrangement of permutations is called the hierarchical order. Details will be given in Section 3.4. As Figure 1 shows—for a simulated time series—all pattern frequencies of a time series for a fixed d can be subsumed in this way by a sequence of refining histograms, and in the limit m by a density function.
The process chosen for Figure 1 is an AR(1) model with exponential noise: X t + 1 = 1 2 X t log U t , where the U t are independent uniformly distributed random numbers in [ 0 , 1 ] . Due to the asymmetry of the noise distribution, p 21 ( 1 ) 0.58 > 1 2 . Successive decreasing steps are frequent: p 321 ( 1 ) 0.3 and p 4321 ( 1 ) 0.13 . We cannot explain why 2143 is the permutation with the second-largest frequency. However, an AR(1) model with the coefficient 1 2 has a short memory. For d = 3 , the patterns already have almost uniform frequencies. The models considered below are more uniform for m = 2 and 3, but they also provide interesting behavior for large d .
The objective of this paper is to find typical ordinal histograms from a theoretical viewpoint. An important objective is to find models where the histograms agree for all d . Two such models are known [1] [Section II.C], that is, white noise, where the histogram is a constant function for all m , and Brownian motion; see Figure 3. We are looking for other models.

1.6. Stationary and Order Stationary Processes

When modeling time series, we suppose that the given time series is an instance of a larger ensemble, called a stochastic process. The process represents the mechanism that generates the given time series and many other ones. We want to know the laws of this process. Formally, a stochastic process, X, is a sequence, X = X 1 , X 2 , , of real random variables on a probability space ( Ω , P ) . The time series, x, is a random choice from the ensemble X .
To find the laws of X from a single time series, we must assume that the distributions and dependencies of X k do not change in time. Usually, X is required to be stationary. That is, the m-dimensional distribution of ( X t , X t + 1 , , X t + m ) does not depend on t , for all m . This assumption is very strong. When we study autocorrelation, we assume weak stationarity. That is, the mean M ( X t ) and the covariance C o v ( X t , X t + k ) do not depend on t [8]. When we consider ordinal patterns, we assume a similar, very weak stationarity assumption. It is required throughout this paper.
Order stationarity. For a given pattern π with length m and delay d , let
P π ( d ) = P { ( X t , X t + d , , X t + ( m 1 ) d )   shows   pattern   π }
denote the probability that the process, X, shows pattern π just after time t . We say that the process, X, is order stationary for patterns of length m if P π ( d ) is the same value for all t . This should hold for all patterns, π , of length m and all delays, d . For a time series of length T , we consider only d = 1 , , d max , where d max must be smaller than T / m .
Brownian motion (where X 0 = 0 and X t + 1 X t are independent increments with standard normal distribution for t = 0 , 1 , ) is an order stationary process that is not even weakly stationary. It is a basic model for the financial time series. Note that the process, X, and the pattern probabilities, P π ( d ) , are theoretical objects. The observed frequency p π ( d ) serves as an estimate of P π ( d ) , and from a theoretical perspective, the estimator Formula (2) has nice properties [9,10,11,12,13]. The practical success of ordinal patterns indicates that the assumption of order-stationarity is at least approximately fulfilled in applications.

1.7. The Topic of This Paper

There are plenty of well-studied stochastic processes, including Gaussian processes, random walks, ARMA and GARCH processes, processes with long-range dependence [11,14], etc. Most of them are either stationary or have stationary (multi-dimensional) increments, and are, therefore, order stationary. However, they are defined by arithmetical operations. This makes the rigorous study of ordinal pattern frequencies difficult, although various interesting theorems could be derived [9,10,11,15,16]. Even for the Brownian motion, exact frequencies of patterns can be determined only for m 4 [17]. Some of them are rational, and some are irrational. Frequencies for length 5 lead to integrals that can only be numerically calculated. Overall, the impression is that ordinal patterns do not fit into the established theory of stochastic processes.
The only proper theoretical model for pattern frequencies is white noise, which is a process consisting of independent and identically distributed random variables, X i . For each m , all permutations have the same probability, p π = 1 / m ! , for any delay, d . This model was repeatedly used as a null hypothesis for testing the serial dependence of time series [1,12,13,18,19,20,21]. A greater variety of models is needed.
In this paper, we attempt to define ordinal random processes in a combinatorial way, directly by their ordinal pattern frequencies. We do not require normal distribution, we require no arithmetic, and we do not even need numerical values to compare objects X i and X j . An algorithmic example in the next section will show that this is possible. In Section 3, we outline some technical concepts and basic statements for a theory of ordinal processes. This could be the first step toward finding more applicable ordinal models. Note that the possible equality of values, an important problem in practice [21], plays no role in this paper; our models will only allow X s < X t or X s > X t but not X s = X t .

1.8. Symmetry and Independence Properties of Ordinal Processes

Of course, it is not enough to just postulate pattern frequencies. Our models should also have nice properties. Here, we list the ordinal version of some of the most common process properties. We start with symmetry in time and space [9].
A process, X, is reversible if the distributions of ( X 1 , , X m ) and ( X m , , X 1 ) agree for every m . In the ordinal version, it is reversible if the permutations π = π 1 π 2 π m and π m π m 1 π 1 have the same pattern probability for every π . The process, X, is invariant under the reversal of values if the distributions of ( X 1 , , X m ) and ( X 1 , X m ) agree for every m . The ordinal version states that the permutations π = π 1 π 2 π m and ( m + 1 π 1 ) ( m + 1 π 2 ) ( m + 1 π m ) have the same pattern probability for every π .
In both cases, the ordinal versions of symmetry are much weaker and easier to check. These symmetry properties are often met in applications. As a consequence, the distribution of ordinal patterns of length 3 often has only one degree of freedom, since p 123 = p 321 , and the probabilities of the other four permutations agree. All Gaussian processes are symmetric in space and time [9,11,17].
A process, X, is said to be Markov if for any t and fixed X t , the distributions of ( X 1 , , X t 1 ) and ( X t + 1 , X t + 2 , ) are independent. In the ordinal version, this means that the appearances of patterns π , π at times t 1 < < t m t and t t 1 < < t m , respectively, are independent. Again, the ordinal property is weaker and easier to check. The Markov property appears in theory more than in applications.
Order self-similarity. The most important symmetry property in this paper is order self-similarity [22] [Section 6]. A random process in continuous time, Y t , t > 0 , is said to be self-similar if there is an exponent, H > 0 , such that Y r t = r H Y t in distribution, for each positive number, r . Such a process cannot be stationary, it must look similar to Brownian motion, and it has a very special mathematical structure [23]. The ordinal concept for discrete time is much weaker, more general, and very natural. It does not depend on a parameter, and it can be statistically checked for real-world data.
A process, X, is said to be order self-similar for patterns π of length m if
P π ( d )   is   the   same   value   for   all   d .
This should hold for all patterns, π , of length m and all delays, d .
For a time series of length T , we consider d = 1 , , d max , where d max is much smaller than T . With such a restriction, order self-similarity was statistically verified in financial time series [22] and EEG measurements [1]. For self-similar Gaussian processes with stationary increments, Sinn and Keller [9] estimated the Hurst exponent, H, from pattern frequencies. For real-world data, however, it is not at all clear to what extent the strong self-similarity is fulfilled and the exponent H is meaningful. Thus, for practical purposes, order self-similarity seems to be the right concept to start with. The present paper is motivated by the desire for order self-similar models.

2. The Coin-Tossing Order—An Ordinal Process Without Values

2.1. Ranking Fighters by Their Strength

Ordinal patterns were derived from numerical values of a time series. Now, we go the opposite way, similar to the world of tennis or chess, where players are first compared pairwise and then assigned a rank or ELO score. Suppose there are four players, A , B , C , D , in a tennis club. We design an algorithm to order them by strength. First, B plays against A . Let us assume B wins. Then we write A < B . Now the next player, C, plays against the previous player, B . If C wins, then A < B < C by transitivity of order. If C loses, however, C must still play against A . Depending on the result, either A < C < B or C < A < B . Now, the last player, D, comes and plays first against the previous player, C , and then against B and/or A , if the comparison is not yet determined by previous games and transitivity.
This procedure is turned into a mathematical model. We want to construct a random order between objects X 1 , X 2 , , which are not numbers. We follow the algorithm for tennis players and replace each match with the toss of a fair coin. As a result, we obtain a stochastic process like white noise or Brownian motion on the ordinal level only. It is possible to determine pattern probabilities much better than for the Brownian motion.

2.2. Definition of the Coin-Tossing Order

Repeated coin tossing is a standard way to represent randomness. Take a simple random walk, for example, and write X n for our position at time n = 0 , 1 , 2 , then X 0 = 0 , and either X n + 1 = X n + 1 or X n + 1 = X n 1 , depending on the result of a coin throw at time n . In the present case, X 1 , X 2 , X 3 , will denote objects, not necessarily numbers. We define a random order between the X i . We throw a fair coin, c j i , to decide whether X i < X j or X i > X j , for any pair of integers, i < j . Let us write 1 for ‘head’ and 0 for ‘tail’. Then our basic probability space Ω is the space of all 0-1-sequences, where each coordinate is 0 or 1 with probability 1 2 , independently of all other coordinates. Formally,
Ω = { ( c 21 , c 32 , c 31 , c 43 , c 42 , ) | c j i { 0 , 1 } f o r j > i 1 }
where c j i = 0 means X i < X j , and c j i = 1 means X i > X j , for 1 i < j . The important point is that c j i is disregarded when the order of X i and X j is already fixed by previous comparisons and transitivity.
The first coin c 21 decides the ordering of X 1 and X 2 . Now suppose X 1 , , X j 1 are already ordered. Then, X j is compared to X j 1 , X j 2 , , X 1 by considering the random numbers c j i . However, when the comparison is fixed by transitivity from the already defined ordering, then c j i is disregarded—that coin need not be thrown.
The resulting random order is called the coin-tossing order. It can be easily simulated. Figure 2 shows the rank numbers of 500 consecutive objects, X j , in the middle of a simulated series of length T = 10,000. The global rank numbers have strange discontinuities. Local rank numbers, obtained by comparing with the following 20 objects on the left and right sides, show a more familiar picture.
Problem. Is there a stochastic process with stationary increments that generates the coin-tossing order?

2.3. Basic Properties of the Coin-Tossing Order

Despite the erratic view of trajectories in Figure 2, it turns out that the coin-tossing order has the same ordinal pattern frequencies for length 3 as Brownian motion, for arbitrary d . In particular, it is order self-similar for patterns of length 3.
Theorem 1
(Basic properties of the coin-tossing order).
(i) 
The coin-tossing order is stationary and has the ordinal Markov property.
(ii) 
For any permutation, π, of length m , the pattern probability is P π ( 1 ) = 2 u with
u = u ( π ) = # { ( i , j ) | 1 i < j m ,   i f   i < k < j   t h e n   π k   i s   n o t   b e t w e e n   π i   a n d   π j } .
(iii) 
The pattern probabilities, P π ( d ) , are invariant under time reversal and the reversal of values.
(iv) 
For d = 1 , 2 , . we have P 12 ( d ) = 1 2 , P 123 ( d ) = P 321 ( d ) = 1 4 , and P π ( d ) = 1 8 for the other permutations of length 3.
Proof. 
(i): The order of ( X t , X t + 1 , , X t + m 1 ) depends on the random numbers c j i , t + m 1 j > i t , in the same way that the order of ( X 1 , X 2 , , X m ) depends on the c j i , m j > i 1 . Since both collections of random numbers have the same distribution, the pattern probabilities do not depend on t , and the defined random order is stationary. Moreover, the comparisons of X s with s t depend on random numbers c j i with j > t while the comparisons of X s with s t depend on c j i with j t . Since the different c j i are independent, this implies the ordinal Markov property.
(ii): u ( π ) is the number of coin flips needed to determine π . Given a permutation π , we determine the c j i variables, which are needed to define the occurrence of π in ( X 1 , X 2 , , X m ) . Of course, m j > i 1 , and j run in increasing order. For the fixed j , the number c j , j 1 is always used, and the other i variables are considered in decreasing order. Now, consider k with i < k < j . If π i < π k < π j , this is determined by the random numbers c k i and c j k , which were drawn before c j i . In that case, c j i is disregarded since π i < π j follows from the transitivity of the order. A similar argument applies when π i > π k > π j . However, if there is no π k between π i and π j , then c j i is needed to determine π . We shall call u ( π ) the energy of π .
(iii): First, we consider d = 1 . Given π , we have to show that the time-reversed and spatially reversed permutations have the same energy u ( π ) . For spatial reversal, this directly follows from the definition (5) with ‘between’. For time reversal, we show that P π ( 1 ) can also be determined backwards by considering c j i with decreasing i = m 1 , m 2 , , 1 , and for fixed i , with increasing j = i + 1 , , m . The point is that when we compare places, i < j , we have already compared both j and i with all k between. So, c i j is needed only if no π k is between π i and π k . Otherwise, the order between π i and π j is already fixed, and c i j is disregarded. This proves reversibility for d = 1 .
Now consider d > 1 and π of length m, which appears as a pattern of ( X 1 , X 1 + d , , X 1 + ( m 1 ) d ) . The probability of this event is P π ( d ) , which may be different from P π ( 1 ) . We have to calculate P π ( d ) from all ‘atom permutations’ of length M = 1 + ( m 1 ) d for ( X 1 , X 2 , , X 1 + ( m 1 ) d ) for which the order among the special m places agrees with the order of π . This would be a lot of work. However, since the reversed permutation of π is composed of the reversed ‘atom permutations’, both in space and in time, we shall obtain the same pattern probability.
(iv): The spatial reversal invariance implies P 12 ( d ) = P 21 ( d ) = 1 2 . Then, P 123 ( d ) = P 321 ( d ) = 1 4 follows from the Markov property. The equality of the four other pattern frequencies is a consequence of the invariance under time and space reversal.    □

2.4. Computer Work and Breakdown of Self-Similarity

In Figure 3, we compare pattern probabilities of length 4 for the Brownian motion, taken from [17], with those of the coin-tossing order. The Brownian motion looks more interesting since coin-tossing allows only the probabilities 2 k . Our model is simple, but not recommended for applications. For length 3, the P π ( d ) of the two processes agree and do not depend on d , according to Theorem 1. The figure shows that we must investigate patterns of length 4 if we want to distinguish processes like these two. Moreover, the Brownian motion is perfectly self-similar so that the probabilities, P π ( d ) , of any length do not depend on d [23]. Is this also true for the coin-tossing order?
Figure 3. Probabilities of patterns of length 4 (brown) on top of those of length 3 (blue) for the Brownian motion (left) and coin-tossing order with d = 1 (right). For length 3, the probabilities coincide.
Figure 3. Probabilities of patterns of length 4 (brown) on top of those of length 3 (blue) for the Brownian motion (left) and coin-tossing order with d = 1 (right). For length 3, the probabilities coincide.
Entropy 27 00610 g003
Equation (5) allows for accurate calculation of all pattern probabilities for d = 1 and m 10 . Note that such pattern probabilities just for m = 5 have never been determined for non-trivial stochastic processes, not even for the Brownian motion or ARMA models of low order [17]. Here is a MATLAB procedure that determines u ( x ) for a permutation x, given as a vector of length m . The vector w contains the logical values 0 and 1.
function u=ctoscore(x);
m=length(x); u=m-1; % coin throws for i=1,...,m-1 and j=i+1 are necessary
for i=1:m-2; for j=i+2:m; v=x(i+1:j-1); % look for numbers k between i, j
  if x(i)<x(j); w=v>x(i)&v<x(j); else w=v<x(i)&v>x(j); end % x(k) as in (5)
  if sum(w)==0; u=u+1; end % if such x(k) do not exist, add a coin throw
end; end;                                      % end of loops for i and j
It takes five minutes to run this procedure over all permutations of length m = 10 . The minimum number u ( x ) is m 1 = 9 and appears only for x = 12 m and x = m ( m 1 ) 1 . The maximum number u ( x ) = 33 is realized by eight permutations, that is, x = 6789 10 12345 , and similar patterns. About three-quarters of the 3.6 million patterns fulfill 23 u ( x ) 27 .
Now, we can determine P π ( d ) for all permutations of length 4 and d = 3 by adding probabilities 2 u ( x ) of ‘atom permutations’ x of length 10, as in the proof of Theorem 1, (iii). For π = 1324 , we consider all x with x ( 1 ) < x ( 7 ) < x ( 4 ) < x ( 10 ) . We add 10 ! / 4 ! = 151 , 200 cases. The result must be a rational number with denominator 2 k , where k 33 . For π = 1324 , we get P π ( 3 ) = 3 · 41 · 59 · 10 , 321 / 2 31 3.49 % . To determine P π ( 2 ) for patterns of length 4, it suffices to consider m = 7 .
The result is shown in Figure 4. Unfortunately, order self-similarity breaks down for patterns of length 4. For π = 1324 , we have P π ( 1 ) = 1 32 3.12 % while P π ( 2 ) 3.42 % and P π ( 3 ) 3.49 % . However, the probabilities for d = 2 or 3 can be taken as a new model, which is nearer to self-similarity and looks not as artificial as coin tossing in Figure 3.
Problem. Let us define new pattern probabilities Q π ( 1 ) = P π ( d ) for d = 2 , 3 , what are the properties of Q? Is there a limit for d ?
The small difference between d = 2 and d = 3 in Figure 4 indicates fast convergence. The mere existence of an order self-similar limit would be interesting, but not yet helpful. After all, Brownian motion is already available as an ordinally self-similar process. The merit of the coin-tossing order lies in its algorithmic flavor and its connection to ranking. We look for an ordinally self-similar model with such an intuitive explanation.
Finally, let us consider the interpretation of u ( π ) as an energy function. Let d = 1 , and let S m denote the set of all permutations of length m . Let P denote the probability measure on S m given by the probabilities, P π = 2 u ( π ) , of the coin-tossing order. The permutation entropy [24] of P , using logarithms with base 2, is just the mean energy with respect to P .
H ( P ) = π P π log 2 P π = π P π u ( π ) = M P ( u ) .
For all other probability measures, Q on S m , the permutation entropy is smaller than the mean energy; see [25] [Chapter 1]. Thus, P is a so-called Gibbs measure on S m .
Problem. Are there other meaningful energy functions for permutations, perhaps even parametric families of Gibbs measures on S m ?

3. Rudiments of a Theory of Ordinal Processes

3.1. Random Stationary Order

Having studied one example, we will now discuss a possible theory of ordinal random processes. As above, X 1 , X 2 , are not numbers, but only objects that are ordered. We have the discrete-time domain { 1 , , T } for time series and N = { 1 , 2 , } for models. An order is a relation < on the time domain such that x x , and x y implies either x < y or y < x , and x < y together with y < z implies x < z . Of course, the order does not apply to the time points t , but to the corresponding objects x t . Then the statement that x t , , x t + m 1 exhibits pattern π has a meaning for every π S m —it is either true or false. When an order on N or { 1 , , T } is given, we can determine pattern frequencies, permutation entropy, and so on. In the following, we always have d = 1 .
We want to construct models of ordinal processes, like the coin-tossing algorithm. For this purpose, we need the concept of random order. To keep things simple, we identify an ordinal process with its law, as follows: A random order is defined as a probability measure on the set of all orders on the time domain. For the finite time domain { 1 , , T } , a random order is just a probability measure, P T , on the set, S T , of permutations of length T . For m T , π S m , and 1 t T + 1 m , the random order allows determining the probability:
P π t = P { x t , , x t + m 1   shows   pattern   π }
The random order is called stationary if the probabilities P π t do not depend on the time point t , for any pattern π of any length m < T . In other words, the numbers P π t must be the same for all admissible t . This is exactly the order stationarity that we defined for numerical processes in (3).

3.2. The Problem of Finding Good Models

The infinite time domain N is considered in the next section. We use the classical measure theory, which is easy. The real problem appears already for finite T , even for T = 20 . We have an abundance of probability measures on S T since we can prescribe P π for every π . When we require stationarity, we have ( T 1 ) ! equations for these T ! parameters, as shown below, which still leaves too many choices.
The objective is to select realistic P π . Most of the patterns, π , for m = 20 will never appear in any real-time series, and we could set P π = 0 , but we do not know for which π . There are three types of properties that we should require for our model.
  • Independence: The Markov property or k-dependence of patterns (cf. [12]).
  • Self-similarity: P π ( d ) should not depend on d . For processes with short-term memory, like Figure 1, this could be replaced by the requirement that P π ( d ) converge to the uniform distribution, exponentially with increasing d . There must be a law connecting P π ( d ) for different d . Otherwise, how can we describe them all?
  • Smoothness: There should not be too much zigzag in the time series. Patterns like 3142 should be exceptions. This can perhaps be reached by minimizing certain energy functions.
Problem. Does there exist on S 100 a stationary random order that is Markov and self-similar (for admissible d and π ) and has parameters different from white noise and Brownian motion; for instance, P 12 1 2 , or 1 3 < P 123 + P 321 < 1 2 ?

3.3. Random Order on N

The infinite time domain N has its merits. Order stationarity, for instance, is very easy to define since every pattern can be shifted to the right as far as we want. It is enough to require
P π t = P π t + 1   for   all   finite   patterns   π   and   t = 1 , 2 ,
It is even enough to require P η 1 = P η 2 for all patterns η of any length. (To prove that this implies (7) for a fixed t and a pattern π of length k , consider all patterns η of length m = t + k 1 , which show the pattern π on their last k positions. The sum of their probabilities P η 1 equals P π t , since P is a measure. And shifting all η from 1 to 2 means shifting π from t to t + 1 . )
On the other hand, infinite patterns require a limit T . Moreover, there are many recent papers on infinite permutations, that is, one-to-one mappings of N onto itself. An overview was given by Pitman and Tang [26]. However, an order on N is a much wider concept than a permutation on N . An infinite permutation π 1 , π 2 , defines an order on N , with the special property that below a given value π k , there are only finitely many other values, for any k .
For an order on N , however, the smallest object rarely exists. Usually, each object has infinitely many other objects below and above. Nevertheless, an order on N is uniquely defined by patterns π m S m , which represent the order of the first m elements, for m = 2 , 3 , for example, 1 2 2 3 1 3 4 1 2 4 5 1 3 2
Theorem 2
(Approximating random order on N ).
(i) 
A sequence of permutations π m S m defines an order on N if for m = 2 , 3 , , the pattern π m is represented by the first m values of the pattern π m + 1 .
(ii) 
A sequence of probability measures P m on S m defines a random order P on N if for m = 2 , 3 , , for π S m and π m + 1 S m + 1 , the following holds:
P m + 1 { π m + 1   s h o w s   p a t t e r n   π   a t   i t s   f i r s t   m   p o s i t i o n s } = P m ( π ) .
(iii) 
The random order P defined by the P m is stationary if and only if, for m = 2 , 3 , , for π S m , and π m + 1 S m + 1 , the following holds:
P m + 1 { π m + 1   s h o w s   p a t t e r n   π   a t   i t s   l a s t   m   p o s i t i o n s } = P m ( π ) .
Proof. 
(i): The condition says that the pattern of the first m objects, defined in step m , will not change during successive steps. So the construction is straightforward. The rank numbers, however, may change in each step, and they may converge to for m .
(ii): The condition says that the probability for a pattern π S m to appear for the first m objects, defined by P m , will remain the same for P m + 1 and successive probability measures. So we can define
P { the   objects   with   numbers   1 , , m   show   pattern   π } = P m ( π )
in a consistent way, and P m determines P . Moreover, P is an inverse limit of the measures P m . Below, we provide a more elementary argument.
(iii): Together with (ii), this condition says that P π 1 = P π 2 for all patterns π of any length m . As noted above, this is just the definition (7) of order stationarity. □

3.4. The Space of Random Orders

Resuming the discussion in Section 1.5, we show that the set of all orders on N can be represented by the unit interval, using a numeration system similar to our decimal numbers. We first assign subintervals of I = [ 0 , 1 ] to the permutations of length 2 , 3 , the pattern 12 corresponds to the interval [ 0 , 1 2 ] , and 21 to [ 1 2 , 1 ] . The permutations 123, 132, and 231, which exhibit the pattern 12 at their first two positions, correspond to [ 0 , 1 6 ] , [ 1 6 , 1 3 ] , and [ 1 3 , 1 2 ] , respectively. For intervals of length 4, see Figure 3.
Instead of the lexicographic order, we define a hierarchical order of permutations, with respect to the patterns shown by the first 2,3,…elements. For any permutation π = π 1 π m of length m and any k with 2 k m , let r k ( π ) denote the number of j < k with π j > π k . This is a kind of rank number of π k with values between 0 and k 1 . For example, r 3 ( 123 ) = 0 , r 3 ( 132 ) = 1 , and r 3 ( 231 ) = 2 while r 2 = 0 for these three patterns. Now, we assign to the permutation π of length m the following interval:
I ( π ) = [ x ( π ) , x ( π ) + 1 m ! ] with x ( π ) = k = 2 m r k k ! .
It is easy to check that the patterns π ( k ) = π 1 π k of the first k < m items of π are assigned to larger intervals, i.e.,
I I ( π ( 2 ) ) I ( π ( 3 ) ) I ( π ( m ) )
where π ( m ) = π . Moreover, π need not be a permutation, just a pattern—it could also be a numerical time series. Only the ordering of π j is used for defining r k and I ( π ) .
When we extend the pattern to the right, we obtain smaller nested subintervals, and for m a single point x = k = 2 r k k ! , which characterizes the limiting order of infinitely many objects. Thus, each order on N corresponds to a unique point x in [ 0 , 1 ] . This is very similar to decimal expansions, where we subdivide an interval into 10 subintervals. In the case of patterns, values r k are the digits, and we subdivide first into 2, then into 3, 4, 5,…intervals. The endpoints of intervals represent two orders on N , but this is an exception, as 0.5 and 0.4999 for the decimals.
Once we represent all orders on N as points in [ 0 , 1 ] , we can better understand the probability measures P 2 , P 3 , of Theorem 2 and the limiting probability measure P, which is called the random order on N . We start with the function F 1 ( x ) = 1 , which denotes uniform distribution on [ 0 , 1 ] . The function F m represents the measure P m , for m = 2 , 3 , for patterns π of length m, it is defined as the histogram of P m :
F m ( x ) = m ! · P m ( π )   for   x I ( π ) .
See Figure 3. The rectangle over I ( π ) has area P m ( π ) . In the case of white noise, F m = 1 for all m , and lim F m = F = 1 is the uniform distribution. We now show that such a limit exists for all sequences F 2 , F 3 , for which P 2 , P 3 , fulfill condition (ii) of Theorem 2. We reformulate (ii) as
I ( π ) F m + 1 ( x ) d x = m ! · P m ( π ) = I ( π ) F m ( x ) d x .
The second equation is obvious, and the first is best illustrated by an example with m = 3 and π = 312 . We have r 2 = r 3 = 1 , so I ( π ) = [ 4 6 , 5 6 ] . The possible extensions π 4 of π are 3124, 4123, 4132, and 4231. Their intervals of length 1 24 partition I ( π ) . Condition (ii) states that P 4 { 3124 , 4123 , 4132 , 4231 } = P 3 { 312 } . Thus, the four rectangles of F 4 over I ( π ) together have the same area as the one rectangle of F 3 . This is expressed in (10).
Equation (10) states that the F m form a martingale. The martingale convergence theorem implies that there is a limit function F , in the sense that 0 1 | F m F | d x converges to zero for m . This limit function is integrable and F = 1 . As a density function, it defines the probability measure P on all orders on N .
Our argument indicates that random orders on N belong to the realm of classical analysis and probability. Of course, the density function F is terribly discontinuous and can hardly be used to discuss stationarity. (An open challenge for experts is to identify examples for which the sequence F m converges in L 2 . )

3.5. Extension of Pattern Distributions

We conclude our paper with an optimistic outlook for practitioners. When you find a distribution of pattern probabilities of length 3 or 4, say, from data, say, there is no need to worry about extending it to longer finite or infinite patterns. Such an extension will always exist.
Theorem 3
(Markov extension of pattern probabilities). Any stationary probability measure P m on S m can be extended to a stationary probability measure P m + 1 on S m + 1 , and, hence, also to a stationary probability measure P on the space of random orders.
Proof. 
To show that P m + 1 is stationary, we only need to verify condition (iii) of Theorem 2; any pattern within { 1 , , m } can be shifted to the right, without changing its probability, until its maximum reaches m , by the stationarity assumption of P m . What remains is to shift the maximum from m to m + 1 , and this is condition (iii).
In the following extension formula, we use the convention that probability measures on permutations also apply to patterns, by replacing a pattern with its representing permutation. Let π = π 1 π 2 π m π m + 1 be a permutation in S m + 1 .
P m + 1 ( π ) = P m ( π 1 π m ) · P m ( π 2 π m + 1 ) P m ( π 2 π m )
This formula is used whenever there exists some π k with 2 k m between π 1 and π m + 1 . However, if π 1 and π m + 1 are neighboring numbers, then the right-hand side of (11) cannot distinguish π and π = π m + 1 π 2 π m π 1 . In such cases, both π and π are assigned half of the value of the right-hand side, in order to avoid double counting.
The denominator on the right refers to a pattern π 2 π m that has a representing permutation κ = κ 1 κ m 1 in S m 1 and can be extended in m ways to a permutation η = η 1 η m S m . Indeed, η m can be chosen from { 1 , , m } . For j < m , either κ j < η m and η j = κ j , or κ j η m and η j = κ j + 1 . Let us write
P m ( κ ) = P m { η | η 1 η m 1   shows   pattern   κ } η P m ( η ) .
In the numerator, π 1 π m is also a pattern that has a representing permutation λ = λ 1 λ m in S m . This term is P m ( λ ) .
We now prove that the defined P m + 1 fulfills condition (ii) of Theorem 2. We calculate
p = P m + 1 { π 1 π m + 1 | π 1 π m   is   represented   by   λ }
using the definition (11) of P m + 1 . There are m + 1 permutations π S m + 1 that fulfill the condition, differing in the value π m + 1 . For each case, π 2 π m + 1 is represented by one of the permutations η introduced above. However, the two cases π with | π 1 π m + 1 | = 1 belong to the same η . This is why we consider them together and divide their probability by two. Now,
p = P m ( λ ) P m ( κ ) · η P m ( η ) = P m ( λ ) .
This proves that P m + 1 extends P m . For the stationarity, the same proof has to be performed with the extension to the left and condition (iii). We must be careful that P m ( κ ) now refers to η 2 η m . However, since we assume that P m is stationary, this is the same number, and the proof runs as above. □
We call this a Markov extension since (11) states that π m + 1 does not depend on π 1 . Since we do not assume any independence properties of P m , we cannot expect more. However, if we start with P 2 ( 12 ) = P 2 ( 21 ) = 1 2 and extend successively to m = 3 , 4 , we obtain the coin-tossing order.
There are many other extensions. To provide an example, we can divide the double cases in an asymmetric way. A careful study of extensions may lead to a better model than coin tossing. But we have to stop here.

4. Conclusions and Outlook

Established models of stochastic processes do not say much about the probabilities of ordinal patterns. It has been suggested that models for ordinal pattern analysis can be found by algorithms for comparison and ranking of objects rather than by arithmetical operations. A paradigmatic example of an ordinal process without numerical values shows that this is possible. Properties like stationarity and self-similarity can be formulated in a weak and very natural way for ordinal processes. As a starting point for further work, we have proved a representation theorem and an extension theorem for stationary orders.
Our example can be improved and generalized in different ways. One way is to define ordinal processes directly from the extension theorem. Another way is to modify the algorithm so that order depends not only on coin tossing but also on already existing comparisons. Furthermore, we can consider the P π ( d ) of the coin-tossing order as d and prove the conjecture that this leads to an order self-similar process, similar to how the simple random walk leads to Brownian motion. On the practical side, we have to study large datasets and estimate the P π ( d ) for m = 4 and various values of d as accurately as possible, in order to classify different behaviors. The fundamental mathematical challenge is to understand connections between pattern probabilities for different scales, d .

Funding

This research received no external funding.

Data Availability Statement

Data are contained within the article. Further inquiries can be directed to the author.

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Bandt, C. Statistics and contrasts of order patterns in univariate time series. Chaos 2023, 33, 033124. [Google Scholar] [CrossRef] [PubMed]
  2. Keller, K.; Sinn, M. Ordinal analysis of time series. Phys. A 2005, 356, 114–120. [Google Scholar] [CrossRef]
  3. Keller, K.; Sinn, M.; Edmonds, J. Time Series from the ordinal viewpoint. Stochastics Dyn. 2007, 7, 247–272. [Google Scholar] [CrossRef]
  4. Amigo, J.M. Permutation Complexity in Dynamical Systems; Springer Series in Synergetics; Springer: Berlin/Heidelberg, Germany, 2010. [Google Scholar]
  5. Amigo, J.; Keller, K.; Kurths, J. Recent progress in symbolic dynamics and permutation complexity. Ten years of permutation entropy. Eur. Phys. J. Spec. Top. 2013, 222, 247–257. [Google Scholar]
  6. Zanin, M.; Zunino, L.; Rosso, O.; Papo, D. Permutation entropy and its main biomedical and econophysics applications: A review. Entropy 2012, 14, 1553–1577. [Google Scholar] [CrossRef]
  7. Levya, I.; Martinez, J.; Masoller, C.; Rosso, O.; Zanin, M. 20 years of ordinal patterns: Perspectives and challenges. arXiv 2022, arXiv:2204.12883. [Google Scholar] [CrossRef]
  8. Grimmett, G.; Stirzaker, D. Probability and Random Processes; Oxford University Press: Oxford, UK, 1995. [Google Scholar]
  9. Sinn, M.; Keller, K. Estimation of ordinal pattern probabilities in Gaussian processes with stationary increments. Comput. Stat. Data Anal. 2011, 55, 1781–1790. [Google Scholar] [CrossRef]
  10. Schnurr, A.; Dehling, H. Testing for structural breaks via ordinal pattern dependence. J. Am. Stat. Assoc. 2017, 112, 706–720. [Google Scholar] [CrossRef]
  11. Betken, A.; Buchsteiner, J.; Dehling, H.; Münker, I.; Schnurr, A.; Woerner, J.H. Ordinal patterns in long-range dependent time series. Scand. J. Stat. 2021, 48, 969–1000. [Google Scholar] [CrossRef]
  12. de Sousa, A.R.; Hlinka, J. Assessing serial dependence in ordinal patterns processes using chi-squared tests with application to EEG data analysis. Chaos 2022, 32, 073126. [Google Scholar] [CrossRef]
  13. Weiß, C. Non-parametric tests for serial dependence in time series based on asymptotic implementations of ordinal-pattern statistics. Chaos 2022, 32, 093107. [Google Scholar] [CrossRef] [PubMed]
  14. Nüßgen, I.; Schnurr, A. Ordinal pattern dependence in the context of long-range dependence. Entropy 2021, 23, 670. [Google Scholar] [CrossRef] [PubMed]
  15. Betken, A.; Wendler, M. Rank-based change-point analysis for long-range dependent time series. Bernoulli 2022, 28, 2209–2233. [Google Scholar] [CrossRef]
  16. Olivares, F. Ordinal language of antipersistent binary walks. Phys. Lett. A 2024, 527, 130017. [Google Scholar] [CrossRef]
  17. Bandt, C.; Shiha, F. Order patterns in time series. J. Time Ser. Anal. 2007, 28, 646–665. [Google Scholar] [CrossRef]
  18. Elsinger, H. Independence Tests Based on Symbolic Dynamics; Working Paper No. 150; Österreichische Nationalbank: Wien, Austria, 2010. [Google Scholar]
  19. DeFord, D.; Moore, K. Random Walk Null Models for Time Series Data. Entropy 2017, 19, 615. [Google Scholar] [CrossRef]
  20. Weiß, C.; Martin, M.R.; Keller, K.; Matilla-Garcia, M. Non-parametric analysis of serial dependence in time series using ordinal patterns. Comput. Stat. Data Anal. 2022, 168, 107381. [Google Scholar] [CrossRef]
  21. Weiß, C.; Schnurr, A. Generalized ordinal patterns in discrete-valued time series: Nonparametric testing for serial dependence. J. Nonparametric Stat. 2024, 36, 573–599. [Google Scholar] [CrossRef]
  22. Bandt, C. Order patterns, their variation and change points in financial time series and Brownian motion. Stat. Pap. 2020, 61, 1565–1588. [Google Scholar] [CrossRef]
  23. Embrechts, P.; Maejima, M. Selfsimilar Processes; Princeton University Press: Princeton, NJ, USA, 2002. [Google Scholar]
  24. Bandt, C.; Pompe, B. Permutation entropy: A natural complexity measure for time series. Phys. Rev. Lett. 2001, 88, 174102. [Google Scholar] [CrossRef]
  25. Keller, G. Equilibrium States in Ergodic Theory; London Mathematical Society Student Texts; Cambridge University Press: Cambridge, UK, 1998; Volume 42. [Google Scholar]
  26. Pitman, J.; Tang, W. Regenerative random permutations of integers. Ann. Prob. 2019, 47, 1378–1416. [Google Scholar] [CrossRef]
Figure 1. Histograms of pattern frequencies of a simulated time series for d = 1 and m = 2 (red), 3 (blue), and 4 (brown). Each pattern π of length m is represented by a subinterval of [ 0 , 1 ] with length 1 / m ! . The corresponding bar has area p π ( d ) , so the height is m ! p π ( d ) . The permutations are arranged in hierarchical order so that each histogram is a refinement of the previous one. For details, see Section 3.4.
Figure 1. Histograms of pattern frequencies of a simulated time series for d = 1 and m = 2 (red), 3 (blue), and 4 (brown). Each pattern π of length m is represented by a subinterval of [ 0 , 1 ] with length 1 / m ! . The corresponding bar has area p π ( d ) , so the height is m ! p π ( d ) . The permutations are arranged in hierarchical order so that each histogram is a refinement of the previous one. For details, see Section 3.4.
Entropy 27 00610 g001
Figure 2. Global and local rank numbers obtained from coin tossing.
Figure 2. Global and local rank numbers obtained from coin tossing.
Entropy 27 00610 g002
Figure 4. Probabilities of patterns of length 4 for the coin-tossing order with d = 2 and d = 3 . They are clearly different from d = 1 while the changes from d = 2 to d = 3 are small.
Figure 4. Probabilities of patterns of length 4 for the coin-tossing order with d = 2 and d = 3 . They are clearly different from d = 1 while the changes from d = 2 to d = 3 are small.
Entropy 27 00610 g004
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bandt, C. Ordinal Random Processes. Entropy 2025, 27, 610. https://doi.org/10.3390/e27060610

AMA Style

Bandt C. Ordinal Random Processes. Entropy. 2025; 27(6):610. https://doi.org/10.3390/e27060610

Chicago/Turabian Style

Bandt, Christoph. 2025. "Ordinal Random Processes" Entropy 27, no. 6: 610. https://doi.org/10.3390/e27060610

APA Style

Bandt, C. (2025). Ordinal Random Processes. Entropy, 27(6), 610. https://doi.org/10.3390/e27060610

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop