Preprint
Article

This version is not peer-reviewed.

Ordinal Random Processes

A peer-reviewed article of this preprint also exists.

Submitted:

06 May 2025

Posted:

07 May 2025

You are already at the latest version

Abstract
For applications in many fields, ordinal patterns are a successful tool. Here we address the need for theoretical models. A paradigmatic example shows that a model for frequencies of ordinal patterns can be determined without any numerical values. We specify the important concept of stationary order and the fundamental problems to be solved in order to establish a genuine statistical methodology for ordinal time series.
Keywords: 
;  ;  ;  

1. Introduction

1.1. Contents of the Paper

In this introductory section, we introduce the basic concepts in a brief manner. For more detail and background, we refer to ([1] Section II) and [2,3,4,5,6,7]. A new point is the visualization of the distribution of ordinal patterns as refining ordinal histogram in Section 1.4. The main topic of the paper is explained in Section 1.6. Order-related properties of processes, in particular order self-similarity, are introduced in Section 1.7.
Section 2 deals with a paradigmatic example of an ordinal process which has no numerical values. The construction is algorithmic and intuitive, and the probabilities for patterns of length 3 are the same as for Brownian motion. Unfortunately, the example is not self-similar and not relevant for applications. We conjecture that there is a self-similar modification.
The final Section 3 gives a more technical outline of a possible theory of ordinal processes. We construct stationary ordinal processes without using numerical values. Order self-similarity remains an open problem.

1.2. Permutations as Patterns in Time Series

A permutation π of length m is a one-to-one mapping from the set { 1 , 2 , . . . , m } onto itself. We write π k for π ( k ) and π = π 1 π 2 . . . π m . Thus π = 231 means π ( 1 ) = 2 , π ( 2 ) = 3 , π ( 3 ) = 1 . However, we do not need mapping properties of permutations like composition. We just consider the graph of π as a geometric pattern - an ordinal pattern.
With x = x 1 , x 2 , . . . , x T we denote a time series of length T . It is also a mapping, assigning to the time points t = 1 , . . . , T the values x t = x ( t ) . For m time points t 1 < t 2 < . . . < t m between 1 and T we consider the pattern of corresponding values. We define
( x t 1 , x t 2 , . . . , x t m )   shows   pattern   π   when x t i < x t j if   and   only   if   π i < π j .
In other words, the correspondence x t j π j , j = 1 , . . . , m is strictly monotonous. This is very intuitive. The pattern π can be determined from the time series by calculating ranks: π k is the number of time points t j for which x j x k .

1.3. Pattern Frequencies in Time Series

Instead of arbitrary t j , we study equally spaced time points t < t + d < t + 2 d < . . . < t + ( m 1 ) d . We call m the length and d the delay or lag of the pattern. We always take m 6 and mainly consider m = 3 and m = 4 . However, we vary d as much as possible.
Ordinal pattern analysis is a statistical method. We fix a length m and determine relative frequencies of all patterns π of length m , for various delays d . We divide the number of occurences of π by the number of time points where π could occur.
p π ( d ) = 1 T ( m 1 ) d # { t | 1 t T ( m 1 ) d , ( x t , x t + d , . . . , x t + ( m 1 ) d )   shows   pattern   π } .
These frequencies are the basic parameters of ordinal time series analysis. They can be combined to calculate permutation entropy, persistence, and other important parameters [1,2,3,4,5,6,7].

1.4. Visualization of Pattern Frequencies by an Ordinal Histogram

This paper is not about parameters like entropy. It studies models for all pattern frequencies together. For this reason, we here introduce a graphical method to subsume the pattern frequencies for fixed d in a kind of density function. The permutations π are assigned subintervals I π of [ 0 , 1 ] as follows: I 12 = [ 0 , 1 2 ] , I 21 = [ 1 2 , 1 ] . We draw bars over I π which have area p π ( d ) . For m = 3 , we note that p 12 = p 123 + p 132 + p 231 (neglecting a possible tiny error due to the last value for p 12 ( d ) in (2)). We choose I 123 = [ 0 , 1 6 ] , I 132 = [ 1 6 , 1 3 ] , I 231 = [ 1 3 , 1 2 ] so that their union is I 12 . Then we again draw bars with area p π ( d ) . The same is done for π = 213 , 312 , 321 . In this way, the histogram for m = 3 is a refinement of the histogram for m = 2 . Sukzessive refinements can be obtained for m = 4 , 5 , . . . although the length of the time series will rarely allow reliable estimates for m > 4 . This arrangement of permutations is called hierarchical order. Details will be given in Section 3.4. As Figure 1 shows for a simulated time series, all pattern frequencies of a time series for a fixed d can be subsumed in this way by a sequence of refining histograms, or in the limit m by a kind of density function.
The process chosen for Figure 1 is an AR(1) model with exponential noise: X t + 1 = 1 2 X t log U t , where the U t are independent uniformly distributed random numbers in [ 0 , 1 ] . Due to the asymmetry of the noise distribution, p 21 ( 1 ) 0 . 58 > 1 2 . Sukzessive decreasing steps are frequent: p 321 ( 1 ) 0 . 3 and p 4321 ( 1 ) 0 . 13 . We cannot explain why 2143 is the permutation with the second largest frequency. However, an AR(1) model with coefficient 1 2 has a short memory. For d = 3 the patterns already have almost uniform frequencies. The models considered below are more uniform for m = 2 and 3, but they provide interesting behavior also for large d .
The problem of this paper is to find typical ordinal histograms from a theoretical viewpoint. An important question is to find models where the histograms agree for all d . Two such models are known ([1] Section II.C): white noise, where the histogram is a constant function for all m , and Brownian motion, see Figure 3 below. We are looking for other models.

1.5. Stationary and Order Stationary Processes

In time series analysis, we do always assume that the given time series is an instance of a larger ensemble, called a stochastical process. The process represents the mechanism which generates the given time series and many other ones. We want to know the laws of this process. Formally, a stochastical process X is a sequence X = X 1 , X 2 , . . . of real random variables on a probability space ( Ω , P ) . The time series x is a random choice from the ensemble X .
To find the laws of X from one single time series, we must assume that the distributions and dependencies of the X k do not change in time. Usually X is required to be stationary. That is, the m-dimensional distribution of ( X t , X t + 1 , . . . , X t + m ) does not depend on t , for all m . This assumption is very strong. When we study autocorrelation, we assume weak stationarity. That is, the mean M ( X t ) and the covariance C o v ( X t , X t + k ) do not depend on t [8]. When we consider ordinal patterns, we assume a similar very weak stationarity assumption. It will be required throughout this paper.
Order stationarity. For a given pattern π with length m and delay d , let
P π ( d ) = P { ( X t , X t + d , . . . , X t + ( m 1 ) d )   shows   pattern   π }
denote the probability that the process X shows pattern π just after time t . We say that the process X is order stationary for patterns of length m if P π ( d ) is the same value for all t . This should hold for all patterns π of length m and all delays d = 1 , . . . , d max where d max is chosen much smaller than the length T of our time series.
Brownian motion (where X 0 = 0 and X t + 1 X t are independent increments with standard normal distribution for t = 0 , 1 , . . . ) is an order stationary process which is not even weakly stationary. It is a basic model for financial time series. Note that the process X and the pattern probabilities P π ( d ) are theoretical objects. The observed frequency p π ( d ) is an estimate for P π ( d ) , and from the viewpoint of theory, the estimator formula (2) has nice properties [9,10,11,12,13]. The practical success of ordinal patterns indicates that the assumption of order-stationarity is at least approximately fulfilled in applications.

1.6. The topic of this paper.

There are plenty of well-studied stochastic processes: Gaussian processes, random walks, ARMA and GARCH processes, processes with long-range dependence [11,14] etc. Most of them are either stationary or have stationary (multidimensional) increments and so are order stationary. They are defined by arithmetical operations, however. This makes the rigorous study of ordinal pattern frequencies difficult although various interesting theorems could be derived [9,10,11,15,16]. Even for Brownian motion, exact frequencies of patterns can be determined only for m 4 [17]. Some of them are rational, some irrational. Frequencies for length 5 lead to integrals which can be only numerically calculated. On the whole, the impression is that ordinal patterns do not fit into the established theory of stochastic processes.
The only proper theoretical model for pattern frequencies is white noise, a process of independent and identically distributed random variables X i . For each m , all permutations have the same probability p π = 1 / m ! for any delay d . This model was repeatedly used as null hypothesis for testing serial dependence of time series [1,12,13,18,19,20,21]. A greater variety of models is needed.
In this paper, we try to define ordinal random processes in a combinatorial way, directly by their ordinal pattern frequencies. We do not require normal distribution, we require no arithmetics, and we even do not need numerical values to compare objects X i and X j . An algorithmic example in the next chapter will show that this is possible. In Section 3, we outline some technical concepts and basic statements for a theory of ordinal processes. This could be a first step towards finding more applicable ordinal models. Note that possible equality of values, an important problem in practice [21], plays no role in this paper: our models will only allow X s < X t or X s > X t but not X s = X t .

1.7. Symmetry and Independence Properties of Ordinal Processes

Of course it is not enough to just postulate pattern frequencies. Our models should also have nice properties. Here we list the ordinal version of some of the most common process properties. We start with symmetry in time and space [9].
A process X is reversible if the distributions of ( X 1 , . . . , X m ) and ( X m , . . . X 1 ) agree for every m . In the ordinal version, it is reversible if the permutations π = π 1 π 2 . . . π m and π m π m 1 . . . π 1 have the same pattern probability for every π . The process X is invariant under reversal of values if the distributions of ( X 1 , . . . , X m ) and ( X 1 , . . . X m ) agree for every m . The ordinal version states that the permutations π = π 1 π 2 . . . π m and ( m + 1 π 1 ) ( m + 1 π 2 ) . . . ( m + 1 π m ) have the same pattern probability for every π .
In both cases, the ordinal versions of symmetry are much weaker and much easier to check. These symmetry properties are often met in applications. As a consequence, the distribution of ordinal patterns of length 3 often has only one degree of freedom, since p 123 = p 321 , and the probabilities of the other four permutations agree. All Gaussian processes are symmetric in space and time [9,11,17].
A process X is said to be Markov if for any t and fixed X t , the distributions of ( X 1 , . . . , X t 1 ) and ( X t + 1 , X t + 2 , . . . ) are independent. In the ordinal version, this means that the appearance of patterns π , π at times t 1 < . . . < t m t and t t 1 < . . . < t m , respectively, are independent. Again, the ordinal property is weaker and easier to check. The Markov property appears in theory rather than in applications. It implies that P 12 . . . m = ( P 12 ) m which is too large to occur in practice, even in financial data.
Order self-similarity. The most important symmetry property in this paper is order self-similarity ([22] Section 6). A random process in continuous time, Y t , t > 0 is said to be self-similar if there is an exponent H > 0 such that Y r t = r H Y t in distribution, for each positive number r [23]. Such a process cannot be stationary, it must look similar to Brownian motion. The ordinal concept is weaker and much simpler.
A process X is said to be order self-similar for patterns π of length m if
P π ( d )   is   the   same   value   for   all   d .
This should hold for all patterns π of length m and all delays d = 1 , . . . , d max where d max is chosen much smaller than the length T of our time series.
Order self-similarity was verified in financial time series [22] and EEG measurements [1]. Our paper is motivated by the desire for models of this behavior.

2. The Coin-Tossing Order—An Ordinal Process Without Values

2.1. Ranking Fighters by Their Strength

Ordinal patterns were derived from numerical values of a time series. Now we go the opposite way, as in the world of tennis or chess where players first pairwise compare each other and afterwards are assigned a rank or ELO score. Suppose there are four players A , B , C , D in a tennis club. We design an algorithm to order them by strength. First B plays against A . Let us assume B wins. Then we write A < B . Now the next player C plays against the previous player B . If C wins, then A < B < C by transitivity of order. If C loses, however, C must still play against A . Depending on the result, either A < C < B or C < A < B . Now the last player D comes and plays first against the previous player C , and then against B and/or A , if the comparison is not yet determined by previous games and transitivity.
This procedure will be turned into a mathematical model. We want to construct a random order between objects X 1 , X 2 , . . . which are not numbers. We follow the algorithm for tennis players, and replace each match by the toss of a fair coin. As a result we obtain a stochastic process like white noise or Brownian motion, on ordinal level only. It will be possible to determine pattern probabilities much better than for Brownian motion.

2.2. Definition of the Coin Tossing Order

Repeated toin cossing is a standard way to represent randomness. Take a simple random walk, for example, and write X n for our position at time n = 0 , 1 , 2 , . . . Then X 0 = 0 , and either X n + 1 = X n + 1 or X n + 1 = X n 1 depending on the result of a coin throw at time n . In the present case, X 1 , X 2 , X 3 , . . . will denote objects, not necessarily numbers. We define a random order between the X i . We throw a fair coin c j i to decide whether X i < X j or X i > X j , for any pair of integers i < j . Let us write 1 for `head’ and 0 for `tail’. Then our basic probability space Ω is the space of all 0-1-sequences, where each coordinate is 0 or 1 with probability 1 2 , independently of all other coordinates. Formally,
Ω = { ( c 21 , c 32 , c 31 , c 43 , c 42 , . . . ) | c j i { 0 , 1 }   for   j > i 1 }
where c j i = 0 means X i < X j , and c j i = 1 means X i > X j , for 1 i < j . The important point is that c j i will be disregarded when the order of X i and X j is already fixed by previous comparisons and transitivity.
The first coin c 21 decides the ordering of X 1 and X 2 . Now suppose X 1 , . . . , X j 1 are already ordered. Then X j is compared to X j 1 , X j 2 , . . . , X 1 by considering the random numbers c j i . However, when the comparison is fixed by transitivity from the already defined ordering, then c j i is disregarded - that coin need not be thrown.
The resulting random order will be called coin tossing order. It can be easily simulated. Figure 2 shows rank numbers of 500 consecutive objects X j in the middle of a simulated series of length T = 10000 . The global rank numbers have strange discontinuities. Local rank numbers, obtained by comparing with the next 20 objects on the left and right, show a more familiar picture.
Problem. Is there a stochastic process with stationary increments which generates the coin tossing order?

2.3. Basic Properties of the Coin Tossing Order

Despite the erratic view of trajectories in Figure 2, it turns out that the coin-tossing order has the same ordinal pattern frequencies for length 3 as Brownian motion, for arbitrary d . In particular, it is order self-similar for patterns of length 3.
Theorem 1 
(Basic properties of the coin tossing order).
(i) 
The coin tossing order is stationary and has the ordinal Markov property.
(ii) 
For any permutation π of length m , the pattern probability is P π ( 1 ) = 2 u with
u = u ( π ) = # { ( i , j ) | 1 i < j m ,   i f   i < k < j   t h e n   π k   i s   n o t   b e t w e e n   π i   a n d   π j } .
(iii) 
The pattern probabilities P π ( d ) are invariant under time reversal and reversal of the values.
(iv) 
For d = 1 , 2 , . . . . we have P 12 ( d ) = 1 2 , P 123 ( d ) = P 321 ( d ) = 1 4 , and P π ( d ) = 1 8 for the other permutations of length 3.
Proof. (i): The order of ( X t , X t + 1 , . . . , X t + m 1 ) depends on the random numbers c j i , t + m 1 j > i t , in the same way as the order of ( X 1 , X 2 , . . . , X m ) depends on the c j i , m j > i 1 . Since both collections of random numbers have the same distribution, the pattern probabilities do not depend on t , and the defined random order is stationary. Moreover, the comparisons of X s with s t depend on random numbers c j i with j > t while the comparisons of X s with s t depend on c j i with j t . Since different c j i are independent, this implies the ordinal Markov property.
(ii): u ( π ) is the number of coin flips needed to determine π . Given a permutation π , we determine the c j i which are needed to define the occurence of π in ( X 1 , X 2 , . . . , X m ) . Of course m j > i 1 , and j runs in increasing order. For fixed j , the number c j , j 1 is always used, and the other i are considered in decreasing order. Now consider some k with i < k < j . If π i < π k < π j , this was determined by the random numbers c k i and c j k which were drawn before c j i . In that case c j i is disregarded since π i < π j follows from the transitivity of the order. Similar for π i > π k > π j . However, if there is no π k between π i and π j , then c j i is needed to determine π . We shall call u ( π ) the energy of π .
(iii): First we consider d = 1 . Given π , we have to show that the time-reversed and spatially reversed permutations have the same energy u ( π ) . For spatial reversal this directly follows from the definition (5) with `between’. For time reversal, we show that P π ( 1 ) can also be determined backwards by considering the c j i with decreasing i = m 1 , m 2 , . . . , 1 , and for fixed i , with increasing j = i + 1 , . . . , m . The point is that when we compare places i < j , we have already compared both j and i with all k between. So c i j is needed only if no π k is between π i and π k . Otherwise the order between π i and π j is already fixed, and c i j is disregarded. This proves reversibility for d = 1 .
Now consider d > 1 and some π of length m which appears as pattern of ( X 1 , X 1 + d , . . . , X 1 + ( m 1 ) d ) . The probability of this event is P π ( d ) which may be different from P π ( 1 ) . We have to calculate P π ( d ) from all `atom permutations’ of length M = 1 + ( m 1 ) d for ( X 1 , X 2 , . . . , X 1 + ( m 1 ) d ) for which the order among the special m places agrees with the order of π . This would be a lot of work. However, since the reversed permutation of π is composed of the reversed `atom permutations’, both in space and in time, we shall get the same pattern probability.
(iv): The spatial reversal invariance implies P 12 ( d ) = P 21 ( d ) = 1 2 . Then P 123 ( d ) = P 321 ( d ) = 1 4 follows from the Markov property. The equality of the four other pattern frequencies is a consequence of the invariance under time and space reversal. □

2.4. Computer Work and Breakdown of Self-Similarity

In Figure 3 we compare pattern probabilities of length 4 for Brownian motion and coin tossing order. Brownian motion looks more interesting since coin-tossing allows only the probabilities 2 k . Our model is a first attempt, not recommended for applications. For length 3, the P π ( d ) of the two processes agree and do not depend on d , according to Theorem 1. The figure shows that we must investigate patterns of length 4 if we want to distinguish processes like these two. Moreover, Brownian motion is perfectly self-similar so that the probabilities P π ( d ) of any length do not depend on d [23]. Is that also true for the coin tossing order?
Figure 3. Probabilities of patterns of length 4 (brown) on top of those of length 3 (blue) for Brownian motion (left) and coin tossing order with d = 1 (right). For length 3, the probabilities coincide.
Figure 3. Probabilities of patterns of length 4 (brown) on top of those of length 3 (blue) for Brownian motion (left) and coin tossing order with d = 1 (right). For length 3, the probabilities coincide.
Preprints 158559 g003
The Equation (5) allows fast calculation of all pattern probabilities for d = 1 and m 10 . So we can determine P π ( d ) for all permutations of length 4 and d = 2 and 3 by adding the probabilities of `atom permutations’ of length 7 and 10, respectively, as sketched in the proof of Theorem 1, (iii). For d = 3 we have to add 10 ! / 4 ! = 151200 cases for each permutation.
The result is shown in Figure 4. Unfortunately, order self-similarity breaks down for patterns of length 4. For π = 1324 we have P π ( 1 ) = 1 32 3 . 12 % while P π ( 2 ) 3 . 42 % and P π ( 3 ) 3 . 49 % . However, the probabilities for d = 2 or 3 can be taken as a new model which is nearer to self-similarity and looks not as artificial as coin tossing in Figure 3.
Problem. Let us define new pattern probabilities Q π ( 1 ) = P π ( d ) for d = 2 , 3 , . . . What are the properties of Q ? Is there a limit for d ?
The small difference between d = 2 and d = 3 in Figure 4 indicates fast convergence. The mere existence of an order self-similar limit would be interesting but not yet helpful. Brownian motion is already available as an order self-similar process. The merit of the coin tossing order is its algorithmic flavour, its connection with ranking. We look for an order self-similar model with an intuitive explanation.
Finally, let us consider the interpretation of u ( π ) as an energy function. Let d = 1 , and let S m denote the set of all permutations of length m . Let P denote the probability measure on S m given by the probabilities P π = 2 u ( π ) of coin tossing order. The permutation entropy [24] of P , using logarithms with base 2, is just the mean energy with respect to P .
H ( P ) = π P π log 2 P π = π P π u ( π ) = M P ( u ) .
For all other probability measures Q on S m , the permutation entropy is smaller than the mean energy, see ([25] Chapter 1). Thus P is a so-called Gibbs measure on S m .
Problem. Are there other meaningful energy functions for permutations, perhaps even parametric families of Gibbs measures on S m ?

3. Rudiments of a Theory of Ordinal Processes

3.1. Random Stationary Order

Having studied one example, we turn to discuss a possible theory of ordinal random processes. As above, X 1 , X 2 , . . . will not be numbers, only objects which are ordered. We have the discrete time domain { 1 , . . . , T } for time series and N = { 1 , 2 , . . . } for models. An order will be a relation < on the time domain such that x ¬ < x , and x < y implies y ¬ < x , and x < y plus y < z implies x < z . Of course, the order does not apply to the time points t , but to the corresponding objects x t . Then the property that x t , . . . , x t + m 1 shows pattern π has a meaning for every π S m – it is either true or false. When an order on N or { 1 , . . . , T } is given, we can determine pattern frequencies, permutation entropy and so on. In the following, we always have d = 1 .
We want to construct models of ordinal processes, like the coin tossing algorithm. For this purpose we need the concept of random order. To keep things simple, a random order is defined as a probability measure on the set of all orders on the time domain. For the finite time domain { 1 , . . . , T } , a random order is just a probability measure P T on the set S T of permutations of length T . For m T and π S m and 1 t T + 1 m the random order allows to determine the probability
P π t = P { x t , . . . , x t + m 1 shows   pattern π }
The random order will be called stationary if the P π t do not depend on the time point t , for any pattern π of any length m < T . In other words, the numbers P π t must be the same for all admissible t . This is exactly the order stationarity which we defined for numerical processes in (3).

3.2. The Problem to Find Good Models

The infinite time domain N is considered in the next section. Using classical measure theory, that will be easy. The real problem appears already for finite T , even for T = 20 . We have an abundance of probability measures on S T since we can prescribe P π for every π . When we require stationarity, we have ( T 1 ) ! equations for these T ! parameters, as shown below, which is still too much choice.
The problem is to select realistic P π . Most of the patterns π for m = 20 will never appear in any real time series, and we could set P π = 0 . But we do not know for which π . There are three types of properties which we should require for our model.
  • Independence: Markov property or k-dependence of patterns (cf. [12]).
  • Self-similarity: P π ( d ) should not depend on d . For processes with short-term memory, like Figure 1, this could be replaced by the requirement that the P π ( d ) converge to the uniform distribution, exponentially with increasing d . There must be a law connecting the P π ( d ) for different d . Otherwise, how can we describe them all?
  • Smoothness: There should not be too much zigzag in the time series. Patterns like 3142 should be exceptions. This can perhaps be reached by minimizing certain energy functions.
Problem. Does there exist on S 100 a stationary random order which is Markov and self-similar (for admissible d and π ) and has parameters different from white noise and Brownian motion? For instance P 12 1 2 , or 1 3 < P 123 + P 321 < 1 2 ?

3.3. Random Order on N

The infinite time domain N has its merits. Order stationarity, for instance, is very easy to define since every pattern can be shifted to the right as far as we want. It is enough to require
P π t = P π t + 1 for   all   finite   patterns   π   and   t = 1 , 2 , . . .
It is even enough to require P η 1 = P η 2 for all patterns η of any length. (To prove that this implies (7) for a fixed t and a pattern π of length k , consider all patterns η of length m = t + k 1 which show the pattern π on their last k positions. The sum of their probabilities P η 1 equals P π t since P is a measure. And shifting all η from 1 to 2 means shifting π from t to t + 1 . )
On the other hand, infinite patterns require a limit T . Actually, there are lots of recent papers on infinite permutations, that is, one-to-one mappings of N onto itself. An overview is given in Pitman and Tang [26]. However, an order on N is a much wider concept than a permutation on N . An infinite permutation π 1 , π 2 , . . . defines an order on N , with the special property that below a given value π k there are only finitely many other values, for any k .
For an order on N , however, there rarely exists a smallest object. Usually each object has infinitely many other objects below and above. Nevertheless, an order on N is uniquely defined by patterns π m S m which represent the order of the first m elements, for m = 2 , 3 , . . . For example 1 2 2 3 1 3 4 1 2 4 5 1 3 2 . . .
Theorem 2 
(Approximating random order on N ).
(i) 
A sequence of permutations π m S m defines an order on N if for m = 2 , 3 , . . . , the pattern π m is represented by the first m values of the pattern π m + 1 .
(ii) 
A sequence of probability measures P m on S m defines a random order P on N if for m = 2 , 3 , . . . , for π S m and π m + 1 S m + 1 holds
P m + 1 { π m + 1   s h o w s   p a t t e r n   π   a t   i t s   f i r s t   m   p o s i t i o n s } = P m ( π ) .
(iii) 
The random order P defined by the P m is stationary if and only if for m = 2 , 3 , . . . , for π S m and π m + 1 S m + 1 holds
P m + 1 { π m + 1   s h o w s   p a t t e r n   π   a t   i t s   l a s t   m   p o s i t i o n s } = P m ( π ) .
Proof. (i): The condition says that the pattern of the first m objects, defined in step m , will not change during successive steps. So the construction is straightforward. The rank numbers, however, may change in each step, and they may converge to for m .
(ii): The condition says that the probability for a pattern π S m to appear for the first m objects, defined by P m , will remain the same for P m + 1 and successive probability measures. So we can define
P { the   objects   with   numbers   1 , . . . , m   show   pattern   π } = P m ( π )
in a consistent way, and the P m determine P . Actually, P is an inverse limit of the measures P m . Below, we provide a more elementary argument.
(iii): Together with (ii), this condition says that P π 1 = P π 2 for all patterns π of any length m . As noted above, this is just the definition (7) of order stationarity. □

3.4. The Space of Random Orders

Resuming the discussion in Section 1.4, we show that the set of all orders on N can be represented by the unit interval, using a numeration system similar to our decimal numbers. We first assign subintervals of I = [ 0 , 1 ] to the permutations of length 2 , 3 , . . . The pattern 12 will correspond to [ 0 , 1 2 ] , and 21 to [ 1 2 , 1 ] . The permutations 123, 132, and 231, which show the pattern 12 at their first two places, will correspond to [ 0 , 1 6 ] , [ 1 6 , 1 3 ] , and [ 1 3 , 1 2 ] , respectively. For intervals of length 4, see Figure 3.
Instead of lexicographic order, we define a hierarchical order of permutations, with respect to the patterns shown by the first 2,3,... elements. For any permutation π = π 1 . . . π m of length m and any k with 2 k m , let r k ( π ) denote the number of j < k with π j > π k . This is a kind of rank number of π k with values between 0 and k 1 . For example r 3 ( 123 ) = 0 , r 3 ( 132 ) = 1 , and r 3 ( 231 ) = 2 while r 2 = 0 for these three patterns. Now we assign to the permutation π of length m the following interval:
I ( π ) = [ x ( π ) , x ( π ) + 1 m ! ] w i t h x ( π ) = k = 2 m r k k ! .
It is easy to check that the patterns π ( k ) = π 1 . . . π k of the first k < m items of π are assigned to larger intervals:
I I ( π ( 2 ) ) I ( π ( 3 ) ) . . . I ( π ( m ) )
where π ( m ) = π . Actually, π need not be a permutation, just a pattern - it could also be a numerical time series. Only the ordering of the π j is used for defining r k and I ( π ) .
When we extend the pattern to the right, we get smaller nested subintervals, and for m a single point x = k = 2 r k k ! which characterizes the limiting order of infinitely many objects. Thus each order on N corresponds to a unique point x in [ 0 , 1 ] . This is very similar to decimal expansions where we subdivide an interval into 10 subintervals. In case of patterns, the r k are the digits, and we subdivide first into 2, then 3, 4, 5,... intervals. The endpoints of intervals represent two orders on N , but this is an exception, as 0.5 and 0 . 4999 . . . for the decimals.
Once we have represented all orders on N as points in [ 0 , 1 ] , we can better understand the probability measures P 2 , P 3 , . . . of Theorem 2 and the limiting probability measure P which is called random order on N . We start with the function F 1 ( x ) = 1 which denotes uniform distribution on [ 0 , 1 ] . The function F m will represent the measure P m , for m = 2 , 3 , . . . For patterns π of length m it is defined as histogram of P m :
F m ( x ) = m ! · P m ( π ) f o r x I ( π ) .
See Figure 3. The rectangle over I ( π ) has area P m ( π ) . In case of white noise, F m = 1 for all m , and the lim F m = F = 1 is the uniform distribution. We now show that such limit exists for all sequences F 2 , F 3 , . . . for which the P 2 , P 3 , . . . fulfil condition (ii) of Theorem 2. We reformulate (ii) as
I ( π ) F m + 1 ( x ) d x = m ! · P m ( π ) = I ( π ) F m ( x ) d x .
The second equation is obvious, and the first is best shown by example, for m = 3 and π = 312 . We have r 2 = r 3 = 1 , so I ( π ) = [ 4 6 , 5 6 ] . The possible extensions π 4 of π are 3124, 4123, 4132, and 4231. Their intervals of length 1 24 partition I ( π ) . Condition (ii) says that P 4 { 3124 , 4123 , 4132 , 4231 } = P 3 { 312 } . Thus the four rectangles of F 4 over I ( π ) together have the same area as the one rectangle of F 3 . This is expressed in (10).
Equation (10) says that the F m form a martingale. The martingale convergence theorem implies that there is a limit function F , in the sense that 0 1 | F m F | d x converges to zero for m . This limit function is integrable and F = 1 . As a density function, it defines the probability measure P on all orders on N .
Our argument indicates that random orders on N belong to the realm of classical analysis and probability. Of course, the density function F will be terribly discontinuous and can hardly be used to discuss stationarity. (Open problem for experts: find examples for which the F m converge in L 2 . )

3.5. Extension of Pattern Distributions

We conclude our paper with an optimistic outlook for practitioners. When you find a distribution of pattern probabilties of length 3 or 4, from data say, you need not care for an extension to longer finite and infinite patterns. Such an extension will always exist.
Theorem 3 
(Markov extension of pattern probabilities). Any stationary probability measure P m on S m can be extended to a stationary probability measure P m + 1 on S m + 1 , and hence also to a stationary probability measure P on the space of random orders.
Proof. 
To show that P m + 1 is stationary, we need only verify condition (iii) of Theorem 2: Any pattern within { 1 , . . . , m } can be shifted to the right, without changing its probability, until its maximum reaches m , by assumption of stationarity of P m . It remains to shift the maximum from m to m + 1 , and this is condition (iii).
In the following extension formula, we use the convention that probability measures on permutations also apply to patterns, by replacing a pattern with its representing permutation. Let π = π 1 π 2 . . . π m π m + 1 be a permutation in S m + 1 .
P m + 1 ( π ) = P m ( π 1 . . . π m ) · P m ( π 2 . . . π m + 1 ) P m ( π 2 . . . π m )
This formula will be used whenever there exists some π k with 2 k m between π 1 and π m + 1 . However, if π 1 and π m + 1 are neighboring numbers, then the right-hand side of (11) cannot distinguish π and π = π m + 1 π 2 . . . π m π 1 . In such cases, both π and π are assigned half of the value of the right-hand side, in order to avoid double counting.
The denominator on the right refers to a pattern π 2 . . . π m which has a representing permutation κ = κ 1 . . . κ m 1 in S m 1 and can be extended in m ways to a permutation η = η 1 . . . η m S m . Indeed, η m can be chosen from { 1 , . . . , m } . For j < m then either κ j < η m and η j = κ j , or κ j η m and η j = κ j + 1 . Let us write
P m ( κ ) = P m { η | η 1 . . . η m 1   shows   pattern   κ } η P m ( η ) .
In the numerator, π 1 . . . π m is also a pattern which has a representing permutation λ = λ 1 . . . λ m in S m . This term is P m ( λ ) .
We now prove that the defined P m + 1 fulfils condition (ii) of Theorem 2. We calculate
p = P m + 1 { π 1 . . . π m + 1 | π 1 . . . π m is   represented   by   λ }
using the definition (11) of P m + 1 . There are m + 1 permutations π S m + 1 which fulfil the condition, differing in the value π m + 1 . For each case π 2 . . . π m + 1 is represented by one of the permutations η introduced above. However, the two cases π with | π 1 π m + 1 | = 1 belong to the same η . That is why we consider them together and divide their probability by two. Now
p = P m ( λ ) P m ( κ ) · η P m ( η ) = P m ( λ ) .
This proves that P m + 1 extends P m . For the stationarity, the same proof has to be performed with extension to the left and condition (iii). We have to care that now P m ( κ ) refers to η 2 . . . η m . However, since we assumed that P m is stationary, this is the same number, and the proof runs as above. □
We called this a Markov extension since (11) says that π m + 1 does not depend on π 1 . Since we do not assume any independence properties of P m , we cannot expect more. However, if we start with P 2 ( 12 ) = P 2 ( 21 ) = 1 2 and extend successively to m = 3 , 4 , . . . we obtain the coin tossing order.
There are many other extensions. Just to give an example, we can divide the double cases in an asymmetric way. A careful study of extensions may lead to a better model than coin tossing. But we have to stop here.

4. Conclusions

Established models of stochastic processes do not say much about probabilities of ordinal patterns. It is suggested that models for ordinal pattern analysis can be found by algorithms for comparison and ranking of objects rather than by arithmetical operations. A paradigmatic example of an ordinal process without numerical values shows that this is possible. Properties like stationarity and self-similarity can be formulated in a weak and very natural way for ordinal processes. As a starting point for further work, we proved a representation theorem and an extension theorem for stationary orders.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Bandt, C. Statistics and contrasts of order patterns in univariate time series. Chaos 2023, 33, 033124. [Google Scholar] [CrossRef]
  2. Keller, K.; Sinn, M. Ordinal analysis of time series. Physica A 2005, 356, 114–120. [Google Scholar] [CrossRef]
  3. Keller, K.; Sinn, M.; Edmonds, J. Time Series from the ordinal viewpoint. Stochastics and Dynamics 2007, 7, 247–272. [Google Scholar] [CrossRef]
  4. Amigo, J.M. Permutation complexity in dynamical systems; Springer Seires in Synergetics, Springer: Berlin, 2010. [Google Scholar]
  5. Amigo, J.; Keller, K.; Kurths, J. Recent progress in symbolic dynamics and permutation complexity. Ten years of permutation entropy. Eur. Phys. J. Special Topics 2013, 222, 247–257. [Google Scholar]
  6. Zanin, M.; Zunino, L.; Rosso, O.; Papo, D. Permutation entropy and its main biomedical and econophysics applications: a review. Entropy 2012, 14, 1553–1577. [Google Scholar] [CrossRef]
  7. Levya, I.; Martinez, J.; Masoller, C.; Rosso, O.; Zanin, M. 20 years of ordinal patterns: perspectives and challenges. Europhysics Letters 2022, 138, 31001, arXiv2204.12883. [Google Scholar] [CrossRef]
  8. Grimmett, G. ; Stirzaker., D. Probability and random processes, 1995. [Google Scholar]
  9. Sinn, M.; Keller, K. Estimation of ordinal pattern probabilities in Gaussian processes with stationary increments. Computational Statistics and Data Analysis 2011, 55, 1781–1790. [Google Scholar] [CrossRef]
  10. Schnurr, A.; Dehling, H. Testing for structural breaks via ordinal pattern dependence. J. Amer. Stat. Assoc. 2017, 112, 706–720. [Google Scholar] [CrossRef]
  11. Betken, A.; Buchsteiner, J.; Dehling, H.; Münker, I.; Schnurr, A.; Woerner, J.H. Ordinal patterns in long-range dependent time series. Scandinavian Journal of Statistics 2021, 48, 969–1000. [Google Scholar] [CrossRef]
  12. de Sousa, A.R.; Hlinka, J. Assessing serial dependence in ordinal patterns processes using chi-squared tests with application to EEG data analysis. Chaos 2022, 32, 073126. [Google Scholar] [CrossRef]
  13. Weiß, C. Non-parametric tests for serial dependence in time series based on asymptotic implementations of ordinal-pattern statistics. Chaos 2022, 32, 093107. [Google Scholar] [CrossRef]
  14. Nüßgen, I.; Schnurr, A. Ordinal pattern dependence in the context of long-range dependence. Entropy 2021, 23, 670. [Google Scholar] [CrossRef]
  15. Betken, A.; Wendler, M. Rank-based change-point analysis for long-range dependent time series. Bernoulli 2022, 28, 2209–2233. [Google Scholar] [CrossRef]
  16. Olivares, F. Ordinal language of antipersistent binary walks. Physics Letters A 2024, 527, 130017. [Google Scholar] [CrossRef]
  17. Bandt, C.; Shiha, F. Order patterns in time series. J. Time Series Analysis 2007, 28, 646–665. [Google Scholar] [CrossRef]
  18. Elsinger, H. Independence tests based on symbolic dynamics. Working paper No. 150, Österreichische Nationalbank.
  19. DeFord, D.; Moore, K. Random Walk Null Models for Time Series Data. Entropy 2017, 19, 615. [Google Scholar] [CrossRef]
  20. Weiß, C.; Martin, M.R.; Keller, K.; Matilla-Garcia, M. Non-parametric analysis of serial dependence in time series using ordinal patterns. Computational Statistics and Data Analysis 2022, 168, 107381. [Google Scholar] [CrossRef]
  21. Weiß, C.; Schnurr, A. Generalized ordinal patterns in discrete-valued time series: nonparametric testing for serial dependence. Journal of Nonparametric Statistics 2024, 36, 573–599. [Google Scholar] [CrossRef]
  22. Bandt, C. Order patterns, their variation and change points in financial time series and Brownian motion. Statistical Papers 2020, 61, 1565–1588. [Google Scholar] [CrossRef]
  23. Embrechts, P.; Maejima, M. Selfsimilar Processes; Princeton University Press: Princeton, NJ, 2002. [Google Scholar]
  24. Bandt, C.; Pompe, B. Permutation entropy: a natural complexity measure for time series. Phys. Rev. Lett. 2001, 88, 174102. [Google Scholar] [CrossRef]
  25. Keller, G. Equilibrium states in ergodic theory; Vol. 42, London Math. Soc. Student Texts, Cambridge University Press, 1998.
  26. Pitman, J.; Tang, W. Regenerative random permutations of integers. Annals Prob. 2019, 47, 1378–1416. [Google Scholar] [CrossRef]
Figure 1. Histograms of pattern frequencies of a simulated time series for d = 1 and m = 2 , 3 , 4 . Permutations are arranged in hierarchical order. Histograms become refined with increasing m .
Figure 1. Histograms of pattern frequencies of a simulated time series for d = 1 and m = 2 , 3 , 4 . Permutations are arranged in hierarchical order. Histograms become refined with increasing m .
Preprints 158559 g001
Figure 2. Global and local rank numbers obtained from coin tossing.
Figure 2. Global and local rank numbers obtained from coin tossing.
Preprints 158559 g002
Figure 4. Probabilities of patterns of length 4 for coin tossing order with d = 2 and d = 3 . They are clearly different from d = 1 while the changes from d = 2 to d = 3 are small.
Figure 4. Probabilities of patterns of length 4 for coin tossing order with d = 2 and d = 3 . They are clearly different from d = 1 while the changes from d = 2 to d = 3 are small.
Preprints 158559 g004
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated