Skip to main content
Entropy logoLink to Entropy
. 2025 Jun 7;27(6):610. doi: 10.3390/e27060610

Ordinal Random Processes

Christoph Bandt 1
Editor: Luciano Zunino1
PMCID: PMC12192520  PMID: 40566197

Abstract

Ordinal patterns have proven to be a valuable tool in many fields. Here, we address the need for theoretical models. A paradigmatic example shows that a model for frequencies of ordinal patterns can be determined without any numerical values. We specify the important concept of stationary order and the fundamental problems to be solved in order to establish a genuine statistical methodology for ordinal time series.

Keywords: ordinal pattern, stochastic process, time series, permutation entropy

MSC: 37A35, 62M10, 60G07

1. Introduction

1.1. The Main Question

Time series analysis has two parts. First, there are methods for describing data series, like means, correlations, and Fourier spectra. Second, there are models, like Brownian motion or ARMA models, which describe mechanisms that could generate similar sequences of data points. Most models postulate that the next state Xt+1 is a function f(Xt,Xt1,) of previous states, with some random influence. A fundamental example is a simple random walk, which states that Xt+1 is either Xt+1 or Xt1, depending on the result of a coin toss.

In recent years, the study of ordinal patterns in time series has proven successful for many practical applications. However, there is a lack of appropriate models. Established numerical models provide little information on the ordinal structure. This paper proposes directly constructing proper ordinal models. We raise the following question:

Can we define mechanisms that explain the order relations between data points in a time series without using any numerical values?

We shall give an affirmative answer, providing an example of a coin-tossing order. It is similar to a simple random walk. Coin throws are used to define the order relation between all data points while the values, Xt, remain undefined. This is a completely new type of model. Due to its simplicity, it can be easily understood. To model real-world data, the example is too simple. It must be modified and extended. Toward this end, we provide a mathematical framework in the last section.

1.2. Contents of the Paper

In this introductory section, we briefly introduce the basic concepts. For more details and background, we refer the reader to [1] [Section II] and [2,3,4,5,6,7]. One new point involves the visualization of the distribution of ordinal patterns by refining ordinal histograms in Section 1.5. The main topic of this paper is explained in Section 1.7. Order-related properties of processes, particularly order self-similarity, are introduced in Section 1.8.

Section 2 deals with a paradigmatic example of an ordinal process that has no numerical values. The construction is algorithmic and intuitive, and the pattern probabilities for length 3 match those of Brownian motion. Unfortunately, the example is not self-similar and is not relevant to applications. We conjecture that there is a self-similar modification.

Section 3 provides a more technical outline of the possible theory of ordinal processes. We construct stationary ordinal processes without using numerical values. Order self-similarity remains an open problem.

1.3. Permutations as Patterns in Time Series

A permutation π of length m is a one-to-one mapping from the set {1,2,,m} onto itself. We write πk for π(k) and π=π1π2πm. Thus, π=231 means π(1)=2,π(2)=3,π(3)=1. However, we do not need mapping properties of permutations, like composition. We just consider the graph of π as a geometric pattern (an ordinal pattern).

With x=x1,x2,,xT denote a time series of length T. This can also be viewed as a mapping that assigns to the time points t=1,,T the values xt=x(t). For m time points t1<t2<<tm between 1 and T, we consider the pattern of corresponding values. We define

(xt1,xt2,,xtm) shows pattern π when xti<xtj if and only if πi<πj. (1)

In other words, the correspondence xtjπj,j=1,,m is strictly monotonous. This is very intuitive. The pattern π can be determined from the time series by calculating ranks; πk is the number of time points tj for which xjxk.

1.4. Pattern Frequencies in Time Series

Instead of arbitrary tj, we study equally spaced time points t<t+d<t+2d<<t+(m1)d. We call m the length and d the delay or lag of the pattern. We always take m6 and mainly consider m=3 and m=4. However, we vary d as much as possible.

Ordinal pattern analysis is a statistical method. We fix a length, m, and determine relative frequencies of all patterns, π, of length m, for various delays, d. We divide the number of occurrences of π by the number of time points where π could occur.

pπ(d)=1T(m1)d#{t|1tT(m1)d,(xt,xt+d,,xt+(m1)d) shows pattern π}. (2)

These frequencies are the basic parameters of ordinal time series analysis. They can be combined to calculate permutation entropy, persistence, and other important parameters [1,2,3,4,5,6,7].

1.5. Visualization of Pattern Frequencies by Ordinal Histograms

This paper is not about parameters like entropy. It studies models for all pattern frequencies together. For this reason, we introduce a graphical method that summarizes the pattern frequencies for a fixed d in a kind of density function. The permutations π are assigned subintervals Iπ of [0,1] as follows: I12=[0,12],I21=[12,1]. We draw bars over Iπ, each with area pπ(d). For m=3, we note that p12=p123+p132+p231 (neglecting a possible tiny error due to the last value for p12(d) in (2)). We choose I123=[0,16], I132=[16,13],I231=[13,12] so that their union is I12. Then we draw bars with area pπ(d). The same is done for π=213,312,321. In this way, the histogram for m=3 is a refinement of the histogram for m=2, in the sense that the area above I12=[0,12] remains the same. In theory, successive refinements can be obtained for m=4,5, but in practice we rarely have reliable estimates for m>4. This arrangement of permutations is called the hierarchical order. Details will be given in Section 3.4. As Figure 1 shows—for a simulated time series—all pattern frequencies of a time series for a fixed d can be subsumed in this way by a sequence of refining histograms, and in the limit m by a density function.

Figure 1.

Figure 1

Histograms of pattern frequencies of a simulated time series for d=1 and m=2 (red), 3 (blue), and 4 (brown). Each pattern π of length m is represented by a subinterval of [0,1] with length 1/m!. The corresponding bar has area pπ(d), so the height is m!pπ(d). The permutations are arranged in hierarchical order so that each histogram is a refinement of the previous one. For details, see Section 3.4.

The process chosen for Figure 1 is an AR(1) model with exponential noise: Xt+1=12XtlogUt, where the Ut are independent uniformly distributed random numbers in [0,1]. Due to the asymmetry of the noise distribution, p21(1)0.58>12. Successive decreasing steps are frequent: p321(1)0.3 and p4321(1)0.13. We cannot explain why 2143 is the permutation with the second-largest frequency. However, an AR(1) model with the coefficient 12 has a short memory. For d=3, the patterns already have almost uniform frequencies. The models considered below are more uniform for m=2 and 3, but they also provide interesting behavior for large d.

The objective of this paper is to find typical ordinal histograms from a theoretical viewpoint. An important objective is to find models where the histograms agree for all d. Two such models are known [1] [Section II.C], that is, white noise, where the histogram is a constant function for all m, and Brownian motion; see Figure 3. We are looking for other models.

1.6. Stationary and Order Stationary Processes

When modeling time series, we suppose that the given time series is an instance of a larger ensemble, called a stochastic process. The process represents the mechanism that generates the given time series and many other ones. We want to know the laws of this process. Formally, a stochastic process, X, is a sequence, X=X1,X2,, of real random variables on a probability space (Ω,P). The time series, x, is a random choice from the ensemble X.

To find the laws of X from a single time series, we must assume that the distributions and dependencies of Xk do not change in time. Usually, X is required to be stationary. That is, the m-dimensional distribution of (Xt,Xt+1,,Xt+m) does not depend on t, for all m. This assumption is very strong. When we study autocorrelation, we assume weak stationarity. That is, the mean M(Xt) and the covariance Cov(Xt,Xt+k) do not depend on t [8]. When we consider ordinal patterns, we assume a similar, very weak stationarity assumption. It is required throughout this paper.

Order stationarity. For a given pattern π with length m and delay d, let

Pπ(d)=P{(Xt,Xt+d,,Xt+(m1)d) shows pattern π} (3)

denote the probability that the process, X, shows pattern π just after time t. We say that the process, X, is order stationary for patterns of length m if Pπ(d) is the same value for all t. This should hold for all patterns, π, of length m and all delays, d. For a time series of length T, we consider only d=1,,dmax, where dmax must be smaller than T/m.

Brownian motion (where X0=0 and Xt+1Xt are independent increments with standard normal distribution for t=0,1,) is an order stationary process that is not even weakly stationary. It is a basic model for the financial time series. Note that the process, X, and the pattern probabilities, Pπ(d), are theoretical objects. The observed frequency pπ(d) serves as an estimate of Pπ(d), and from a theoretical perspective, the estimator Formula (2) has nice properties [9,10,11,12,13]. The practical success of ordinal patterns indicates that the assumption of order-stationarity is at least approximately fulfilled in applications.

1.7. The Topic of This Paper

There are plenty of well-studied stochastic processes, including Gaussian processes, random walks, ARMA and GARCH processes, processes with long-range dependence [11,14], etc. Most of them are either stationary or have stationary (multi-dimensional) increments, and are, therefore, order stationary. However, they are defined by arithmetical operations. This makes the rigorous study of ordinal pattern frequencies difficult, although various interesting theorems could be derived [9,10,11,15,16]. Even for the Brownian motion, exact frequencies of patterns can be determined only for m4 [17]. Some of them are rational, and some are irrational. Frequencies for length 5 lead to integrals that can only be numerically calculated. Overall, the impression is that ordinal patterns do not fit into the established theory of stochastic processes.

The only proper theoretical model for pattern frequencies is white noise, which is a process consisting of independent and identically distributed random variables, Xi. For each m, all permutations have the same probability, pπ=1/m!, for any delay, d. This model was repeatedly used as a null hypothesis for testing the serial dependence of time series [1,12,13,18,19,20,21]. A greater variety of models is needed.

In this paper, we attempt to define ordinal random processes in a combinatorial way, directly by their ordinal pattern frequencies. We do not require normal distribution, we require no arithmetic, and we do not even need numerical values to compare objects Xi and Xj. An algorithmic example in the next section will show that this is possible. In Section 3, we outline some technical concepts and basic statements for a theory of ordinal processes. This could be the first step toward finding more applicable ordinal models. Note that the possible equality of values, an important problem in practice [21], plays no role in this paper; our models will only allow Xs<Xt or Xs>Xt but not Xs=Xt.

1.8. Symmetry and Independence Properties of Ordinal Processes

Of course, it is not enough to just postulate pattern frequencies. Our models should also have nice properties. Here, we list the ordinal version of some of the most common process properties. We start with symmetry in time and space [9].

A process, X, is reversible if the distributions of (X1,,Xm) and (Xm,,X1) agree for every m. In the ordinal version, it is reversible if the permutations π=π1π2πm and πmπm1π1 have the same pattern probability for every π. The process, X, is invariant under the reversal of values if the distributions of (X1,,Xm) and (X1,Xm) agree for every m. The ordinal version states that the permutations π=π1π2πm and (m+1π1)(m+1π2)(m+1πm) have the same pattern probability for every π.

In both cases, the ordinal versions of symmetry are much weaker and easier to check. These symmetry properties are often met in applications. As a consequence, the distribution of ordinal patterns of length 3 often has only one degree of freedom, since p123=p321, and the probabilities of the other four permutations agree. All Gaussian processes are symmetric in space and time [9,11,17].

A process, X, is said to be Markov if for any t and fixed Xt, the distributions of (X1,,Xt1) and (Xt+1,Xt+2,) are independent. In the ordinal version, this means that the appearances of patterns π,π at times t1<<tmt and tt1<<tm, respectively, are independent. Again, the ordinal property is weaker and easier to check. The Markov property appears in theory more than in applications.

Order self-similarity. The most important symmetry property in this paper is order self-similarity [22] [Section 6]. A random process in continuous time, Yt,t>0, is said to be self-similar if there is an exponent, H>0, such that Yrt=rHYt in distribution, for each positive number, r. Such a process cannot be stationary, it must look similar to Brownian motion, and it has a very special mathematical structure [23]. The ordinal concept for discrete time is much weaker, more general, and very natural. It does not depend on a parameter, and it can be statistically checked for real-world data.

A process, X, is said to be order self-similar for patterns π of length m if

Pπ(d) is the same value for all d. (4)

This should hold for all patterns, π, of length m and all delays, d.

For a time series of length T, we consider d=1,,dmax, where dmax is much smaller than T. With such a restriction, order self-similarity was statistically verified in financial time series [22] and EEG measurements [1]. For self-similar Gaussian processes with stationary increments, Sinn and Keller [9] estimated the Hurst exponent, H, from pattern frequencies. For real-world data, however, it is not at all clear to what extent the strong self-similarity is fulfilled and the exponent H is meaningful. Thus, for practical purposes, order self-similarity seems to be the right concept to start with. The present paper is motivated by the desire for order self-similar models.

2. The Coin-Tossing Order—An Ordinal Process Without Values

2.1. Ranking Fighters by Their Strength

Ordinal patterns were derived from numerical values of a time series. Now, we go the opposite way, similar to the world of tennis or chess, where players are first compared pairwise and then assigned a rank or ELO score. Suppose there are four players, A,B,C,D, in a tennis club. We design an algorithm to order them by strength. First, B plays against A. Let us assume B wins. Then we write A<B. Now the next player, C, plays against the previous player, B. If C wins, then A<B<C by transitivity of order. If C loses, however, C must still play against A. Depending on the result, either A<C<B or C<A<B. Now, the last player, D, comes and plays first against the previous player, C, and then against B and/or A, if the comparison is not yet determined by previous games and transitivity.

This procedure is turned into a mathematical model. We want to construct a random order between objects X1,X2,, which are not numbers. We follow the algorithm for tennis players and replace each match with the toss of a fair coin. As a result, we obtain a stochastic process like white noise or Brownian motion on the ordinal level only. It is possible to determine pattern probabilities much better than for the Brownian motion.

2.2. Definition of the Coin-Tossing Order

Repeated coin tossing is a standard way to represent randomness. Take a simple random walk, for example, and write Xn for our position at time n=0,1,2, then X0=0, and either Xn+1=Xn+1 or Xn+1=Xn1, depending on the result of a coin throw at time n. In the present case, X1,X2,X3, will denote objects, not necessarily numbers. We define a random order between the Xi. We throw a fair coin, cji, to decide whether Xi<Xj or Xi>Xj, for any pair of integers, i<j. Let us write 1 for ‘head’ and 0 for ‘tail’. Then our basic probability space Ω is the space of all 0-1-sequences, where each coordinate is 0 or 1 with probability 12, independently of all other coordinates. Formally,

Ω={(c21,c32,c31,c43,c42,)|cji{0,1}forj>i1}

where cji=0 means Xi<Xj, and cji=1 means Xi>Xj, for 1i<j. The important point is that cji is disregarded when the order of Xi and Xj is already fixed by previous comparisons and transitivity.

The first coin c21 decides the ordering of X1 and X2. Now suppose X1,,Xj1 are already ordered. Then, Xj is compared to Xj1,Xj2,,X1 by considering the random numbers cji. However, when the comparison is fixed by transitivity from the already defined ordering, then cji is disregarded—that coin need not be thrown.

The resulting random order is called the coin-tossing order. It can be easily simulated. Figure 2 shows the rank numbers of 500 consecutive objects, Xj, in the middle of a simulated series of length T = 10,000. The global rank numbers have strange discontinuities. Local rank numbers, obtained by comparing with the following 20 objects on the left and right sides, show a more familiar picture.

Figure 2.

Figure 2

Global and local rank numbers obtained from coin tossing.

Problem. Is there a stochastic process with stationary increments that generates the coin-tossing order?

2.3. Basic Properties of the Coin-Tossing Order

Despite the erratic view of trajectories in Figure 2, it turns out that the coin-tossing order has the same ordinal pattern frequencies for length 3 as Brownian motion, for arbitrary d. In particular, it is order self-similar for patterns of length 3.

Theorem 1

(Basic properties of the coin-tossing order).

  • (i) 

    The coin-tossing order is stationary and has the ordinal Markov property.

  • (ii) 
    For any permutation, π, of length m, the pattern probability is Pπ(1)=2u with
    u=u(π)=#{(i,j)|1i<jm, if i<k<j then πk is not between πi and πj}. (5)
  • (iii) 

    The pattern probabilities, Pπ(d), are invariant under time reversal and the reversal of values.

  • (iv) 

    For d=1,2,. we have P12(d)=12,P123(d)=P321(d)=14, and Pπ(d)=18 for the other permutations of length 3.

Proof. 

(i): The order of (Xt,Xt+1,,Xt+m1) depends on the random numbers cji,t+m1j>it, in the same way that the order of (X1,X2,,Xm) depends on the cji,mj>i1. Since both collections of random numbers have the same distribution, the pattern probabilities do not depend on t, and the defined random order is stationary. Moreover, the comparisons of Xs with st depend on random numbers cji with j>t while the comparisons of Xs with st depend on cji with jt. Since the different cji are independent, this implies the ordinal Markov property.

(ii): u(π) is the number of coin flips needed to determine π. Given a permutation π, we determine the cji variables, which are needed to define the occurrence of π in (X1,X2,,Xm). Of course, mj>i1, and j run in increasing order. For the fixed j, the number cj,j1 is always used, and the other i variables are considered in decreasing order. Now, consider k with i<k<j. If πi<πk<πj, this is determined by the random numbers cki and cjk, which were drawn before cji. In that case, cji is disregarded since πi<πj follows from the transitivity of the order. A similar argument applies when πi>πk>πj. However, if there is no πk between πi and πj, then cji is needed to determine π. We shall call u(π) the energy of π.

(iii): First, we consider d=1. Given π, we have to show that the time-reversed and spatially reversed permutations have the same energy u(π). For spatial reversal, this directly follows from the definition (5) with ‘between’. For time reversal, we show that Pπ(1) can also be determined backwards by considering cji with decreasing i=m1,m2,,1, and for fixed i, with increasing j=i+1,,m. The point is that when we compare places, i<j, we have already compared both j and i with all k between. So, cij is needed only if no πk is between πi and πk. Otherwise, the order between πi and πj is already fixed, and cij is disregarded. This proves reversibility for d=1.

Now consider d>1 and π of length m, which appears as a pattern of (X1,X1+d,,X1+(m1)d). The probability of this event is Pπ(d), which may be different from Pπ(1). We have to calculate Pπ(d) from all ‘atom permutations’ of length M=1+(m1)d for (X1,X2,,X1+(m1)d) for which the order among the special m places agrees with the order of π. This would be a lot of work. However, since the reversed permutation of π is composed of the reversed ‘atom permutations’, both in space and in time, we shall obtain the same pattern probability.

(iv): The spatial reversal invariance implies P12(d)=P21(d)=12. Then, P123(d)=P321(d)=14 follows from the Markov property. The equality of the four other pattern frequencies is a consequence of the invariance under time and space reversal.    □

2.4. Computer Work and Breakdown of Self-Similarity

In Figure 3, we compare pattern probabilities of length 4 for the Brownian motion, taken from [17], with those of the coin-tossing order. The Brownian motion looks more interesting since coin-tossing allows only the probabilities 2k. Our model is simple, but not recommended for applications. For length 3, the Pπ(d) of the two processes agree and do not depend on d, according to Theorem 1. The figure shows that we must investigate patterns of length 4 if we want to distinguish processes like these two. Moreover, the Brownian motion is perfectly self-similar so that the probabilities, Pπ(d), of any length do not depend on d [23]. Is this also true for the coin-tossing order?

Figure 3.

Figure 3

Probabilities of patterns of length 4 (brown) on top of those of length 3 (blue) for the Brownian motion (left) and coin-tossing order with d=1 (right). For length 3, the probabilities coincide.

Equation (5) allows for accurate calculation of all pattern probabilities for d=1 and m10. Note that such pattern probabilities just for m=5 have never been determined for non-trivial stochastic processes, not even for the Brownian motion or ARMA models of low order [17]. Here is a MATLAB procedure that determines u(x) for a permutation x, given as a vector of length m. The vector w contains the logical values 0 and 1.

function u=ctoscore(x);

m=length(x); u=m-1; % coin throws for i=1,...,m-1 and j=i+1 are necessary

for i=1:m-2; for j=i+2:m; v=x(i+1:j-1); % look for numbers k between i, j

  if x(i)<x(j); w=v>x(i)&v<x(j); else w=v<x(i)&v>x(j); end % x(k) as in (5)

  if sum(w)==0; u=u+1; end % if such x(k) do not exist, add a coin throw

end; end;                                      % end of loops for i and j

It takes five minutes to run this procedure over all permutations of length m=10. The minimum number u(x) is m1=9 and appears only for x=12m and x=m(m1)1. The maximum number u(x)=33 is realized by eight permutations, that is, x=67891012345, and similar patterns. About three-quarters of the 3.6 million patterns fulfill 23u(x)27.

Now, we can determine Pπ(d) for all permutations of length 4 and d=3 by adding probabilities 2u(x) of ‘atom permutations’ x of length 10, as in the proof of Theorem 1, (iii). For π=1324, we consider all x with x(1)<x(7)<x(4)<x(10). We add 10!/4!=151,200 cases. The result must be a rational number with denominator 2k, where k33. For π=1324, we get Pπ(3)=3·41·59·10,321/2313.49%. To determine Pπ(2) for patterns of length 4, it suffices to consider m=7.

The result is shown in Figure 4. Unfortunately, order self-similarity breaks down for patterns of length 4. For π=1324, we have Pπ(1)=1323.12% while Pπ(2)3.42% and Pπ(3)3.49%. However, the probabilities for d=2 or 3 can be taken as a new model, which is nearer to self-similarity and looks not as artificial as coin tossing in Figure 3.

Figure 4.

Figure 4

Probabilities of patterns of length 4 for the coin-tossing order with d=2 and d=3. They are clearly different from d=1 while the changes from d=2 to d=3 are small.

Problem. Let us define new pattern probabilities Qπ(1)=Pπ(d) for d=2,3, what are the properties of Q? Is there a limit for d ?

The small difference between d=2 and d=3 in Figure 4 indicates fast convergence. The mere existence of an order self-similar limit would be interesting, but not yet helpful. After all, Brownian motion is already available as an ordinally self-similar process. The merit of the coin-tossing order lies in its algorithmic flavor and its connection to ranking. We look for an ordinally self-similar model with such an intuitive explanation.

Finally, let us consider the interpretation of u(π) as an energy function. Let d=1, and let Sm denote the set of all permutations of length m. Let P denote the probability measure on Sm given by the probabilities, Pπ=2u(π), of the coin-tossing order. The permutation entropy [24] of P, using logarithms with base 2, is just the mean energy with respect to P.

H(P)=πPπlog2Pπ=πPπu(π)=MP(u).

For all other probability measures, Q on Sm, the permutation entropy is smaller than the mean energy; see [25] [Chapter 1]. Thus, P is a so-called Gibbs measure on Sm.

Problem. Are there other meaningful energy functions for permutations, perhaps even parametric families of Gibbs measures on Sm?

3. Rudiments of a Theory of Ordinal Processes

3.1. Random Stationary Order

Having studied one example, we will now discuss a possible theory of ordinal random processes. As above, X1,X2, are not numbers, but only objects that are ordered. We have the discrete-time domain {1,,T} for time series and N={1,2,} for models. An order is a relation < on the time domain such that xx, and xy implies either x<y or y<x, and x<y together with y<z implies x<z. Of course, the order does not apply to the time points t, but to the corresponding objects xt. Then the statement that xt,,xt+m1 exhibits pattern π has a meaning for every πSm—it is either true or false. When an order on N or {1,,T} is given, we can determine pattern frequencies, permutation entropy, and so on. In the following, we always have d=1.

We want to construct models of ordinal processes, like the coin-tossing algorithm. For this purpose, we need the concept of random order. To keep things simple, we identify an ordinal process with its law, as follows: A random order is defined as a probability measure on the set of all orders on the time domain. For the finite time domain {1,,T}, a random order is just a probability measure, PT, on the set, ST, of permutations of length T. For mT, πSm, and 1tT+1m, the random order allows determining the probability:

Pπt=P{xt,,xt+m1 shows pattern π} (6)

The random order is called stationary if the probabilities Pπt do not depend on the time point t, for any pattern π of any length m<T. In other words, the numbers Pπt must be the same for all admissible t. This is exactly the order stationarity that we defined for numerical processes in (3).

3.2. The Problem of Finding Good Models

The infinite time domain N is considered in the next section. We use the classical measure theory, which is easy. The real problem appears already for finite T, even for T=20. We have an abundance of probability measures on ST since we can prescribe Pπ for every π. When we require stationarity, we have (T1)! equations for these T! parameters, as shown below, which still leaves too many choices.

The objective is to select realistic Pπ. Most of the patterns, π, for m=20 will never appear in any real-time series, and we could set Pπ=0, but we do not know for which π. There are three types of properties that we should require for our model.

  • Independence: The Markov property or k-dependence of patterns (cf. [12]).

  • Self-similarity: Pπ(d) should not depend on d. For processes with short-term memory, like Figure 1, this could be replaced by the requirement that Pπ(d) converge to the uniform distribution, exponentially with increasing d. There must be a law connecting Pπ(d) for different d. Otherwise, how can we describe them all?

  • Smoothness: There should not be too much zigzag in the time series. Patterns like 3142 should be exceptions. This can perhaps be reached by minimizing certain energy functions.

Problem. Does there exist on S100 a stationary random order that is Markov and self-similar (for admissible d and π) and has parameters different from white noise and Brownian motion; for instance, P1212, or 13<P123+P321<12?

3.3. Random Order on N

The infinite time domain N has its merits. Order stationarity, for instance, is very easy to define since every pattern can be shifted to the right as far as we want. It is enough to require

Pπt=Pπt+1 for all finite patterns π and t=1,2, (7)

It is even enough to require Pη1=Pη2 for all patterns η of any length. (To prove that this implies (7) for a fixed t and a pattern π of length k, consider all patterns η of length m=t+k1, which show the pattern π on their last k positions. The sum of their probabilities Pη1 equals Pπt, since P is a measure. And shifting all η from 1 to 2 means shifting π from t to t+1.)

On the other hand, infinite patterns require a limit T. Moreover, there are many recent papers on infinite permutations, that is, one-to-one mappings of N onto itself. An overview was given by Pitman and Tang [26]. However, an order on N is a much wider concept than a permutation on N. An infinite permutation π1,π2, defines an order on N, with the special property that below a given value πk, there are only finitely many other values, for any k.

For an order on N, however, the smallest object rarely exists. Usually, each object has infinitely many other objects below and above. Nevertheless, an order on N is uniquely defined by patterns πmSm, which represent the order of the first m elements, for m=2,3, for example, 12231341245132

Theorem 2

(Approximating random order on N).

  • (i) 

    A sequence of permutations πmSm defines an order on N if for m=2,3,, the pattern πm is represented by the first m values of the pattern πm+1.

  • (ii) 
    A sequence of probability measures Pm on Sm defines a random order P on N if for m=2,3,, for πSm and πm+1Sm+1, the following holds:
    Pm+1{πm+1 shows pattern π at its first m positions}=Pm(π).
  • (iii) 
    The random order P defined by the Pm is stationary if and only if, for m=2,3,, for πSm, and πm+1Sm+1, the following holds:
    Pm+1{πm+1 shows pattern π at its last m positions}=Pm(π).

Proof. 

(i): The condition says that the pattern of the first m objects, defined in step m, will not change during successive steps. So the construction is straightforward. The rank numbers, however, may change in each step, and they may converge to for m.

(ii): The condition says that the probability for a pattern πSm to appear for the first m objects, defined by Pm, will remain the same for Pm+1 and successive probability measures. So we can define

P{the objects with numbers 1,,m show pattern π}=Pm(π)

in a consistent way, and Pm determines P. Moreover, P is an inverse limit of the measures Pm. Below, we provide a more elementary argument.

(iii): Together with (ii), this condition says that Pπ1=Pπ2 for all patterns π of any length m. As noted above, this is just the definition (7) of order stationarity. □

3.4. The Space of Random Orders

Resuming the discussion in Section 1.5, we show that the set of all orders on N can be represented by the unit interval, using a numeration system similar to our decimal numbers. We first assign subintervals of I=[0,1] to the permutations of length 2,3, the pattern 12 corresponds to the interval [0,12], and 21 to [12,1]. The permutations 123, 132, and 231, which exhibit the pattern 12 at their first two positions, correspond to [0,16],[16,13], and [13,12], respectively. For intervals of length 4, see Figure 3.

Instead of the lexicographic order, we define a hierarchical order of permutations, with respect to the patterns shown by the first 2,3,…elements. For any permutation π=π1πm of length m and any k with 2km, let rk(π) denote the number of j<k with πj>πk. This is a kind of rank number of πk with values between 0 and k1. For example, r3(123)=0,r3(132)=1, and r3(231)=2 while r2=0 for these three patterns. Now, we assign to the permutation π of length m the following interval:

I(π)=[x(π),x(π)+1m!]withx(π)=k=2mrkk!. (8)

It is easy to check that the patterns π(k)=π1πk of the first k<m items of π are assigned to larger intervals, i.e.,

II(π(2))I(π(3))I(π(m))

where π(m)=π. Moreover, π need not be a permutation, just a pattern—it could also be a numerical time series. Only the ordering of πj is used for defining rk and I(π).

When we extend the pattern to the right, we obtain smaller nested subintervals, and for m a single point x=k=2rkk!, which characterizes the limiting order of infinitely many objects. Thus, each order on N corresponds to a unique point x in [0,1]. This is very similar to decimal expansions, where we subdivide an interval into 10 subintervals. In the case of patterns, values rk are the digits, and we subdivide first into 2, then into 3, 4, 5,…intervals. The endpoints of intervals represent two orders on N, but this is an exception, as 0.5 and 0.4999 for the decimals.

Once we represent all orders on N as points in [0,1], we can better understand the probability measures P2,P3, of Theorem 2 and the limiting probability measure P, which is called the random order on N. We start with the function F1(x)=1, which denotes uniform distribution on [0,1]. The function Fm represents the measure Pm, for m=2,3, for patterns π of length m, it is defined as the histogram of Pm:

Fm(x)=m!·Pm(π) for xI(π). (9)

See Figure 3. The rectangle over I(π) has area Pm(π). In the case of white noise, Fm=1 for all m, and limFm=F=1 is the uniform distribution. We now show that such a limit exists for all sequences F2,F3, for which P2,P3, fulfill condition (ii) of Theorem 2. We reformulate (ii) as

I(π)Fm+1(x)dx=m!·Pm(π)=I(π)Fm(x)dx. (10)

The second equation is obvious, and the first is best illustrated by an example with m=3 and π=312. We have r2=r3=1, so I(π)=[46,56]. The possible extensions π4 of π are 3124, 4123, 4132, and 4231. Their intervals of length 124 partition I(π). Condition (ii) states that P4{3124,4123,4132,4231}=P3{312}. Thus, the four rectangles of F4 over I(π) together have the same area as the one rectangle of F3. This is expressed in (10).

Equation (10) states that the Fm form a martingale. The martingale convergence theorem implies that there is a limit function F, in the sense that 01|FmF|dx converges to zero for m. This limit function is integrable and F=1. As a density function, it defines the probability measure P on all orders on N.

Our argument indicates that random orders on N belong to the realm of classical analysis and probability. Of course, the density function F is terribly discontinuous and can hardly be used to discuss stationarity. (An open challenge for experts is to identify examples for which the sequence Fm converges in L2.)

3.5. Extension of Pattern Distributions

We conclude our paper with an optimistic outlook for practitioners. When you find a distribution of pattern probabilities of length 3 or 4, say, from data, say, there is no need to worry about extending it to longer finite or infinite patterns. Such an extension will always exist.

Theorem 3

(Markov extension of pattern probabilities). Any stationary probability measure Pm on Sm can be extended to a stationary probability measure Pm+1 on Sm+1, and, hence, also to a stationary probability measure P on the space of random orders.

Proof. 

To show that Pm+1 is stationary, we only need to verify condition (iii) of Theorem 2; any pattern within {1,,m} can be shifted to the right, without changing its probability, until its maximum reaches m, by the stationarity assumption of Pm. What remains is to shift the maximum from m to m+1, and this is condition (iii).

In the following extension formula, we use the convention that probability measures on permutations also apply to patterns, by replacing a pattern with its representing permutation. Let π=π1π2πmπm+1 be a permutation in Sm+1.

Pm+1(π)=Pm(π1πm)·Pm(π2πm+1)Pm(π2πm) (11)

This formula is used whenever there exists some πk with 2km between π1 and πm+1. However, if π1 and πm+1 are neighboring numbers, then the right-hand side of (11) cannot distinguish π and π=πm+1π2πmπ1. In such cases, both π and π are assigned half of the value of the right-hand side, in order to avoid double counting.

The denominator on the right refers to a pattern π2πm that has a representing permutation κ=κ1κm1 in Sm1 and can be extended in m ways to a permutation η=η1ηmSm. Indeed, ηm can be chosen from {1,,m}. For j<m, either κj<ηm and ηj=κj, or κjηm and ηj=κj+1. Let us write

Pm(κ)=Pm{η|η1ηm1 shows pattern κ}ηPm(η).

In the numerator, π1πm is also a pattern that has a representing permutation λ=λ1λm in Sm. This term is Pm(λ).

We now prove that the defined Pm+1 fulfills condition (ii) of Theorem 2. We calculate

p=Pm+1{π1πm+1|π1πm is represented by λ}

using the definition (11) of Pm+1. There are m+1 permutations πSm+1 that fulfill the condition, differing in the value πm+1. For each case, π2πm+1 is represented by one of the permutations η introduced above. However, the two cases π with |π1πm+1|=1 belong to the same η. This is why we consider them together and divide their probability by two. Now,

p=Pm(λ)Pm(κ)·ηPm(η)=Pm(λ).

This proves that Pm+1 extends Pm. For the stationarity, the same proof has to be performed with the extension to the left and condition (iii). We must be careful that Pm(κ) now refers to η2ηm. However, since we assume that Pm is stationary, this is the same number, and the proof runs as above. □

We call this a Markov extension since (11) states that πm+1 does not depend on π1. Since we do not assume any independence properties of Pm, we cannot expect more. However, if we start with P2(12)=P2(21)=12 and extend successively to m=3,4, we obtain the coin-tossing order.

There are many other extensions. To provide an example, we can divide the double cases in an asymmetric way. A careful study of extensions may lead to a better model than coin tossing. But we have to stop here.

4. Conclusions and Outlook

Established models of stochastic processes do not say much about the probabilities of ordinal patterns. It has been suggested that models for ordinal pattern analysis can be found by algorithms for comparison and ranking of objects rather than by arithmetical operations. A paradigmatic example of an ordinal process without numerical values shows that this is possible. Properties like stationarity and self-similarity can be formulated in a weak and very natural way for ordinal processes. As a starting point for further work, we have proved a representation theorem and an extension theorem for stationary orders.

Our example can be improved and generalized in different ways. One way is to define ordinal processes directly from the extension theorem. Another way is to modify the algorithm so that order depends not only on coin tossing but also on already existing comparisons. Furthermore, we can consider the Pπ(d) of the coin-tossing order as d and prove the conjecture that this leads to an order self-similar process, similar to how the simple random walk leads to Brownian motion. On the practical side, we have to study large datasets and estimate the Pπ(d) for m=4 and various values of d as accurately as possible, in order to classify different behaviors. The fundamental mathematical challenge is to understand connections between pattern probabilities for different scales, d.

Data Availability Statement

Data are contained within the article. Further inquiries can be directed to the author.

Conflicts of Interest

The author declares no conflicts of interest.

Funding Statement

This research received no external funding.

Footnotes

Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

References

  • 1.Bandt C. Statistics and contrasts of order patterns in univariate time series. Chaos. 2023;33:033124. doi: 10.1063/5.0132602. [DOI] [PubMed] [Google Scholar]
  • 2.Keller K., Sinn M. Ordinal analysis of time series. Phys. A. 2005;356:114–120. doi: 10.1016/j.physa.2005.05.022. [DOI] [Google Scholar]
  • 3.Keller K., Sinn M., Edmonds J. Time Series from the ordinal viewpoint. Stochastics Dyn. 2007;7:247–272. doi: 10.1142/S0219493707002025. [DOI] [Google Scholar]
  • 4.Amigo J.M. Permutation Complexity in Dynamical Systems. Springer; Berlin/Heidelberg, Germany: 2010. (Springer Series in Synergetics). [Google Scholar]
  • 5.Amigo J., Keller K., Kurths J. Recent progress in symbolic dynamics and permutation complexity. Ten years of permutation entropy. Eur. Phys. J. Spec. Top. 2013;222:247–257. [Google Scholar]
  • 6.Zanin M., Zunino L., Rosso O., Papo D. Permutation entropy and its main biomedical and econophysics applications: A review. Entropy. 2012;14:1553–1577. doi: 10.3390/e14081553. [DOI] [Google Scholar]
  • 7.Levya I., Martinez J., Masoller C., Rosso O., Zanin M. 20 years of ordinal patterns: Perspectives and challenges. arXiv. 2022 doi: 10.1209/0295-5075/ac6a72.2204.12883 [DOI] [Google Scholar]
  • 8.Grimmett G., Stirzaker D. Probability and Random Processes. Oxford University Press; Oxford, UK: 1995. [Google Scholar]
  • 9.Sinn M., Keller K. Estimation of ordinal pattern probabilities in Gaussian processes with stationary increments. Comput. Stat. Data Anal. 2011;55:1781–1790. doi: 10.1016/j.csda.2010.11.009. [DOI] [Google Scholar]
  • 10.Schnurr A., Dehling H. Testing for structural breaks via ordinal pattern dependence. J. Am. Stat. Assoc. 2017;112:706–720. doi: 10.1080/01621459.2016.1164706. [DOI] [Google Scholar]
  • 11.Betken A., Buchsteiner J., Dehling H., Münker I., Schnurr A., Woerner J.H. Ordinal patterns in long-range dependent time series. Scand. J. Stat. 2021;48:969–1000. doi: 10.1111/sjos.12478. [DOI] [Google Scholar]
  • 12.de Sousa A.R., Hlinka J. Assessing serial dependence in ordinal patterns processes using chi-squared tests with application to EEG data analysis. Chaos. 2022;32:073126. doi: 10.1063/5.0096954. [DOI] [PubMed] [Google Scholar]
  • 13.Weiß C. Non-parametric tests for serial dependence in time series based on asymptotic implementations of ordinal-pattern statistics. Chaos. 2022;32:093107. doi: 10.1063/5.0094943. [DOI] [PubMed] [Google Scholar]
  • 14.Nüßgen I., Schnurr A. Ordinal pattern dependence in the context of long-range dependence. Entropy. 2021;23:670. doi: 10.3390/e23060670. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Betken A., Wendler M. Rank-based change-point analysis for long-range dependent time series. Bernoulli. 2022;28:2209–2233. doi: 10.3150/21-BEJ1416. [DOI] [Google Scholar]
  • 16.Olivares F. Ordinal language of antipersistent binary walks. Phys. Lett. A. 2024;527:130017. doi: 10.1016/j.physleta.2024.130017. [DOI] [Google Scholar]
  • 17.Bandt C., Shiha F. Order patterns in time series. J. Time Ser. Anal. 2007;28:646–665. doi: 10.1111/j.1467-9892.2007.00528.x. [DOI] [Google Scholar]
  • 18.Elsinger H. Independence Tests Based on Symbolic Dynamics. Österreichische Nationalbank; Wien, Austria: 2010. Working Paper No. 150. [Google Scholar]
  • 19.DeFord D., Moore K. Random Walk Null Models for Time Series Data. Entropy. 2017;19:615. doi: 10.3390/e19110615. [DOI] [Google Scholar]
  • 20.Weiß C., Martin M.R., Keller K., Matilla-Garcia M. Non-parametric analysis of serial dependence in time series using ordinal patterns. Comput. Stat. Data Anal. 2022;168:107381. doi: 10.1016/j.csda.2021.107381. [DOI] [Google Scholar]
  • 21.Weiß C., Schnurr A. Generalized ordinal patterns in discrete-valued time series: Nonparametric testing for serial dependence. J. Nonparametric Stat. 2024;36:573–599. doi: 10.1080/10485252.2023.2231565. [DOI] [Google Scholar]
  • 22.Bandt C. Order patterns, their variation and change points in financial time series and Brownian motion. Stat. Pap. 2020;61:1565–1588. doi: 10.1007/s00362-020-01171-7. [DOI] [Google Scholar]
  • 23.Embrechts P., Maejima M. Selfsimilar Processes. Princeton University Press; Princeton, NJ, USA: 2002. [Google Scholar]
  • 24.Bandt C., Pompe B. Permutation entropy: A natural complexity measure for time series. Phys. Rev. Lett. 2001;88:174102. doi: 10.1103/PhysRevLett.88.174102. [DOI] [PubMed] [Google Scholar]
  • 25.Keller G. Equilibrium States in Ergodic Theory. Volume 42 Cambridge University Press; Cambridge, UK: 1998. London Mathematical Society Student Texts. [Google Scholar]
  • 26.Pitman J., Tang W. Regenerative random permutations of integers. Ann. Prob. 2019;47:1378–1416. doi: 10.1214/18-AOP1286. [DOI] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Data Availability Statement

Data are contained within the article. Further inquiries can be directed to the author.


Articles from Entropy are provided here courtesy of Multidisciplinary Digital Publishing Institute (MDPI)

RESOURCES