Hostname: page-component-745bb68f8f-b6zl4 Total loading time: 0 Render date: 2025-02-11T12:27:23.556Z Has data issue: false hasContentIssue false

Classification of backward filtrations and factor filtrations: examples from cellular automata

Published online by Cambridge University Press:  21 July 2021

PAUL LANTHIER
Affiliation:
Laboratoire de Mathématiques Raphaël Salem, Université de Rouen, CNRS, Avenue de l’Université, F-76801 Saint Étienne du Rouvray, France (e-mail: planthier76@outlook.fr)
THIERRY DE LA RUE*
Affiliation:
Laboratoire de Mathématiques Raphaël Salem, Université de Rouen, CNRS, Avenue de l’Université, F-76801 Saint Étienne du Rouvray, France (e-mail: planthier76@outlook.fr)
Rights & Permissions [Opens in a new window]

Abstract

We consider backward filtrations generated by processes coming from deterministic and probabilistic cellular automata. We prove that these filtrations are standard in the classical sense of Vershik’s theory, but we also study them from another point of view that takes into account the measure-preserving action of the shift map, for which each sigma-algebra in the filtrations is invariant. This initiates what we call the dynamical classification of factor filtrations, and the examples we study show that this classification leads to different results.

Type
Original Article
Copyright
© The Author(s), 2021. Published by Cambridge University Press

1 Introduction

The topic of this paper is the study of some backward filtrations, or filtrations in discrete negative time, which are non-decreasing sequences of the form $\mathscr {F}=(\mathscr {F}_n)_{n\leq 0}$ where each $\mathscr {F}_n$ is a sub-sigma-algebra on a given probability space. In the sequel, we will simply refer to them as filtrations.

In this work, all measure spaces are implicitly assumed to be Polish spaces equipped with their Borel sigma-algebra, and we only consider random variables taking their values in such spaces. If X is a random variable defined on some probability space $(\Omega ,\mathbb {P})$ , taking values in a Polish space E, the law of X is the probability measure $\mathscr {L}(X)$ on E which is the pushforward measure of $\mathbb {P}$ by the measurable map X. We call a copy of X any random variable $X'$ , possibly defined on another probability space, such that $\mathscr {L}(X')=\mathscr {L}(X)$ (in particular, $X'$ takes its values in the same space as X). Given a family $(X_m)_{m\in I}$ of random variables defined on a given probability space $(\Omega ,\mathbb {P})$ , we denote by $\boldsymbol {\Sigma }(X_m: m\in I)$ the sub-sigma algebra generated by the random variables $X_m$ , $m\in I$ and the negligible sets. All sigma-algebras are supposed to be complete and essentially separable: they are generated (modulo negligible sets) by countably many events (or, equivalently, by countably many random variables).

Therefore, each filtration $\mathscr {F}=(\mathscr {F}_n)_{n\leq 0}$ we consider can be generated by a process in negative time (that is to say, a family $X=(X_n)_{n\leq 0}$ of random variables), which means that for each $n\leq 0$ , $\mathscr {F}_n=\boldsymbol {\Sigma }(X_m: m\leq n)$ . The filtration $\mathscr {F}$ is then the mathematical object that describes the acquisition of information as the process $(X_n)_{n\leq 0}$ evolves from $-\infty $ up to the present time. Different processes can generate the same filtration, and the classification of filtrations is roughly equivalent to answering the following question: given a filtration $\mathscr {F}$ , can we find a ‘nice’ process generating it?

In this direction, the simplest structure we can have is called a filtration of product type, which means a filtration which can be generated by a sequence $X=(X_n)_{n\leq 0}$ of independent random variables. Then there is the slightly more general notion of standard filtration: a filtration which is immersible into a filtration of product type (see Definition 2.3 below).

Classification of backward filtrations was initiated by Vershik in the 1970s [Reference Vershik16]. His work, which was written in the language of ergodic theory, did not receive much attention until a later publication in the 1990s [Reference Vershik17]. Then some authors started to bring the subject into probability theory where they found nice applications (see, in particular, [Reference Dubins, Feldman, Smorodinsky and Tsirelson5, Reference Émery and Schachermayer6, Reference Tsirelson15]). In the present paper, among other things we propose a return trip to ergodic theory by considering special families of filtrations on measure-theoretic dynamical systems (such systems are given by the action of a measure-preserving transformation T on a probability space $(\Omega ,\mathscr {A},\mathbb {P})$ ). Filtrations in measure-theoretic dynamical systems have already been considered by many authors (see, for example, [Reference Bressaud, Maass, Martinez and San Martin2, Reference Ceillier and Leuridan3, Reference Hoffman8, Reference Hoffman and Rudolph9]), but in situations where the transformation acts ‘in the direction of the filtration’: by this we mean that, in these works, the filtrations satisfy $\mathscr {F}_{n-1}=T^{-1}\mathscr {F}_n\varsubsetneq \mathscr {F}_n$ for each $n\le 0$ . Here we adopt a transverse point of view: we consider examples where the sigma-algebras forming the filtration are invariant with respect to the underlying transformation: $\mathscr {F}_n=T^{-1}\mathscr {F}_n$ for each $n\le 0$ . We get what we call factor filtrations, and we would like to classify these objects by taking into account the transverse action of the transformation. To emphasize the role of the underlying dynamical system, we will call the classification of factor filtrations the dynamical classification of factor filtrations. In contrast, we will refer to the usual classification of filtrations as the static classification.

The examples that we consider in this paper are all derived from the theory of cellular automata, where the underlying measure-preserving transformation is the shift of coordinates. We provide in this context natural examples of factor filtrations, some of which show a different behaviour depending on whether we look at them from a static or a dynamical point of view.

1.1 Examples of filtrations built from cellular automata

We define in this section the two filtrations which will be studied in the paper. The first is built from an algebraic deterministic cellular automaton $\tau $ , and the second from a probabilistic cellular automaton (PCA) which is a random perturbation of $\tau $ , depending on a parameter $\varepsilon $ and denoted by ${\tau _{\varepsilon }}$ . Both automata are one-dimensional (cells are indexed by $\mathbb {Z}$ ) and the state of each cell is an element of a fixed finite Abelian group $(\mathbb {A},+)$ , which we assume non-trivial: $|\mathbb {A}|\ge 2$ .

Throughout the paper the index n is viewed as the time during which the cellular automata evolve (represented vertically from top to bottom in the figures), whereas the index i should be interpreted as the position within a given configuration (represented horizontally). For a given configuration $x=(x(i):i\in \mathbb {Z})$ and $j\le k\in \mathbb {Z}$ , we denote by $x[j,k]$ the restriction $(x(i):j\le i\le k)$ of x to the sites between j and k.

1.1.1 The deterministic cellular automaton $\tau $ and the filtration ${\mathscr {F}_{}^{\tau ^{-1}}}$

We first define the deterministic cellular automaton $\tau :\mathbb {A}^{\mathbb {Z}}\to \mathbb {A}^{\mathbb {Z}}$ . Given a configuration $x=(x(i))_{i\in \mathbb {Z}}$ , the image configuration $\tau x$ is given by

(1) $$ \begin{align} \text{for all } i\in\mathbb{Z},\quad \tau x(i) := x(i)+x(i+1). \end{align} $$

Observe that $\tau $ can be written as $\tau =\sigma +\operatorname {\mathrm {Id}}$ , where $\sigma $ is the left shift on $\mathbb {A}^{\mathbb {Z}}$ .

For any finite set S, we denote by $\mathcal {U}_S$ the uniform probability measure on S. We consider on $\mathbb {A}^{\mathbb {Z}}$ the product probability measure $\mu :=\mathcal {U}_{\mathbb {A}}^{\otimes \mathbb {Z}}$ , which is the normalized Haar measure on the compact Abelian group $\mathbb {A}^{\mathbb {Z}}$ . The measure $\mu $ is invariant by $\tau $ .

On the probability space $(\mathbb {A}^{\mathbb {Z}},\mu )$ , we construct a negative-time process $(X_n)_{n\le 0}$ as follows: we first consider the random variable $X_0$ taking values in $\mathbb {A}^{\mathbb {Z}}$ and simply defined by the identity on $\mathbb {A}^{\mathbb {Z}}$ . Then we set for each integer $n\le 0$ ,

$$ \begin{align*} X_n := \tau^{|n|} X_0. \end{align*} $$

Since the law $\mu $ of $X_0$ is preserved by $\tau $ , the process $(X_n)_{n\le 0}$ defined in this way is stationary. In particular, all the random variables $X_n$ , $n\le 0$ , have the same law $\mu $ .

We denote by ${\mathscr {F}_{}^{\tau ^{-1}}}$ the filtration generated by this process $(X_n)_{n\le 0}$ : for each $n\le 0$ , ${\mathscr {F}_{n}^{\tau ^{-1}}}=\boldsymbol {\Sigma }(X_m:m\le n)=\boldsymbol {\Sigma }(X_m)$ . We use $\tau ^{-1}$ instead of $\tau $ in this notation to emphasize the fact that the time of the process generating this filtration goes in the other direction than that of the cellular automaton $\tau $ .

1.1.2 The probabilistic cellular automaton ${\tau _{\varepsilon }}$ and the filtration ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$

We also introduce a random perturbation of the deterministic cellular automaton $\tau $ , depending on a parameter $\varepsilon $ which is the probability of making an error in the computation of the new state of a given cell. More precisely, we fix $\varepsilon>0$ and we define ${\tau _{\varepsilon }}$ as the Markov kernel on $\mathbb {A}^{\mathbb {Z}}$ given by the following: for each $x\in \mathbb {A}^{\mathbb {Z}}$ , ${\tau _{\varepsilon }}(x,\cdot )$ is the law of a random variable of the form $\tau x+\xi $ , where $\xi $ is a random error, with law

$$ \begin{align*} \mathscr{L}(\xi) = \bigotimes_{i\in\mathbb{Z}}\bigg((1-\varepsilon)\delta_{0_{\mathbb{A}}}+ \frac{\varepsilon}{|\mathbb{A}| -1}\sum_{a\in\mathbb{A}\setminus\{0_{\mathbb{A}}\}}\delta_a\bigg). \end{align*} $$

In the sequel, we will always assume that

(2) $$ \begin{align} \varepsilon < \frac{[\mathbb{A}|-1}{|\mathbb{A}| }, \end{align} $$

which is equivalent to $1-\varepsilon>\varepsilon /(|\mathbb {A}|-1)$ . In other words, we give more weight to $0_{\mathbb {A}}$ . Setting

(3) $$ \begin{align} \tilde{\varepsilon}:=\varepsilon\dfrac{|\mathbb{A}| }{|\mathbb{A}| -1}<1, \end{align} $$

we can also write the law of the random error $\xi $ as

(4) $$ \begin{align} \mathscr{L}(\xi) = \bigotimes_{i\in\mathbb{Z}}\bigg((1-\tilde{\varepsilon})\delta_{0_{\mathbb{A}}}+ \frac{\tilde{\varepsilon}}{|\mathbb{A}| }\sum_{a\in\mathbb{A}}\delta_a\bigg). \end{align} $$

The following lemma shows that the probability measure $\mu $ on $\mathbb {A}^{\mathbb {Z}}$ defined in §1.1.1 is invariant by this Markov kernel (in fact, as we will prove later, it is the only one; see Corollary 2.17).

Lemma 1.1. Let $X=(X(i):i\in \mathbb {Z})$ and $\xi =(\xi (i):i\in \mathbb {Z})$ be two independent random variables taking values in $\mathbb {A}^{\mathbb {Z}}$ , where $\mathscr {L}(X)=\mu $ . Then $\mathscr {L}(\tau X + \xi )=\mu $ .

Proof We already know that $\mathscr {L}(\tau X)=\mu $ . By independence of $\tau X$ and $\xi $ , and since $\mu $ is the Haar measure on $\mathbb {A}^{\mathbb {Z}}$ , for $a\in \mathbb {A}^{\mathbb {Z}}$ we have

$$ \begin{align*} \mathscr{L}(\tau X+\xi | \xi=a) = \mathscr{L}(\tau X + a) =\mu.\\[-34pt] \end{align*} $$

(Note that the proof of the lemma is valid regardless of the law of the error $\xi $ , and that it also proves the independence of $\tau X+\xi $ and $\xi $ .)

By invariance of $\mu $ under ${\tau _{\varepsilon }}$ , we can construct on some probability space $(\Omega ,\mathbb {P})$ a stationary Markov chain $(X_n)_{n\in \mathbb {Z}}$ , where for each $n\in \mathbb {Z}$ , $X_n$ takes its values in $\mathbb {A}^{\mathbb {Z}}$ , the law of $X_n$ is $\mathscr {L}(X_n)=\mu $ , and for each $x\in \mathbb {A}^{\mathbb {Z}}$ , the conditional distribution of $X_{n+1}$ given $X_n=x$ is

$$ \begin{align*} \mathscr{L}(X_{n+1} | X_n=x) = {\tau_{\varepsilon}}(x,\cdot). \end{align*} $$

We will denote by ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$ the filtration generated by the negative-time part of such a process: for each $n\le 0$ , ${\mathscr {F}_{n}^{{\tau _{\varepsilon }}}}=\Sigma (X_m:m\le n)$ . If we want to define ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$ in a canonical way, we can always assume that in this construction, $\Omega =(\mathbb {A}^{\mathbb {Z}})^{\mathbb {Z}}$ and $\mathbb {P}$ is the law of the stationary Markov chain on $\Omega $ .

2 Usual (static) classification of filtrations

2.1 A very short abstract of the theory

We recall the main points in the usual theory of classification of filtrations. We refer to [Reference Émery and Schachermayer6, Reference Laurent11] for details.

We start with the following definition of isomorphic filtrations, which is of course equivalent to the definition provided in the cited references.

Definition 2.1. (Isomorphism of filtrations)

Let $\mathscr {F}$ and $\mathscr {F}'$ be two filtrations, possibly defined on two different probability spaces. We say that they are isomorphic if we can find two processes $X=(X_n)_{n\leq 0}$ and $X'=(X^{\prime }_n)_{n\leq 0}$ with the same law generating respectively $\mathscr {F}$ and $\mathscr {F}'$ . We write in this case $\mathscr {F}\sim \mathscr {F}'$ , or $\mathscr {F}\stackrel {X,X'}{\sim }\mathscr {F}'$ to specify the processes involved in the isomorphism.

It may not seem obvious in the above formulation that this notion of isomorphism is transitive. For details on this point, see [Reference Lanthier10, §1.2.1].

Let $\mathscr {F}$ and $\mathscr {F}'$ be two isomorphic filtrations, and let X and $X'$ be two processes such that $\mathscr {F}\stackrel {X,X'}{\sim }\mathscr {F}'$ . If Y is an $\mathscr {F}_0$ -measurable random variable, it is a classical result in measure theory that there exists some measurable map $\phi $ satisfying $Y=\phi (X)$ almost surely (a.s.). Then the isomorphism implicitly provides an $\mathscr {F}^{\prime }_0$ -measurable copy of Y, namely the random variable $Y':=\phi (X')$ . Of course, this copy depends on the choice of the processes X and $X'$ used in the isomorphism, but once these processes are fixed, the copy $Y'$ of Y is a.s. unique (because the map $\phi $ is a.s. unique with respect to the law of X).

Let $\mathscr {A}$ , $\mathscr {B}$ and $\mathscr {C}$ be three sub-sigma-algebras on the same probability space, with $\mathscr {C}\subset \mathscr {A}\cap \mathscr {B}$ . We recall that $\mathscr {A}$ and $\mathscr {B}$ are independent conditionally on $\mathscr {C}$ if, whenever X and Y are bounded real random variables, respectively $\mathscr {A}$ and $\mathscr {B}$ measurable, we have

$$ \begin{align*} \mathbb{E}[XY | \mathscr{C}] = \mathbb{E}[X | \mathscr{C}]\, \mathbb{E}[Y | \mathscr{C}] \quad(\text{a.s.}). \end{align*} $$

We note in this case that

Definition 2.2. (Immersion)

Let $\mathscr {F}$ and $\mathscr {G}$ be two filtrations on the same probability space. We say that $\mathscr {F}$ is immersed in $\mathscr {G}$ if the following conditions are satisfied:

  • for each $n\le 0$ , $\mathscr {F}_n\subset \mathscr {G}_n$ ( $\mathscr {F}$ is included in $\mathscr {G}$ );

  • for each $n\le -1$ , .

We write in this case $\mathscr {F}\prec \mathscr {G}$ .

We say that $\mathscr {F}$ and $\mathscr {G}$ are jointly immersed if both $\mathscr {F}$ and $\mathscr {G}$ are immersed in the filtration $\mathscr {F}\vee \mathscr {G}$ .

Definition 2.3. (Immersibility)

Let $\mathscr {F}$ and $\mathscr {G}$ be two filtrations, possibly defined on different probability spaces. We say that $\mathscr {F}$ is immersible in $\mathscr {G}$ if there exists a filtration $\mathscr {F}'$ immersed in $\mathscr {G}$ such that $\mathscr {F}'$ and $\mathscr {F}$ are isomorphic. We write in this case $\mathscr {F}\underset {\sim }{\prec }\mathscr {G}$ .

We can now define the following basic hierarchy of properties for filtrations.

Definition 2.4. The filtration $\mathscr {F}$ is said to be of product type if it can be generated by a process formed of independent random variables.

The filtration $\mathscr {F}$ is said to be standard if it is immersible in a filtration of product type.

The filtration $\mathscr {F}$ is said to be Kolmogorovian if its tail sigma-algebra

$$ \begin{align*} \mathscr{F}_{-\infty} := \bigcap_{n\le 0}\mathscr{F}_n \end{align*} $$

is trivial (it only contains events of probability 0 or 1).

It is a direct consequence of the definitions and of Kolmogorov’s 0–1 law that the following chain of implications holds:

$$ \begin{align*} \text{Product type} \Longrightarrow \text{Standard} \Longrightarrow \text{Kolmogorovian}. \end{align*} $$

A simple example of a standard filtration which is not of product type has been provided by Vinokurov (see [Reference Émery and Schachermayer6]). The construction of a Kolmogorovian filtration which is not standard was one of the first spectacular achievements of Vershik in this theory.

2.2 Filtrations of product type: example of ${\mathscr {F}_{}^{\tau ^{-1}}}$

We now return to the filtration ${\mathscr {F}_{}^{\tau ^{-1}}}$ defined in §1.1.1, and we use here the same notation as in that section. In particular, the process $X=(X_n)_{n\le 0}$ generating ${\mathscr {F}_{}^{\tau ^{-1}}}$ is a process where each coordinate $X_n$ takes its values in $\mathbb {A}^{\mathbb {Z}}$ , $\mathscr {L}(X_n)=\mu $ , and $X_{n-1}=\tau X_n$ .

The purpose of this section is to prove the following result.

Theorem 2.5. The filtration ${\mathscr {F}_{}^{\tau ^{-1}}}$ is of product type.

The above theorem is a direct consequence of the two following lemmas.

Lemma 2.6. For each sequence $(i_n)_{n\le 0}$ of integers, the random variables $X_n(i_n)$ , $n\le 0$ , are independent.

Lemma 2.7. Consider the sequence $(i_n)_{n\le 0}$ of integers defined by $i_n:=-\lfloor |n|/2 \rfloor $ . Then the process $(X_n(i_n))_{n\le 0}$ generates the filtration ${\mathscr {F}_{}^{\tau ^{-1}}}$ .

Proof of Lemma 2.6

Let us first consider, on a probability space $(\Omega ,\mathbb {P})$ , a random variable $Y=(Y(i):i\in \mathbb {Z})$ taking values in $\mathbb {A}^{\mathbb {Z}}$ , and such that $\mathscr {L}(Y)=\mu $ . Let us fix some integers $i,j,k$ such that $j\le k$ and $i\in \{j,\ldots ,k,k+1\}$ . For each block $w\in \mathbb {A}^{\{j,\ldots ,k\}}$ and each $a\in \mathbb {A}$ , it is straightforward to check from the construction of $\tau $ that there exists a unique block $w'\in \mathbb {A}^{\{j,\ldots ,k,k+1\}}$ such that:

  • $w'(i)=a$ ;

  • $( w'(h)+w'(h+1):j\le h\le k)=w$ .

Let us denote this block $w'$ by $\tau _{a,i}^{-1}(w)$ . We then have the following equivalence:

$$ \begin{align*} \tau Y[j,k]=w \Longleftrightarrow \text{ there exists } a\in\mathbb{A},\ Y[j,k+1]=\tau_{a,i}^{-1}(w). \end{align*} $$

But by construction of $\mu $ , all the $|\mathbb {A}| $ events $(Y[j,k+1]=\tau _{a,i}^{-1}(w))$ , $a\in \mathbb {A}$ , have the same probability $|\mathbb {A}|^{-(k-j+2)}$ (and are of course disjoint). It follows that, for each $a\in \mathbb {A}$ ,

$$ \begin{align*} \mathbb{P}(Y(i)=a\,|\,\tau Y[j,k]=w) = \mathbb{P}(Y[j,k+1]=\tau_{a,i}^{-1}(w)\,|\,\tau Y[j,k]=w) = \frac{1}{|\mathbb{A}| }. \end{align*} $$

Since this holds for any block $w\in \mathbb {A}^{\{j,\ldots ,k\}}$ , this proves that $Y(i)$ is independent of $\tau Y[j,k]$ . Then, since this is true for each $j,k$ such that $i\in \{j,\ldots ,k,k+1\}$ , $Y(i)$ is independent of $\tau Y$ .

Let us apply this with $Y=X_n$ for some $n\le 0$ : we get that, for each $i_n\in \mathbb {Z}$ , $X_n(i_n)$ is independent of $\tau X_n=X_{n-1}$ . Now if we have an arbitrary sequence $(i_n)_{n\le 0}$ of integers, we note that for each $n\ge 0$ , the random variables $X_m(i_m)$ , $m\le n-1$ , are all ${\mathscr {F}_{n-1}^{\tau ^{-1}}}$ -measurable, hence $X_n(i_n)$ is independent of $\boldsymbol {\Sigma }(X_m(i_m):m\le n-1)$ .

Lemma 2.7 will be derived from the following result (see Figure 1).

Figure 1 Illustration of Lemma 2.8: the variables marked with a circle can be computed from the ones marked with a square.

Lemma 2.8. Consider integers $m \leq n \leq 0$ , and let $(i_m,i_{m+1},\ldots ,i_n)$ be a finite sequence of integers such that, for each $m \leq \ell \leq n-1$ ,

(5) $$ \begin{align} i_{\ell+1} = i_{\ell} \quad\text{or}\quad i_{\ell+1} = i_{\ell}+1. \end{align} $$

Then $X_n[i_m,i_m+n-m]$ is measurable with respect to $\Sigma ( X_m(i_m),\ldots , X_n(i_n))$ .

Proof For a fixed m, we prove the result by induction on n. If $n=m$ the result is obvious. Assume now that, for some $m+1\le n\le 0$ , the result holds up to $n-1$ . Then $X_{n-1}[i_m,i_m+n-1-m]$ is measurable with respect to $\Sigma ( X_m(i_m),\ldots , X_{n-1}(i_{n-1}))$ . Remembering the notation $\tau _{a,i}^{-1}(w)$ from the proof of Lemma 2.6, and since by assumption we have $i_n\in \{i_m,\ldots ,i_m+n-m\}$ , we can then write

$$ \begin{align*} X_n[i_m,i_m+n-m] = \tau_{X_n(i_n),i_n}^{-1}(X_{n-1}[i_m,i_m+n-1-m]), \end{align*} $$

hence $X_n[i_m,i_m+n-m]$ is measurable with respect to $\Sigma ( X_m(i_m),\ldots , X_n(i_n))$ as claimed.

Proof of Lemma 2.7

The sequence $(i_n)$ defined in the statement of Lemma 2.7 obviously satisfies (5). Moreover, we have $i_m\to -\infty $ as $m\to -\infty $ , and for any fixed $n\le 0$ , $i_m+n-m\to +\infty $ as $m\to -\infty $ . By application of Lemma 2.8, it follows that $X_n$ is measurable with respect to $\boldsymbol {\Sigma }(X_m(i_m):m\le n)$ . But conversely, all the random variables $X_m(i_m)$ , $m\le n$ , are ${\mathscr {F}_{n}^{\tau ^{-1}}}$ -measurable. Hence these variables generate ${\mathscr {F}_{n}^{\tau ^{-1}}}$ .

2.3 Standardness and I-cosiness

The example of the filtration ${\mathscr {F}_{}^{\tau ^{-1}}}$ is very special, as it is not so hard to explain a process with independent coordinates which generates the filtration. In general, when we have a standard filtration, it may be very hard to find such a process from which the filtration is built. This is one of the reasons why several criteria of standardness have been developed, allowing the standardness of a filtration to be proved without explicitly giving the process with independent coordinates. The first such criterion was given by Vershik [Reference Vershik17], but here we will be interested in another one, called I-cosiness, introduced by Émery and Schachermayer [Reference Émery and Schachermayer6] and strongly inspired by ideas of Tsirelson [Reference Tsirelson15] and Smorodinsky [Reference Smorodinsky14].

We first have to define the concept of real-time coupling for a filtration.

Definition 2.9. (Real-time coupling)

Let $\mathscr {F}=(\mathscr {F}_n)_{n\le 0}$ a filtration on a probability space $(\Omega ,\mathbb {P})$ . A real-time coupling of $\mathscr {F}$ is a pair $(\mathscr {F}',\mathscr {F}")$ of filtrations, both defined on the same probability space (but possibly different from $(\Omega ,\mathbb {P})$ ), such that:

  • $\mathscr {F}'\sim \mathscr {F}$ ;

  • $\mathscr {F}"\sim \mathscr {F}$ ;

  • $\mathscr {F}'$ and $\mathscr {F}"$ are jointly immersed.

Such a real-time coupling of $\mathscr {F}$ is said to be independent in the distant past if there exists some integer $n_0\le 0$ such that $\mathscr {F}^{\prime }_{n_0}$ and $\mathscr {F}^{\prime \prime }_{n_0}$ are independent. In this case, we also say $n_0$ -independent if we want to highlight $n_0$ .

In practice, if we want to construct a real-time coupling of a filtration generated by a process $X=(X_n)_{n\le 0}$ , we have to build on the same probability space two copies $X'=(X^{\prime }_n)_{n\le 0}$ and $X"=(X^{\prime \prime }_n)_{n\le 0}$ of X. The joint immersion of the filtrations they generate amounts to the following conditions for each $n\le -1$ :

$$ \begin{align*} &\mathscr{L}(X^{\prime}_{n+1} | \mathscr{F}^{\prime}_n \vee \mathscr{F}^{\prime\prime}_n) = \mathscr{L}(X^{\prime}_{n+1} | \mathscr{F}^{\prime}_n),\\ &\mathscr{L}(X^{\prime\prime}_{n+1} | \mathscr{F}^{\prime}_n \vee \mathscr{F}^{\prime\prime}_n) = \mathscr{L}(X^{\prime\prime}_{n+1} | \mathscr{F}^{\prime\prime}_n). \end{align*} $$

So, we can construct such a coupling step by step: assuming that we have already defined $X^{\prime }_m$ and $X^{\prime \prime }_m$ for each $m\le n$ , the construction continues at time $n+1$ with the realization, conditionally on $\mathscr {F}^{\prime }_n\vee \mathscr {F}^{\prime \prime }_n$ , of a coupling of the two conditional laws $\mathscr {L}(X^{\prime }_{n+1} | \mathscr {F}^{\prime }_n)$ and $\mathscr {L}(X^{\prime \prime }_{n+1} | \mathscr {F}^{\prime \prime }_n)$ . This explains the name real-time coupling.

Definition 2.10. (I-cosiness)

Let $\mathscr {F}$ be a filtration, and let Y be an $\mathscr {F}_0$ -measurable random variable taking values in some Polish metric space $(E,d)$ . Then Y is said to be I-cosy (with respect to $\mathscr {F}$ ) if, for each real number $\delta>0$ , we can find a real-time coupling $(\mathscr {F}',\mathscr {F}")$ of $\mathscr {F}$ which is independent in the distant past, and such that the two copies $Y'$ and $Y"$ of Y in $\mathscr {F}'$ and $\mathscr {F}"$ respectively satisfy

(6) $$ \begin{align} d(Y',Y")<\delta \text{ with probability }> 1-\delta. \end{align} $$

The filtration itself is said to be I-cosy if each $\mathscr {F}_0$ -measurable random variable Y is I-cosy with respect to $\mathscr {F}$ .

The variable Y whose copies we want to be close together is called a target. Classical arguments of measure theory allow us to considerably reduce the number of targets to test when we want to establish I-cosiness for a given filtration (see [Reference Laurent11] or [Reference Lanthier10, §1.2.4]). In particular, we will use the following proposition.

Proposition 2.11. Assume that $\mathscr {F}_0$ is generated by a countable family $(Z_i)_{i\in I}$ of random variables taking values in finite sets. Assume also that we have written the countable set I as an increasing union of finite sets $I_k$ , $k\ge 0$ . For each integer $k\ge 0$ , denote by $Y_k$ the random variable $(Z_i)_{i\in I_k}$ . If $Y_k$ is I-cosy for each k, then the filtration is I-cosy.

Note that, in the context of the above proposition, since the set of all possible values of $Y_k$ is finite, we can endow it with the discrete metric and replace condition (6) with

$$ \begin{align*} Y^{\prime}_k=Y^{\prime\prime}_k \text{ with probability }> 1-\delta. \end{align*} $$

The following result is proved in [Reference Émery and Schachermayer6] using ideas from [Reference Vershik17] (see also [Reference Laurent11]).

Theorem 2.12. A filtration $\mathscr {F}$ is standard if and only if it is I-cosy.

2.4 Standardness of the filtration ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$

We will apply the I-cosiness criterion to prove the standardness of the filtration ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$ defined in §1.1.1, from which we will also be able to derive that this filtration is of product type. We use now the notation introduced in that section: $X=(X_n)_{n\in \mathbb {Z}}$ is a stationary Markov process with transitions given by the Markov kernel ${\tau _{\varepsilon }}$ , and for each n, $X_n$ takes its values in $\mathbb {A}^{\mathbb {Z}}$ and follows the law $\mu $ . The filtration ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$ is generated by the negative-time part of this process.

Theorem 2.13. For each $0 < \varepsilon < ({[\mathbb {A}|-1})/{|\mathbb {A}| }$ , the filtration ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$ is I-cosy, hence it is standard.

Proof Since ${\mathscr {F}_{0}^{{\tau _{\varepsilon }}}}$ is generated by the countable family of random variables $(X_n(i):$ $n\leq 0,i\in \mathbb {Z} )$ , it is sufficient by Proposition 2.11 to check that, for each integer $k\ge 0$ , the random variable $Y_k:=(X_n[-k,k]:-k\le n\leq 0 )$ is I-cosy. In fact, we will see at the end of the proof that it is enough by stationarity of the process X to consider simpler targets, which are the random variables of the form $X_0[-k,k]$ , $(k\ge 0)$ .

So we fix an integer $k\ge 0$ , we consider the target $X_0[-k,k]$ , and we fix a real number $\delta>0$ . To check the I-cosiness of $X_0[-k,k]$ , we have to construct on some probability space $(\Omega ,\mathbb {P})$ two copies $X'$ and $X"$ of the process X, such that:

  • for some $n_0\le 0$ ,

    (7) $$ \begin{align} \mathscr{F}^{\prime}_{n_0}=\boldsymbol{\Sigma}(X^{\prime}_m:m\le n_0)\text{ and }\mathscr{F}^{\prime\prime}_{n_0}=\boldsymbol{\Sigma}(X^{\prime\prime}_m:m\le n_0)\text{ are independent} \end{align} $$
    (to ensure that the filtrations $\mathscr {F}'$ and $\mathscr {F}"$ generated by the negative parts of these process are independent in the distant past);
  • for each $n_0+1\le n\le 0$ ,

    (8) $$ \begin{align} \mathscr{L}(X^{\prime}_{n} | \mathscr{F}^{\prime}_{n-1}\vee\mathscr{F}^{\prime\prime}_{n-1}) = {\tau_{\varepsilon}}(X^{\prime}_{n-1}, \cdot) \text{ and }\mathscr{L}(X^{\prime\prime}_{n} | \mathscr{F}^{\prime}_{n-1}\vee\mathscr{F}^{\prime\prime}_{n-1}) = {\tau_{\varepsilon}}(X^{\prime\prime}_{n-1}, \cdot) \end{align} $$
    (to get the joint immersion of the filtrations $\mathscr {F}'$ and $\mathscr {F}"$ ); and
  • the copies $(X^{\prime }_0[-k,k])$ and $(X^{\prime \prime }_0[-k,k])$ of the target random variable satisfy

    (9) $$ \begin{align} \mathbb{P}(X^{\prime}_0[-k,k] = X^{\prime\prime}_0[-k,k])> 1-\delta. \end{align} $$

Here is how we will proceed. We fix some $n_0\le 0$ and consider a probability space $(\Omega ,\mathbb {P})$ in which we have two independent copies $(X^{\prime }_n:n\le n_0)$ and $(X^{\prime \prime }_n:n\le n_0)$ of $(X_n:n\le n_0)$ , and an independent family of i.i.d. random variables ${(U_n(i):n\ge n_0+1,i\in \mathbb {Z})}$ which are all uniformly distributed on the interval $[0,1]$ . These random variables will be used to construct inductively the error processes $(\xi ^{\prime }_n(i):n\ge n_0+1,i\in \mathbb {Z})$ and $(\xi ^{\prime \prime }_n(i):n\ge n_0+~1, i\in \mathbb {Z})$ and the rest of the Markov processes $(X^{\prime }_n:n\ge n_0+1)$ and ${(X^{\prime \prime }_n:n\ge n_0~+~1)}$ through the formula

$$ \begin{align*} X^{\prime}_{n}:=\tau X^{\prime}_{n-1} + \xi^{\prime}_{n}\quad\text{and}\quad X^{\prime\prime}_{n}:=\tau X^{\prime\prime}_{n-1} + \xi^{\prime\prime}_{n}. \end{align*} $$

We will use the auxiliary process $Z:=X'-X"$ , and we note that for $n\ge n_0+1$ we have

(10) $$ \begin{align} Z_n=\tau Z_{n-1}+\xi^{\prime}_n-\xi^{\prime\prime}_n. \end{align} $$

Rewriting (9) with this notation, we want to achieve the coupling in such a way that, provided $|n_0|$ is large enough,

(11) $$ \begin{align} \mathbb{P}(Z_0[-k,k] = (0_{\mathbb{A}},\ldots,0_{\mathbb{A}}))> 1-\delta. \end{align} $$

We explain now how we construct the error processes inductively. Assuming that, for some $n\ge n_0+1$ , we have already defined the processes $X'$ and $X"$ up to time $n-1$ , we choose for each $i\in \mathbb {Z}$ two random maps $g^{\prime }_{n,i}$ and $g^{\prime \prime }_{n_i}$ from $[0,1]$ to $\mathbb {A}$ and set $\xi ^{\prime }_{n}(i):=g^{\prime }_{n,i}(U_n(i))$ and $\xi ^{\prime \prime }_{n}(i):=g^{\prime \prime }_{n,i}(U_n(i))$ . These maps are random in the sense that they depend on the realizations of the processes up to time $n-1$ . However, they have to comply with the required law for the errors, which we ensure by choosing them among the family of maps $(g_a)_{a\in \mathbb {A}}$ defined as follows. Recalling the definition (3) of $\tilde {\varepsilon }$ and the formulation (4) of the law of the error, we cut the interval $[0,1]$ into $|\mathbb {A}| +1$ subintervals: the first $|\mathbb {A}|$ of them are of the same length $\tilde {\varepsilon }/|\mathbb {A}|$ , and we denote them by $J_b$ , $b\in \mathbb {A}$ (they correspond to the uniform measure part of the error); the last one is $[\tilde {\varepsilon },1]$ . Then, for each $a\in \mathbb {A}$ , we define the map $g_a:[0,1]\to \mathbb {A}$ by setting

$$ \begin{align*} g_a(u):= \begin{cases} a+b &\text{ if }u\in J_b\ (b\in\mathbb{A}),\\ {0_{\mathbb{A}}} &\text{ if }u\in [\varepsilon,1]. \end{cases} \end{align*} $$

By choosing $g^{\prime }_{n,i}$ and $g^{\prime \prime }_{n_i}$ in the set $\{g_a: a\in \mathbb {A}\}$ , we get the correct conditional law of the errors $\xi ^{\prime }_n$ and $\xi ^{\prime \prime }_n$ knowing $\mathscr {F}^{\prime }_{n-1}\vee \mathscr {F}^{\prime \prime }_{n-1}$ , and we have (8). It remains to explain how we make this choice for our purposes.

Figure 2 Choice of $g^{\prime }_{n,i}$ and $g^{\prime \prime }_{n,i}$ when we apply the strategy $S_a$ .

Given a site $(n,i)$ , for each $a\in \mathbb {A}$ we define the strategy $S_a$ as the choice $(g^{\prime }_{n,i},g^{\prime \prime }_{n,i}):=(g_a,g_{{0_{\mathbb {A}}}})$ (see Figure 2). In this way, when we apply the strategy $S_a$ , we obtain

$$ \begin{align*} \xi^{\prime}_n(i)-\xi^{\prime\prime}_n(i)=\begin{cases} a &\text { if }U_n(i)\in[0,\tilde{\varepsilon}),\\ {0_{\mathbb{A}}} &\text{ otherwise.} \end{cases} \end{align*} $$

The special case $a={0_{\mathbb {A}}}$ gives $\xi ^{\prime \prime }_n(i)-\xi ^{\prime }_n(i)={0_{\mathbb {A}}}$ with probability $1$ , and thus the choice of the strategy $S_{0_{\mathbb {A}}}$ on a given site $(n,i)$ ensures that, locally, the evolution of the Z process is given by the deterministic action of the cellular automaton $\tau $ (remember (10)).

The set of sites $(n,i)$ for which we have to define $g^{\prime }_{n,i}$ and $g^{\prime \prime }_{n,i}$ is partitioned into ‘diagonals’ ${\mathcal {D}} (j)$ , $j\in \mathbb {Z}$ , where

$$ \begin{align*} {\mathcal{D}} (j) := \{ (n,i) / n\geq n_0+1,\ i=j-n\} \end{align*} $$

(see Figure 3). We observe that, for $i<j\in \mathbb {Z}$ , any error added on some site of ${\mathcal {D}} (j)$ has no influence on $Z_0(i)$ .

Figure 3 The choice of the local strategies to construct the real-time coupling.

On all the diagonals ${\mathcal {D}} (j)$ , $j<-k$ or $j>k$ , we systematically choose the strategy $S_{0_{\mathbb {A}}}$ , so that Z follows the deterministic evolution of the cellular automaton $\tau $ in these areas. It remains to explain the choice of the strategies on the sites of the diagonals ${\mathcal {D}} (j)$ , $-k\leq j\leq k$ .

Consider an integer n, $n_0<n\leq 0$ , and assume that we know the Z process up to time $n-1$ . We compute $\tau ^{|n-1|}(Z_{n-1})$ , and according to what we get we make the following choices.

  • If $\tau ^{|n-1|}(Z_{n-1})[-k,k]=({0_{\mathbb {A}}},\ldots ,{0_{\mathbb {A}}})$ , we can easily ensure that $Z_0[-k,k]=({0_{\mathbb {A}}},\ldots ,{0_{\mathbb {A}}})$ by choosing the strategy $S_{{0_{\mathbb {A}}}}$ for all the following sites. In this case we set $j_{n-1}:=k+1$ .

  • Otherwise, we define $j_{n-1}$ as the smallest integer $j\in \{-k,\ldots ,k\}$ such that

    $$ \begin{align*} \tau^{|n-1|}(Z_{n-1})(j)\neq {0_{\mathbb{A}}}, \end{align*} $$
    and we apply the strategy $S_{{0_{\mathbb {A}}}}$ at all sites $(n,i)$ except when $(n,i)$ falls on $\mathcal {D}_{j_{n-1}}$ (that is, when $i=j_{n-1}-n$ ), where we apply $S_{-a_{n-1}}$ with $a_{n-1}:= \tau ^{|n-1|}(Z_{n-1})(j_{n-1})$ .

With this method, if $j_{n-1}\in \{-k,\ldots ,k\}$ , the only i on line n for which $Z_{n,i}$ is not determined is $i=j_{n-1}-n$ , and the value of $Z_{n,j_{n-1}-n}$ now depends only on the random variable $U_n(j_{n-1}-n)$ :

  • If $U_n(j_{n-1}-n) \in [0,\tilde {\varepsilon })$ , the difference $\xi ^{\prime }_n(j_{n-1}-n)-\xi ^{\prime \prime }_n(j_{n-1}-n)$ takes the value $-a_{n-1}$ . This exactly compensates the deterministic action of $\tau $ for the target site $(0,j_{n-1})$ , which ensures that, at the next step, we get $\tau ^{|n|}(Z_{n})(j_{n-1})={0_{\mathbb {A}}}$ , and thus $j_n\ge j_{n-1}~+~1$ .

  • Otherwise, we have $\xi ^{\prime }_n(j_{n-1}-n)-\xi ^{\prime \prime }_n(j_{n-1}-n)=0$ , so that $\tau ^{|n|}(Z_{n})(j_{n-1})=a_{n-1}$ and $j_n=j_{n-1}$ .

We construct the real-time coupling inductively by applying the above method for $n_0+1\le n\le 0$ . How we build the coupling for times $n\geq 0$ is irrelevant; we can, for example, choose always the strategy $S_{{0_{\mathbb {A}}}}$ for these positive times. Now we have to prove that (11) is achieved for $|n_0|$ large enough.

We inductively define random times $T(-k),\ldots ,T(k)$ . We start with $T(-k)$ , which only depends on the random variables $U_n(i)$ , $(n,i)\in {\mathcal {D}} (-k)$ :

$$ \begin{align*} T(-k):=\min\{n \geq n_0+1 :\, U_n(-k-n) \in [0,\tilde{\varepsilon}) \}. \end{align*} $$

(With the usual convention that $\min \emptyset =+\infty $ , but observe that $T(-k)<\infty $ with probability 1.) The random variable $T(-k)-n_0$ then follows a geometric law of parameter $\tilde {\varepsilon }$ . And by construction of the coupling, for each $n_0+1\le n\le 0$ we have

$$ \begin{align*} T(-k)\leq n \Longrightarrow j_n> -k. \end{align*} $$

Then, if we have already defined $T(j)$ for some $-k\le j\le k-1$ , we set

$$ \begin{align*} T(j+1):=\min\{n\geq T(j)+1 :\, U_n(j+1-n) \in [0,\tilde{\varepsilon}) \}. \end{align*} $$

Note that $T(j+1)$ only depends on $T(j)$ and on the random variables $U_n(i)$ , $(n,i)\in {\mathcal {D}} (j+1)$ . And by construction of the coupling, we get by induction that, for each $j\in \{-k,\ldots ,k\}$ and each $n\leq 0$ ,

$$ \begin{align*}T(j)\leq n \Longrightarrow j_n> j.\end{align*} $$

In particular,

(12) $$ \begin{align} T(k)\leq 0 \Longrightarrow j_0> k \Longrightarrow Z_0[-k,k]=({0_{\mathbb{A}}},\ldots,{0_{\mathbb{A}}}). \end{align} $$

For $j\in \{-k,\ldots ,k-1\}$ , the difference $T(j+1) -T(j)$ is distributed according to a geometric law of parameter $\tilde {\varepsilon }$ , and is independent of $T(-k),\ldots ,T(j)$ . Hence $T(k)-n_0$ is distributed as the sum of $2k+1$ independent geometric random variables of parameter $\tilde {\varepsilon }$ . Therefore, as $\tilde {\varepsilon }$ and k are fixed, we get

(13) $$ \begin{align} \mathbb{P}(T(k)\leq 0) {\xrightarrow[{n_0}\to{-\infty}]{}} 1. \end{align} $$

By (13) and (12), we get that (11) is satisfied for $|n_0|$ large enough.

This proves that for each $k\ge 0$ , the target $X_0[-k,k]$ is I-cosy. Now, the same argument works also to prove the I-cosiness of $X_{-k}[-2k,2k]$ for each $k\ge 0$ . This gives us a real-time coupling up to time $-k$ for which

$$ \begin{align*} \mathbb{P}(Z_{-k}[-2k,2k]=({0_{\mathbb{A}}},\ldots,{0_{\mathbb{A}}}))>1-\delta. \end{align*} $$

Note that

$$ \begin{align*} Z_{-k}[-2k,2k]=({0_{\mathbb{A}}},\ldots,{0_{\mathbb{A}}}) \Longrightarrow \text{for all } 0\le \ell\le k,\ \tau^\ell Z_k[-k,k]=({0_{\mathbb{A}}},\ldots,{0_{\mathbb{A}}}). \end{align*} $$

So we can continue this coupling by always using the strategy $S_{{0_{\mathbb {A}}}}$ on lines $-k+1,\ldots ,0$ to prove the I-cosiness of $Y_k=(X_n[-k,k]:-k\le n\leq 0 )$ .

We further observe that the filtration ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$ is homogeneous: for each $n\le 0$ , ${\mathscr {F}_{n}^{{\tau _{\varepsilon }}}}={\mathscr {F}_{n-1}^{{\tau _{\varepsilon }}}}\vee \boldsymbol {\Sigma }(\xi _n)$ , where

  • $\xi _n$ is independent of ${\mathscr {F}_{n-1}^{{\tau _{\varepsilon }}}}$ ,

  • the law of $\xi _n$ is diffuse.

Then a direct application of Theorem A in [Reference Laurent11] yields the following corollary.

Corollary 2.14. For each $0 < \varepsilon < ({[\mathbb {A}|-1})/{|\mathbb {A}| }$ , the filtration ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$ is of product type.

2.5 Uniform cosiness and ergodicity of the Markov kernel

If we look carefully at the proof that the filtration ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$ is I-cosy, we see that the probability that the two copies of the target coincide in the real-time coupling converges to 1 as $|n_0|\to \infty $ uniformly with respect to the states of the two copies at time $n_0$ . We in fact get a stronger property than I-cosiness, which we call uniform cosiness, and which implies not only the I-cosiness of the filtration, but also the ergodicity of the Markov kernel.

Recall that a Markov kernel on a compact metric space $(\mathcal {X},d)$ is Feller if $x\mapsto P(x,\cdot )$ is continuous for the weak* topology. Any Feller Markov kernel on a compact metric space admits an invariant probability distribution on X. The Markov kernel is said to be ergodic in the Markov sense if there exists a probability distribution $\mu $ on X such that, for each probability measure $\nu $ on $\mathcal {X}$ ,

$$ \begin{align*} \nu Q^n {\xrightarrow[{n}\to{\infty}]{w*}} \mu. \end{align*} $$

In this case, $\mu $ is the unique ${Q}$ -invariant probability measure.

Definition 2.15. (Uniform cosiness)

Let ${Q}$ be a Markov kernel on a compact metric space $({\mathcal {X}},d)$ . We say that ${Q}$ is uniformly cosy if, for each $\delta>0$ , there exists $M=M(\delta )>0$ such that, whenever $n_0$ is a negative integer with $|n_0|\ge M$ , for each $x',x"\in {\mathcal {X}}$ , there exists a probability measure $m_{x',x"}$ on $({\mathcal {X}}^{\{n_0,\ldots ,0\}})^2$ , depending measurably on $(x',x")$ such that, denoting by $(X^{\prime }_n)_{n_0\leq n\leq 0}$ and $(X^{\prime \prime }_n)_{n_0\leq n\leq 0}$ the canonical processes defined by the coordinates on $({\mathcal {X}}^{\{n_0,\ldots ,0\}})^2$ , the following conditions hold.

  • The starting points of the processes are given by $X^{\prime }_{n_0}=x'$ and $X^{\prime \prime }_{n_0}=x"$ , $m_{x',x"}$ -a.s.

  • Under $m_{x',x"}$ , for each $n_0\leq n\leq -1$ , the conditional distribution

    $$ \begin{align*}\mathscr{L}( (X^{\prime}_{n+1},X^{\prime\prime}_{n+1}) \,|\, \boldsymbol{\Sigma}(X^{\prime}_{n_0},\ldots,X^{\prime}_n,X^{\prime\prime}_{n_0},\ldots,X^{\prime\prime}_n))\end{align*} $$
    is a.s. a coupling of the two probability distributions ${Q}(X^{\prime }_n,\cdot )$ and ${Q}(X^{\prime \prime }_n,\cdot )$ (we thus realize from $n_0$ a real-time coupling of two Markov processes of transitions probabilities given by ${Q}$ , one starting at $x'$ and the other at $x"$ ).
  • We have

    $$ \begin{align*} m_{x',x"}(d(X^{\prime}_0,X^{\prime\prime}_0)>\delta)<\delta. \end{align*} $$

For example, the proof of Theorem 2.13 shows that the Markov kernel defined by the PCA ${\tau _{\varepsilon }}$ is uniformly cosy.

Theorem 2.16. Let ${Q}$ be a Feller Markov kernel on the compact metric space $({\mathcal {X}},d)$ . If ${Q}$ is uniformly cosy, then ${Q}$ is ergodic in the Markov sense.

Proof Let $\mu $ be a Q-invariant probability measure on ${\mathcal {X}}$ , and let $\nu $ be any probability measure on ${\mathcal {X}}$ . We want to prove that, for each continuous function $f:{\mathcal {X}}\to \mathbb {R}$ and each $\varepsilon>0$ , for each n large enough we have

$$ \begin{align*} \bigg|\int_{{\mathcal{X}}} f\, d(\nu Q^n) - \int_{{\mathcal{X}}} f\, d\mu\bigg| \le\ \varepsilon. \end{align*} $$

Given f and $\varepsilon $ , by uniform continuity of f on the compact $\mathcal {X}$ , we get $\delta>0$ such that,

$$ \begin{align*} \text{for all } x,y\in{\mathcal{X}},\quad d(x,y)\leq \delta\Longrightarrow | f(x)-f(y) | < \varepsilon/2. \end{align*} $$

We can also assume that $\delta \|f\|_\infty <\varepsilon /4$ .

By uniform cosiness, we can take $M>0$ such that, given any integer $n_0\leq -M$ , there exists a family of probability measures $(m_{x',x"})_{x',x"\in {\mathcal {X}}}$ associated to this $\delta $ as in the statement of the definition of uniform cosiness. We then define the probability measure $m_{\mu ,\nu }$ on $({\mathcal {X}}^{\{n_0,\ldots ,0\}})^2$ by

$$ \begin{align*} m_{\mu,\nu} := \int_{{\mathcal{X}}} \int_{{\mathcal{X}}} m_{x',x"}\, d\mu(x')\, d\nu(x"). \end{align*} $$

Under $m_{\mu ,\nu }$ , the processes $(X^{\prime }_n)_{n_0\leq n\leq 0}$ and $(X^{\prime \prime }_n)_{n_0\leq n\leq 0}$ are two Markov chains with transitions given by Q, with respective initial laws $\mathscr {L}(X^{\prime }_{n_0})=\mu $ and $\mathscr {L}(X^{\prime \prime }_{n_0})=\nu $ . We then have $\mathscr {L} (X^{\prime }_{0})=\mu {Q}^{|n_0|}\mu = \mu $ by invariance of $\mu $ , and $ \mathscr {L} (X^{\prime \prime }_{0})={\nu Q}^{|n_0|}$ . The uniform cosiness yields

$$ \begin{align*} m_{\mu,\nu}(d(X^{\prime}_0,X^{\prime\prime}_0)>\delta) = \int_{{\mathcal{X}}} \int_{{\mathcal{X}}} \underbrace{m_{x',x"}(d(X^{\prime}_0,X^{\prime\prime}_0)>\delta)}_{<\delta}\, d\mu(x')\, d\nu(x") < \delta. \end{align*} $$

We then have

$$ \begin{align*} \bigg|\int_{{\mathcal{X}}} f\, d(\nu{Q}^{|n_0|}) - \int_{{\mathcal{X}}} f\, d\mu\bigg| & \leq \mathbb{E}_{m_{\mu,\nu}}[ | f(X^{\prime\prime}_0)-f(X^{\prime}_0) | ] \\ & \leq \int_{d(X^{\prime}_0,X^{\prime\prime}_0)>\delta} | f(X^{\prime\prime}_0)-f(X^{\prime}_0) | \,dm_{\mu,\nu} + \varepsilon/2 \\ & \le 2\delta\|f\|_\infty + \varepsilon/2 \le \varepsilon.\\[-38pt] \end{align*} $$

Corollary 2.17. The Markov kernel given by the cellular automaton ${\tau _{\varepsilon }}$ is ergodic in the Markov sense, and $\mu =\mathcal {U}_{\mathbb {A}}^{\otimes \mathbb {Z}}$ is the unique ${\tau _{\varepsilon }}$ -invariant probability measure.

3 Dynamical classification of factor filtrations

3.1 Factor filtrations

In the two cases we have studied above, we observe that the filtrations ${\mathscr {F}_{}^{\tau ^{-1}}}$ and ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$ both enjoy an important property: if we consider their canonical construction, which is on $\mathbb {A}^{\mathbb {Z}}$ for ${\mathscr {F}_{}^{\tau ^{-1}}}$ and on $(\mathbb {A}^{\mathbb {Z}})^{\mathbb {Z}}$ for ${\tau _{\varepsilon }}$ , we have on the ambient probability space the action of the left shift $\sigma $ , which is a measure-preserving transformation, and moreover all the sigma-algebras in those filtrations are invariant with respect to this automorphism. Our purpose in this section is to formalize such a situation, and adapt the study of the filtration to this context.

Here a dynamical system is any system of the form $(\Omega ,\mathbb {P},T)$ , where $(\Omega ,\mathbb {P})$ is a probability space, and $T:\Omega \to \Omega $ is an invertible, bi-measurable transformation which preserves the probability measure $\mathbb {P}$ . Given such a system, a factor sigma-algebra of $(\Omega ,\mathbb {P},T)$ is any sub-sigma algebra $\mathscr {A}$ of the Borel sigma algebra of $\Omega $ which is invariant by T: for each $A\in \mathscr {A}$ , $T^{-1}A$ and $TA$ are also in $\mathscr {A}$ . For any random variable $X_0$ defined on $(\Omega ,\mathbb {P})$ , as T preserves $\mathbb {P}$ , the process $X=(X(i))_{i\in \mathbb {Z}}$ defined by

(14) $$ \begin{align} X(i):=X_0\circ T^i,\quad\text{for all } i\in\mathbb{Z}, \end{align} $$

is stationary. Such a stationary process will be called a T-process. Whenever X is a T-process, the sigma-algebra generated by X is clearly a factor sigma-algebra of $(\Omega ,\mathbb {P},T)$ . Conversely, any factor sigma-algebra of $(\Omega ,\mathbb {P},T)$ is generated by some T-process (remember that all sigma-algebras are assumed to be essentially separable).

Definition 3.1. A factor filtration is any pair $(\mathscr {F},T)$ where

  • $\mathscr {F}=(\mathscr {F}_n)_{n\le 0}$ is a filtration on some probability space $(\Omega ,\mathbb {P})$ ,

  • $(\Omega ,\mathbb {P},T)$ is a dynamical system,

  • for each $n\le 0$ , $\mathscr {F}_n$ is a factor sigma-algebra of $(\Omega ,\mathbb {P},T)$ .

In view of the above discussion, in a factor filtration $(\mathscr {F},T)$ the filtration $\mathscr {F}$ is always generated by a process $(X_n)_{n\le 0}$ , where for each n, $X_n=(X_n(i))_{i\in \mathbb {Z}}$ is a T-process.

The dynamical classification of factor filtrations that we want to introduce now aims at distinguishing these objects up to the following notion of isomorphism, which is the adaptation of Definition 2.1 to the dynamical context.

Definition 3.2. (Dynamical isomorphism of factor filtrations)

Let $(\mathscr {F},T)$ and $(\mathscr {F}',T')$ be two factor filtrations, possibly defined on two different dynamical systems. We say that they are dynamically isomorphic if we can find two processes $X=(X_n)_{n\leq 0}$ and $X'=(X^{\prime }_n)_{n\leq 0}$ generating respectively $\mathscr {F}$ and $\mathscr {F}'$ , and such that:

  • for each n, $X_n=(X_n(i))_{i\in \mathbb {Z}}$ is a T-process, and $X^{\prime }_n=(X^{\prime }_n(i))_{i\in \mathbb {Z}}$ is a $T'$ -process;

  • $\mathscr {L}(X)=\mathscr {L}(X')$ .

We write in this case $(\mathscr {F},T)\sim (\mathscr {F}',T')$ , or $(\mathscr {F},T)\stackrel {X,X'}{\sim }(\mathscr {F}',T')$ to specify the processes involved in the isomorphism.

We note the following specificity of dynamical isomorphism: if $(\mathscr {F},T)\stackrel {X,X'}{\sim }(\mathscr {F}',T')$ , and if $Y'$ is the copy of an $\mathscr {F}_0$ random variable Y provided by this isomorphism, then $Y'\circ T'$ is the corresponding copy of $Y\circ T$ .

The notion of immersion is unchanged: $(\mathscr {F},T)$ and $(\mathscr {G},T)$ being two factor filtrations in the same dynamical system, we simply say that $(\mathscr {F},T)$ is immersed in $(\mathscr {G},T)$ if $\mathscr {F}$ is immersed in $\mathscr {G}$ . However, the notion of immersibility takes into account the above definition of dynamical isomorphism.

Definition 3.3. (Dynamical immersibility of factor filtrations)

Let $(\mathscr {F},T)$ and $(\mathscr {G},S)$ be two factor filtrations, possibly defined on two different dynamical systems. We say that $(\mathscr {F},T)$ is dynamically immersible in $(\mathscr {G},S)$ if there exists a factor filtration $(\mathscr {F}',S)$ immersed in $(\mathscr {G},S)$ which is dynamically isomorphic to $(\mathscr {F},T)$ .

We can now adapt the notions of product type and standardness to factor filtrations.

Definition 3.4. The factor filtration $(\mathscr {F},T)$ is said to be dynamically of product type if there exists a family $(\mathscr {G}_n)_{n\le 0}$ of independent factor sigma-algebras such that for each $n\le 0$ , $\mathscr {F}_n=\bigvee _{m\le n}\mathscr {G}_m$ .

Equivalently, $(\mathscr {F},T)$ is of product type if $\mathscr {F}$ can be generated by a process $X=(X_n)_{n\leq 0}$ whose coordinates $X_n$ are independent T-processes. Of course, if the factor filtration $(\mathscr {F},T)$ is of product type, then the filtration $\mathscr {F}$ is itself of product type, but the converse is not true (see the example in §3.2).

Definition 3.5. The factor filtration $(\mathscr {F},T)$ is said to be dynamically standard if it is dynamically immersible in some factor filtration dynamically of product type.

It is natural also to translate the notions of real-time coupling and I-cosiness to the dynamical context. The corresponding notions are formally the same as in the static case, except that we have to replace the isomorphism of filtrations by the dynamical isomorphism of factor filtrations.

Definition 3.6. (Dynamical real-time coupling)

Let $(\mathscr {F},T)$ be a factor filtration on a dynamical system $(\Omega ,\mathbb {P},T)$ . A dynamical real-time coupling of $(\mathscr {F},T)$ is a pair $((\mathscr {F}',S),(\mathscr {F}",S))$ of factor filtrations, both defined on the same dynamical system (but possibly different from $(\Omega ,\mathbb {P},T)$ ), such that:

  • $(\mathscr {F}',S)\sim (\mathscr {F},T)$ ;

  • $(\mathscr {F}",S)\sim (\mathscr {F},T)$ ;

  • $\mathscr {F}'$ and $\mathscr {F}"$ are jointly immersed.

Definition 3.7. (Dynamical I-cosiness)

Let $(\mathscr {F},T)$ be a factor filtration, and let Y be an $\mathscr {F}_0$ -measurable random variable taking values in some Polish metric space $(E,d)$ . Then Y is said to be dynamically I-cosy with respect to $(\mathscr {F},T)$ if, for each real number $\delta>0$ , we can find a dynamical real-time coupling $((\mathscr {F}',S),(\mathscr {F}",S))$ of $(\mathscr {F},T)$ which is independent in the distant past, and such that the two copies $Y'$ and $Y"$ of Y in $\mathscr {F}'$ and $\mathscr {F}"$ respectively satisfy

(15) $$ \begin{align} d(Y',Y")<\delta \text{ with probability }> 1-\delta. \end{align} $$

The factor filtration $(\mathscr {F},T)$ is said to be dynamically I-cosy if each $\mathscr {F}_0$ -measurable random variable Y is dynamically I-cosy with respect to $(\mathscr {F},T)$ .

We expect of course some relationship between dynamical I-cosiness and dynamical standardness. In the static case, showing that standardness implies I-cosiness is not very difficult, and in fact exactly the same arguments apply in the dynamical case. We provide them below.

Lemma 3.8. Let $(\mathscr {F},T)$ be a factor filtration which is dynamically of product type. Then $(\mathscr {F},T)$ is dynamically I-cosy.

Proof Let $(\Omega ,\mathbb {P},T)$ be the dynamical system where the factor filtration is defined, and let $X=(X_n)_{n\le 0}$ be a process generating $\mathscr {F}$ , where the coordinates $X_n$ are independent T-processes: $X_n=(X_n(i))_{i\in \mathbb {Z}}=(X_n(0)\circ T^i)_{i\in \mathbb {Z}}$ .

We have to check the dynamical I-cosiness with respect to $(\mathscr {F},T)$ of a target random variable Y which is $\mathscr {F}_0$ -measurable. It is enough to consider the case where Y is measurable with respect to $\boldsymbol {\Sigma }(X_{n_0+1},\ldots ,X_0)$ for some $n_0\le -1$ .

On the product dynamical system $(\Omega \times \Omega ,\mathbb {P}\otimes \mathbb {P},T\times T)$ , we have two independent copies $X^{(1)}$ and $X^{(2)}$ of X, whose coordinates are independent $(T\times T)$ processes. Let us now consider the two copies $X'$ and $X"$ of X defined by $X':=X^{(1)}$ , and

$$ \begin{align*} X^{\prime\prime}_n := \begin{cases} X^{(2)}_n & \text{ if }n\le n_0,\\ X^{(1)}_n & \text{ if }n\ge n_0+1.\ \end{cases} \end{align*} $$

Define $\mathscr {F}'$ (respectively, $\mathscr {F}"$ ) as the filtration generated by $X'$ (respectively, $X"$ ). Then $((\mathscr {F}',T\times T),(\mathscr {F}",T\times T))$ is an $n_0$ -independent dynamical coupling of $(\mathscr {F},T)$ . Moreover, since by construction we have $X^{\prime }_n=X^{\prime \prime }_n$ for $n\ge n_0+1$ , the corresponding copies of $Y'$ and $Y"$ of Y satisfy $Y'=Y"$ . This concludes the proof.

Lemma 3.9. Let $(\mathscr {F},T)$ and $(\mathscr {G},S)$ be two factor filtrations. Assume that $(\mathscr {G},S)$ is dynamically I-cosy, and that $(\mathscr {F},T)$ is dynamically immersible in $(\mathscr {G},S)$ . Then $(\mathscr {F},T)$ is dynamically I-cosy.

Proof Since cosiness is preserved by isomorphism, we can assume without loss of generality that $(\mathscr {F},T)$ is immersed in some dynamically I-cosy factor filtration $(\mathscr {G},T)$ . Let $X=(X_n)_{n\le 0}$ be a process generating $\mathscr {F}$ , where the coordinates $X_n$ are T-processes. Let Y be an $\mathscr {F}_0$ -measurable random variable, taking values in the Polish metric space $(E,d)$ , and let $\delta>0$ . Since $\mathscr {F}_0\subset \mathscr {G}_0$ , and since $(\mathscr {G},T)$ is dynamically I-cosy, there exists a dynamical real-time coupling $((\mathscr {G}',S),(\mathscr {G}",S))$ of $(\mathscr {G},T)$ , independent in the distant past, such that the corresponding copies $Y'$ and $Y"$ of Y satisfy $d(Y',Y")<\delta $ with probability at least $1-\delta $ . But the generating process X is $\mathscr {G}_0$ -measurable, so we can consider its copies $X'$ and $X"$ in $\mathscr {G}'$ and $\mathscr {G}"$ , respectively. Let $\mathscr {F}'$ (respectively, $\mathscr {F}"$ ) be the filtration generated by $X'$ (respectively, $X"$ ). Since these copies are provided by a dynamical isomorphism, the coordinates of $X'$ and $X"$ are S-processes, hence we get two dynamically isomorphic copies $(\mathscr {F}',S)$ and $(\mathscr {F}",S)$ of the factor filtration $(\mathscr {F},T)$ . Since $\mathscr {G}'$ and $\mathscr {G}"$ are jointly immersed, and $\mathscr {F}'$ (respectively, $\mathscr {F}"$ ) is immersed in $\mathscr {G}'$ (respectively, $\mathscr {G}"$ ), $\mathscr {F}'$ and $\mathscr {F}"$ are jointly immersed. Hence we get a dynamical real-time coupling $((\mathscr {F}',S),(\mathscr {F}",S))$ of the factor filtration $(\mathscr {F},T)$ . Moreover this coupling is independent in the distant past because this is the case for $((\mathscr {G}',S),(\mathscr {G}",S))$ . But $Y'$ and $Y"$ are respectively $\mathscr {F}^{\prime }_0$ - and $\mathscr {F}^{\prime \prime }_0$ -measurable, as Y is $\mathscr {F}_0$ -measurable, so they are also the corresponding copies of Y in the dynamical real-time coupling $((\mathscr {F}',S),(\mathscr {F}",S))$ .

From Lemmas 3.8 and 3.9, we immediately derive the following theorem.

Theorem 3.10. If the factor filtration $(\mathscr {F},T)$ is dynamically standard, then it is dynamically I-cosy.

Whether the converse of the above theorem is true remains an open question (see a partial result in this direction in §3.4). However, the fact that dynamical I-cosiness is necessary for dynamical standardness can already be used to establish that some factor filtrations are not dynamically standard, as in the following section.

3.2 Dynamical versus static classification of filtrations: the example of ${\mathscr {F}_{}^{\tau ^{-1}}}$

We return again to the filtration ${\mathscr {F}_{}^{\tau ^{-1}}}$ associated to the deterministic cellular automaton $\tau $ , and defined in §1.1.1. The probability measure $\mu =\mathcal {U}_{\mathbb {A}}^{\otimes \mathbb {Z}}$ on $\mathbb {A}^{\mathbb {Z}}$ is invariant by the shift map $\sigma $ . Furthermore, since the transformations $\tau $ and $\sigma $ commute, each coordinate $X_n=(X_n(i))_{i\in \mathbb {Z}}$ of the process generating the filtration ${\mathscr {F}_{}^{\tau ^{-1}}}$ is itself a stationary $\sigma $ -process: for each $i\in \mathbb {Z}$ , $X_n(i)=X_n(0)\circ \sigma ^i$ . Thus, $({\mathscr {F}_{}^{\tau ^{-1}}},\sigma )$ is a factor filtration in the dynamical system $(\mathbb {A}^{\mathbb {Z}},\mu ,\sigma )$ .

We recall that, if we look at the filtration ${\mathscr {F}_{}^{\tau ^{-1}}}$ from the static point of view, it is of product type (Theorem 2.5). However the following result shows that dynamical classification of factor filtrations may lead to different results than the static one.

Theorem 3.11. The factor filtration $({\mathscr {F}_{}^{\tau ^{-1}}},\sigma )$ is not dynamically standard.

Proof The strategy is to show that the factor filtration $({\mathscr {F}_{}^{\tau ^{-1}}},\sigma )$ , and more precisely the target $X_0(0)$ , is not dynamically I-cosy. For this, we consider a real-time dynamical coupling of this factor filtration in a dynamical system $(\Omega ,\mathbb {P},T)$ , so that we have two copies $X'=(X^{\prime }_n)_{n\leq 0}$ and $X"=(X^{\prime \prime }_n)_{n\leq 0}$ of the process $(X_n)$ , where for each $n\leq 0$ ,

  • $X^{\prime }_n=(X^{\prime }_n(i))_{i\in \mathbb {Z}}=\tau ^{|n|} X^{\prime }_0$ ,

  • $X^{\prime \prime }_n=(X^{\prime \prime }_n(i))_{i\in \mathbb {Z}}=\tau ^{|n|} X^{\prime \prime }_0$ ,

and for each $i\in \mathbb {Z}$ ,

  • $X^{\prime }_n(i)=X^{\prime }_n(0)\circ T^i$ ,

  • $X^{\prime \prime }_n(i)=X^{\prime \prime }_n(0)\circ T^i$ .

For each $n\leq 0$ , set $Z_n:=X^{\prime }_n-X^{\prime \prime }_n$ . Then, for each n, $Z_n=(Z_n(i))_{i\in \mathbb {Z}}=(Z_n(0)\circ T^i)_{i\in \mathbb {Z}}$ is a stationary T-process taking values in $\mathbb {A}$ . And since $\tau $ is an endomorphism of the group $\mathbb {A}^{\mathbb {Z}}$ , we also have $Z_n=\tau ^{|n|}Z_0$ for each $n\leq 0$ .

Assume that this coupling is $n_0$ -independent for some $n_0\leq 0$ ; in other words, $X^{\prime }_{n_0}$ and $X^{\prime \prime }_{n_0}$ are independent. Then the probability distribution of $Z_{n_0}$ is nothing more than $\mu =\mathcal {U}_{\mathbb {A}}^{\otimes \mathbb {Z}}$ . Therefore, the Kolmogorov–Sinaï entropy of this T-process is

$$ \begin{align*} h(Z_{n_0},T)=\log|\mathbb{A}|. \end{align*} $$

But we have

$$ \begin{align*} Z_{n_0}=\tau^{|n_0|}Z_0, \end{align*} $$

so that the T-process $Z_{n_0}$ is a factor of the T-process $Z_{0}$ . In particular, since the Kolmogorov–Sinaï entropy cannot increase when we pass to a factor, and observing that $Z_0$ is a T process taking values in $\mathbb {A}$

$$ \begin{align*} \log|\mathbb{A}| = h(Z_{n_0},T)\leq h(Z_{0},T)\leq \log|\mathbb{A}|. \end{align*} $$

However the only T-process taking values in $\mathbb {A}$ and whose entropy is $\log |\mathbb {A}|$ is the uniform Bernoulli process. It follows that the probability distribution of $Z_0$ is necessarily also $\mu $ . Therefore,

$$ \begin{align*} \mathbb{P}(X^{\prime}_0(0)=X^{\prime\prime}_0(0)) = \mathbb{P}(Z_0(0)=0) = \frac{1}{|\mathbb{A}|}, \end{align*} $$

which prevents the dynamical I-cosiness of the target $X_0(0)$ .

3.2.1 An alternative proof for the case of an infinite group

In this section only we allow the non-trivial Abelian group $(\mathbb {A},+)$ to be metrizable compact, not necessarily finite. The action of the deterministic cellular automaton $\tau $ on $\mathbb {A}^{\mathbb {Z}}$ can be defined by the same formula as in the finite case (1). We denote by $\mathcal {U}_{\mathbb {A}}$ the Haar measure on $\mathbb {A}$ (which is the uniform measure in the case of a finite group). Still, $\mu =\mathcal {U}_{\mathbb {A}}^{\otimes \mathbb {Z}}$ is invariant by $\tau $ and by the shift map $\sigma $ , and we can consider the filtration ${\mathscr {F}_{}^{\tau ^{-1}}}$ defined in §1.1.1, as well as the factor filtration $({\mathscr {F}_{}^{\tau ^{-1}}},\sigma )$ in this more general context. As in the finite case, we can show by exactly the same arguments that ${\mathscr {F}_{}^{\tau ^{-1}}}$ is of product type even in the non-finite case. But the above proof of Theorem 3.11 makes essential use of the fact that $\mathbb {A}$ is a finite Abelian group. We propose below an alternative proof, which replaces the entropy argument by a coupling argument, and which is still valid in the infinite case.

We first need to recall some basic facts about joinings of stationary processes in ergodic theory.

Definition 3.12. (Joining)

Let I be a finite or countable set of indices. For $i\in I$ , let $X_i$ be a $T_i$ -process in a dynamical system $(\Omega _i,{\mathbb {P}}_i,T_i)$ . A dynamical coupling or joining of the processes $(X_i)_{i\in I}$ is the simultaneous realization in the same dynamical system $(\overline \Omega ,\overline {\mathbb {P}},\overline T)$ of $\overline T$ -processes $\overline X_i$ , $i\in I$ , such that $\mathscr {L}(\overline X_i)=\mathscr {L}(X_i)$ for each $i\in I$ .

Definition 3.13. (Disjointness)

The stationary processes $X_1$ and $X_2$ are said to be disjoint if, for each joining $(\overline X_1,\overline X_2)$ of $X_1$ and $X_2$ , $\overline X_1$ and $\overline X_2$ are independent.

Disjointness was introduced by Furstenberg in 1967 in his famous paper [Reference Furstenberg7], where he also provided one of the first examples of disjointness.

Theorem 3.14. Let $X_1$ be a Bernoulli process (that is, a stationary process with independent coordinates), and let $X_2$ be a stationary process with zero entropy. Then $X_1$ and $X_2$ are disjoint.

The following result is a direct application of the construction of a relatively independent joining over a common factor (see, for example, [Reference de la Rue4]).

Lemma 3.15. Let $X_1$ and $Y_1$ be two $T_1$ -processes in a dynamical system $(\Omega _1,\mathbb {P}_1,T_1)$ , and let Let $X_2$ and $Y_2$ be two $T_2$ -processes in a second dynamical system $(\Omega _2,\mathbb {P}_2,T_2)$ . Assume that $\mathscr {L}(Y_1)=\mathscr {L}(Y_2)$ . Then there exists a joining $(\overline {X}_1,,\overline {Y}_1,\overline {X}_2,\overline {Y}_2)$ of $X_1$ , $Y_1$ , $X_2$ , $Y_2$ , such that:

  • $\mathscr {L}(\overline {X}_1,\overline {Y}_1)=\mathscr {L}({X}_1,{Y}_1)$ ;

  • $\mathscr {L}(\overline {X}_2,\overline {Y}_2)=\mathscr {L}({X}_2,{Y}_2)$ ;

  • $\overline {Y}_1=\overline {Y}_2$ a.s.

The invariance of $\mu $ by $\tau $ tells us that, if a random variable X takes its values in $\mathbb {A}^{\mathbb {Z}}$ and follows the distribution $\mu $ , then the law of $\tau X$ is also $\mu $ . The following lemma, which is the key ingredient in the alternative proof, can be viewed as the converse of this fact in the stationary case.

Lemma 3.16. Let X be a random variable taking values in $\mathbb {A}^{\mathbb {Z}}$ , whose law is $\sigma $ -invariant. If $\mathscr {L}(\tau X)=\mu $ , then $\mathscr {L}(X)=\mu $ .

Proof Let $\rho $ be a $\sigma $ -invariant probability measure on $\mathbb {A}^{\mathbb {Z}}$ , whose pushforward measure by $\tau $ is $\tau _*(\rho )=\mu $ . In the dynamical system $(\mathbb {A}^{\mathbb {Z}},\rho ,\sigma )$ , we consider the $\sigma $ -process $X_1$ defined by the coordinates, and we set $Y_1:=\tau (X_1)$ , so that $Y_1$ is a $\sigma $ -process with law $\mu $ . We also consider a second dynamical system $(\mathbb {A}^{\mathbb {Z}},\mu ,\sigma )$ , where we define the $\sigma $ -process $X_2$ by the coordinates, and we set $Y_2:=\tau (X_2)$ . We have $\mathscr {L}(X_2)=\mathscr {L}(Y_2)=\mu $ . By Lemma 3.15 there exists a joining $(\overline {X}_1,,\overline {Y}_1,\overline {X}_2,\overline {Y}_2)$ realized in a dynamical system $(\overline \Omega ,\overline {\mathbb {P}},\overline T)$ , such that:

  • $\mathscr {L}(\overline {X}_1)=\rho $ ;

  • $\mathscr {L}(\overline {X}_2)=\mu $ ;

  • $\tau (\overline {X}_1)=\overline {Y}_1=\overline {Y}_2=\tau (\overline {X}_2)$ .

Set $Z:=\overline {X}_1-\overline {X}_2$ . We have $\tau (Z)=\tau (\overline {X}_1)-\tau (\overline {X}_2)=0$ , therefore for each $i\in \mathbb {Z}$ ,

$$ \begin{align*} Z(i)=Z(0)\circ \overline{T}^i= (-1)^i Z(0). \end{align*} $$

The process Z being 2-periodic, its Kolmogorov–Sinaï entropy vanishes:

$$ \begin{align*} h(Z,\overline{T})=0. \end{align*} $$

Moreover, $\overline {X}_2$ has independent coordinates since its law is $\mu $ . By Theorem 3.14, $\overline {X}_2$ is disjoint of Z, hence independent of Z. But we have $\overline {X}_1=Z+\overline {X}_2$ , and the law of $\overline {X}_2$ is the Haar measure $\mu $ on $\mathbb {A}^{\mathbb {Z}}$ . It follows that the law $\rho $ of $\overline {X}_1$ is also $\mu $ .

Note that the assumption of $\sigma $ -invariance in Lemma 3.16 cannot be removed, because we could take a random variable Y of law $\mu $ , and define X as the only preimage of Y by $\tau $ whose coordinate $X_0$ satisfies $X_0=0_{\mathbb {A}}$ . Then we would have $\mathscr {L}(\tau X)=\mathscr {L}(Y)=\mu $ , but $\mathscr {L}(X)\neq \mu $ since $\mathbb {A}$ is not the trivial group.

We now have all the tools to modify the end of the proof of Theorem 3.11, so that it can still be valid in the infinite case.

Alternative proof of Theorem 3.11. We consider the same situation as in the first proof, but we just modify the final argument. We know that $Z_{n_0}$ has law $\mu $ , and $\tau ^{|n_0|}Z_0=Z_{n_0}$ . All processes are stationary, so we can apply Lemma 3.16 $|n_0|$ times to prove recursively that each $Z_n$ , $n_0\le n\le 0$ , has law $\mu $ . Observe also that we can choose the metric d on $\mathbb {A}$ to be invariant by translation (if we start with an arbitrary metric $d'$ defining the topology on $\mathbb {A}$ , we can set $d(a,b):=d'(a-b,b-a)$ ). Then, for each $\delta>0$ , denoting by $B_\delta $ the ball of center $0_{\mathbb {A}}$ and with radius $\delta $ , we get

$$ \begin{align*} \mathbb{P}(d(X^{\prime}_0(0),X^{\prime\prime}_0(0))<\delta)= \mathbb{P}(Z_0(0)\in B_\delta)=\mathcal{U}_{\mathbb{A}}(B_\delta). \end{align*} $$

But, as $\delta \to 0$ , $\mathcal {U}_{\mathbb {A}}(B_\delta )$ converges to $\mathcal {U}_{\mathbb {A}}(\{0_{\mathbb {A}}\})$ , and we have $\mathcal {U}_{\mathbb {A}}(\{0_{\mathbb {A}}\})<1$ since $\mathbb {A}$ is not the trivial group. This shows that the target $X_0(0)$ is not dynamically I-cosy.

3.3 The factor filtration $({\mathscr {F}_{}^{{\tau _{\varepsilon }}}},\sigma )$

We return now to the case of a finite Abelian group, and we wish to study the dynamical standardness of the factor filtration $({\mathscr {F}_{}^{{\tau _{\varepsilon }}}},\sigma )$ , where ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$ is the filtration associated to the probabilistic cellular automaton ${\tau _{\varepsilon }}$ . This section is essentially devoted to the proof of the following result.

Theorem 3.17. There exists $0<\varepsilon _0<({|\mathbb {A}|-1})/{|\mathbb {A}|}$ such that, for each $\varepsilon _0<\varepsilon \le ({|\mathbb {A}|-1})/{|\mathbb {A}|}$ , $({\mathscr {F}_{}^{{\tau _{\varepsilon }}}},\sigma )$ is dynamically standard.

For this we will construct the filtration ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$ on a bigger space than $(\mathbb {A}^{\mathbb {Z}})^{\mathbb {Z}}$ , namely we will consider the probability space $\Omega :=[0,1]^{\mathbb {Z}\times \mathbb {Z}}$ equipped with the product probability measure $\mathbb {P}:=\lambda ^{\otimes (\mathbb {Z}\times \mathbb {Z})}$ ( $\lambda $ denotes the Lebesgue measure on $[0,1]$ here). On $(\Omega ,\mathbb {P})$ we define the measure-preserving $\mathbb {Z}^2$ -action of shift maps: for $m,j\in \mathbb {Z}$ , we define

$$ \begin{align*} \sigma_{m,j}:\omega=(\omega_{n,i})_{n\in\mathbb{Z},i\in\mathbb{Z}}\mapsto \sigma_{m,j}\omega:=(\omega_{n+m,i+j})_{n\in\mathbb{Z},i\in\mathbb{Z}}, \end{align*} $$

and we simply denote by $\sigma $ the usual left-shift, $\sigma :=\sigma _{0,1}$ . For each $n\in \mathbb {Z}$ , $i\in \mathbb {Z}$ , we set $U_n(i)(\omega ):=\omega _{n,i}$ , so that $(U_n(i))_{n\in \mathbb {Z},i\in \mathbb {Z}}$ is a family of independent variables, uniformly distributed in $[0,1]$ . For each $n\in \mathbb {Z}$ , we denote by $U_n$ the $\sigma $ -process $U_n:=(U_n(i))_{i\in \mathbb {Z}}$ . Let $\mathscr {G}=(\mathscr {G}_n)_{n\leq 0}$ be the filtration generated by $(U_n)_{n\leq 0}$ . Then $(\mathscr {G},\sigma )$ is a factor filtration, and as the random variables $U_n$ , $n\in \mathbb {Z}$ , are independent, $(\mathscr {G},\sigma )$ is dynamically of product type.

Our purpose is to construct, in the dynamical system $(\Omega ,\mathbb {P},\sigma )$ , a stationary Markov process $(X_n)_{n\in \mathbb {Z}}$ with transitions given by the Markov kernel ${\tau _{\varepsilon }}$ . We will still denote by ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$ the natural filtration of the negative-time part of this process, and we want to proceed in such a way that $({\mathscr {F}_{}^{{\tau _{\varepsilon }}}},\sigma )$ is a factor filtration immersed in $(\mathscr {G},\sigma )$ . More precisely, we want the following five conditions to be realized:

  1. (a) $X_n=(X_n(i))_{i\in \mathbb {Z}}=(X_n(0)\circ \sigma ^i)_{i\in \mathbb {Z}}$ is a $\sigma $ -process;

  2. (b) $\mathscr {L}(X_n)=\mu $ ;

  3. (c) for each $x\in \mathbb {A}^{\mathbb {Z}}$ , $\mathscr {L}(X_{n+1}|X_n=x)={\tau _{\varepsilon }}(x,\cdot )$ ;

  4. (d) $X_n$ is $\mathscr {G}_n$ -measurable, namely, is a measurable function of $(U_m)_{m\le n}$ ;

  5. (e) $X_{n+1}$ is measurable with respect to $\boldsymbol {\Sigma }(X_n,U_{n+1})$ .

Conditions (a), (b) and (c) ensure that $(X_n)_{n\le 0}$ generates the correct factor filtration $({\mathscr {F}_{}^{{\tau _{\varepsilon }}}},\sigma )$ that we want to study. Condition (d) shows that ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$ is a subfiltration of $\mathscr {G}$ , and condition (e) implies that ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$ is immersed in $\mathscr {G}$ .

The key condition (d) is the most difficult to obtain. Our strategy will be strongly inspired by clever techniques introduced by Marcovici [Reference Marcovici, Beckmann, Bienvenu and Jonoska13] to prove ergodicity in the Markov sense for some PCAs. Let us start by sketching the argument. It relies on the following fact, which is a straightforward consequence of (4): for each $n,i\in \mathbb {Z}$ , the conditional distribution of $X_{n+1}(i)$ always satisfies

(16) $$ \begin{align} \mathscr{L}(X_{n+1}(i) \,|\, {\mathscr{F}_{n}^{{\tau_{\varepsilon}}}}) = \mathscr{L}(X_{n+1}(i) \,|\, X_n(i),X_n(i+1)) \ge \tilde{\varepsilon}\, \mathcal{U}_{\mathbb{A}}. \end{align} $$

Therefore, we are able to parametrize our Markov chain with the uniform random variables $U_n(i)$ , through a recursion relation of the form

$$ \begin{align*} X_{n+1}(i)={\varphi_\varepsilon}(X_n(i),X_n(i+1),U_{n+1}(i)), \end{align*} $$

where the updating function ${\varphi _\varepsilon }(\cdot ,\cdot ,u)$ only depends on u provided $u\ge 1-\tilde {\varepsilon }$ ; that is, we split the interval $[1-\tilde {\varepsilon },1]$ into $|\mathbb {A}|$ subintervals $I_a$ with length $\tilde {\varepsilon }/|\mathbb {A}|$ each, and set for each $a,b,c\in \mathbb {A}$ and $u\in I_a$ ,

$$ \begin{align*} {\varphi_\varepsilon}(b,c,u):=a. \end{align*} $$

(The complete construction of the updating function is given in §3.3.3.) In this way, the knowledge of the process U is sufficient to recover the values of the random variables $X_n(i)$ at some good sites $(n,i)$ , directly if $U_n(i)>\tilde {\varepsilon }$ , or indirectly through the recursion relation. Actually, these good sites are exactly those which are not connected to $-\infty $ in an oriented site percolation process introduced in the next section. We will see that, if $\tilde {\varepsilon }$ is not too small, then with probability $1$ there is no site connected to $-\infty $ , hence all sites are good and we get condition (d).

The whole construction of the process also borrows from [Reference Marcovici, Beckmann, Bienvenu and Jonoska13] the concept of envelope PCA. This is an auxiliary probabilistic cellular automaton, presented in §3.3.2, which will be coupled with the percolation process in §3.3.4.

3.3.1 Oriented site percolation

The percolation model that we will use is built on an infinite oriented graph whose vertices (sites) are the elements of $\mathbb {Z}^2$ , and the oriented edges connect each site $(n,i)$ to the sites $(n+1,i)$ and $(n+1,i-1)$ . We fix $p\in [0,1]$ (this parameter will be specified later), and we use the random variables $U_n(i)$ to define a random configuration of the graph: the site $(n,i)$ is declared to be open if $U_n(i)\in [0,p[$ , and closed otherwise. In this random configuration, we say that a site $(n,i)$ leads to a site $(m,j)$ if $n\le m$ and there exists in the oriented graph a path starting from $(n,i)$ , ending in $(m,j)$ and passing only through open sites, which in particular requires that both sites $(n,i)$ and $(m,j)$ be open (see Figure 4).

Figure 4 The oriented site percolation model.

For $n\geq 0$ , we denote by $O_{n}$ the event ‘there exists $i\in \mathbb {Z}$ such that $(0,0)$ leads to $(n,i)$ ’, and similarly, when $n\leq 0$ , $O_n$ stands for the events ‘there exists $i\in \mathbb {Z}$ such that $(n,i)$ leads to $(0,0)$ ’. We also set

$$ \begin{align*} O_\infty:= \bigcap_{n\geq0} O_n\quad \text{and}\quad O_{-\infty}:= \bigcap_{n\leq0}O_n. \end{align*} $$

We note that $\mathbb {P}(O_n)=\mathbb {P}(O_{-n})$ since exchanging each random variable $U_{n}(i)$ with $U_{-n}(-i)$ does not affect the law of the process, but exchanges $O_n$ and $O_{-n}$ . It follows that $\mathbb {P}(O_\infty )=\mathbb {P}(O_{-\infty })$ .

The essential question in percolation theory asks whether the probability of $O_\infty $ is positive or not. In the model presented here, this question has been addressed by Liggett [Reference Liggett12], who proved the following theorem.

Theorem 3.18. (Liggett)

There exists a critical value $p_c \in [\tfrac {2}{3},\tfrac {3}{4}]$ such that:

  • $p<p_c\Longrightarrow \mathbb {P}(O_\infty )=0$ ;

  • $p>p_c\Longrightarrow \mathbb {P}(O_\infty )>0$ .

Corollary 3.19. If $0\le p<p_c$ , the integer-valued random variable

$$ \begin{align*} N_0:=\min\{n\le0:O_n\text{ does not hold}\} \end{align*} $$

is a.s. well defined, and it is measurable with respect to $\boldsymbol {\Sigma }(U_n,n\le 0)$ .

3.3.2 Envelope automaton

The envelope automaton associated to ${\tau _{\varepsilon }}$ is an auxiliary PCA acting on configurations built on a larger alphabet: $\mathbb {A}_{\mathrm {env}}:=\mathbb {A}\cup \{{\mathop {?}} \}$ . We associate to each finite word $w\in (\mathbb {A}_{\mathrm {env}})^*$ the set $\overline {w}$ of all words in $\mathbb {A}^*$ obtained from w by replacing each question mark with any symbol from $\mathbb {A}$ . For example, if $\mathbb {A}=\{0,1\}$ and $w={\mathop {?}} 1{\mathop {?}} $ , we have $\overline {w}=\{010,011,110,111\}$ . Of course, if $w\in \mathbb {A}^*$ , we have $\overline {w}=\{w\}$ .

To define the envelope automaton we first have to introduce the local transition rules of ${\tau _{\varepsilon }}$ : if $(X_n)$ is a Markov chain evolving according the Markov kernel ${\tau _{\varepsilon }}$ , we define, for $a,b,c\in \mathbb {A}$ ,

$$ \begin{align*} f_{\tau_{\varepsilon}} (a \,|\, bc) &:= \mathbb{P} (X_{n+1}(i)=a \,|\, X_n(i)=b,X_n(i+1)=c) \\ & = \begin{cases} 1-\varepsilon &\text{ if }a=b+c,\\ \displaystyle\frac{\varepsilon}{|\mathbb{A}|-1} &\text{ otherwise. } \end{cases} \end{align*} $$

Now we define the local transition rules of the envelope PCA: if w is a word of length 2 over the alphabet $\mathbb {A}_{\mathrm {env}}$ , and $a\in \mathbb {A}$ , we set

$$ \begin{align*} f_{\tau_{\varepsilon}}^{\mathrm{env}}(a \,|\, w) := \min_{bc\in\overline{w}} f_{\tau_{\varepsilon}} (a \,|\, bc) \end{align*} $$

and

$$ \begin{align*} f_{\tau_{\varepsilon}}^{\mathrm{env}}({\mathop{?}} \,|\, w) := 1-\sum_{a\in\mathbb{A}}\min_{bc\in\overline{w}} f_{\tau_{\varepsilon}} (a \,|\, bc). \end{align*} $$

Note that, if w contains at least one ${\mathop {?}} $ , for each $a\in \mathbb {A}$ there exists $bc\in \overline {w}$ such that ${b+c\neq a}$ . Recalling the assumption (2) we made on $\varepsilon $ and the definition (3) of $\tilde {\varepsilon }$ , we thus get the following rules: for each $a,b,c\in \mathbb {A}_{\mathrm {env}}$ ,

$$ \begin{align*} f_{\tau_{\varepsilon}}^{\mathrm{env}}(a \,|\, bc) = \begin{cases} f_{\tau_{\varepsilon}}(a\,|\,bc) &\text{ if }a,b,c\in\mathbb{A},\\ 0 &\text{ if }a={\mathop{?}} \text{ and }b,c\in\mathbb{A},\\ \displaystyle\frac{\varepsilon}{|\mathbb{A}|-1} &\text{ if }a\in\mathbb{A}\text{ and }{\mathop{?}} \in\{b,c\},\\ 1-\tilde{\varepsilon} &\text{ if }a={\mathop{?}} \text{ and }{\mathop{?}} \in\{b,c\}.\\ \end{cases} \end{align*} $$

In particular, the symbol ${\mathop {?}} $ cannot appear if the two preceding symbols are in $\mathbb {A}$ , and in this case the local transition rules of the envelope PCA coincide with those of ${\tau _{\varepsilon }}$ .

For each $b,c\in \mathbb {A}_{\mathrm {env}}$ , $f_{\tau _{\varepsilon }}^{\mathrm {env}}(\cdot \,|\, bc)$ can be viewed as a probability measure on $\mathbb {A}_{\mathrm {env}}$ . Then the envelope PCA ${\tau _{\varepsilon }^{\mathrm {env}}}$ is defined as the Markov kernel on $\mathbb {A}_{\mathrm {env}}^{\mathbb {Z}}$ given by

$$ \begin{align*} {\tau_{\varepsilon}^{\mathrm{env}}}(\cdot\,|\,x):=\bigotimes_{i\in\mathbb{Z}}f_{\tau_{\varepsilon}}^{\mathrm{env}}(\cdot \,|\, x(i) x(i+1)),\quad\text{for all } x\in\mathbb{A}_{\mathrm{env}}^{\mathbb{Z}}. \end{align*} $$

3.3.3 Simultaneous implementation of the two PCAs

We will realize a coupling of ${\tau _{\varepsilon }}$ and ${\tau _{\varepsilon }^{\mathrm {env}}}$ on our space $\Omega $ , using the independent uniform random variables $(U_n(i))_{n,i\in \mathbb {Z}}$ . This will be done by means of two updating functions,

$$ \begin{align*} {\varphi_\varepsilon}:\mathbb{A}\times\mathbb{A}\times [0,1]\to\mathbb{A}\end{align*} $$

and

$$ \begin{align*} {\varphi_\varepsilon^{\mathrm{env}}}:\mathbb{A}_{\mathrm{env}}\times\mathbb{A}_{\mathrm{env}}\times [0,1]\to\mathbb{A}_{\mathrm{env}}, \end{align*} $$

adapted respectively to the local transition rules of ${\tau _{\varepsilon }}$ and ${\tau _{\varepsilon }^{\mathrm {env}}}$ . Here is how we define these updating functions (see Figure 5). We split the interval $[0,1]$ into $|\mathbb {A}|+1$ disjoint subintervals, labelled by symbols from $\mathbb {A}_{\mathrm {env}}$ . The first is

$$ \begin{align*} I_{\mathop{?}} := \bigg[0,1-\frac{\varepsilon|\mathbb{A}|}{|\mathbb{A}|-1} \bigg[ = [0,1-\tilde{\varepsilon} [, \end{align*} $$

and the complement of $I_{\mathop {?}} $ is itself split into $|\mathbb {A}|$ subintervals of equal length ${\varepsilon }/({|\mathbb {A}|-1})$ , denoted by $I_a$ , $a\in \mathbb {A}$ . Now for each $b,c\in \mathbb {A}$ , $u\in [0,1]$ , we set

$$ \begin{align*} {\varphi_\varepsilon}(b,c,u):= \begin{cases} b+c &\text{ if } u \in I_{\mathop{?}} , \\ a &\text{ if } u \in I_a,\ a\in\mathbb{A}. \end{cases} \end{align*} $$

In this way, for a random variable U uniformly distributed in $[0,1]$ , we have, for each $b,c\in \mathbb {A}$ ,

$$ \begin{align*} \mathscr{L}( {\varphi_\varepsilon}(b,c,U) ) = f_{\tau_{\varepsilon}}(\cdot\,|\,bc). \end{align*} $$

Figure 5 Values of the updating functions and coupling with the percolation model.

To define ${\varphi _\varepsilon ^{\mathrm {env}}}$ , we use the same partition of $[0,1]$ and we set, for $b,c\in \mathbb {A}_{\mathrm {env}}$ and $u\in [0,1]$ ,

$$ \begin{align*} {\varphi_\varepsilon^{\mathrm{env}}}(b,c,u):= \begin{cases} b+c &\text{ if } b,c\in\mathbb{A} \text{ and } u \in I_{\mathop{?}} , \\ {\mathop{?}} &\text{ if } b \text{ or }c\text{ is }{\mathop{?}} \text{ and }u \in I_{\mathop{?}} , \\ a &\text{ if } u \in I_a,\ a\in\mathbb{A}. \end{cases} \end{align*} $$

Likewise, we see that for a random variable U uniformly distributed in $[0,1]$ , we have for each $b,c\in \mathbb {A}_{\mathrm {env}}$

$$ \begin{align*} \mathscr{L}( {\varphi_\varepsilon^{\mathrm{env}}}(b,c,U) ) = f_{\tau_{\varepsilon}}^{\mathrm{env}}(\cdot\,|\,bc). \end{align*} $$

Let $n_0\le 0$ be a fixed integer and $x\in \mathbb {A}^{\mathbb {Z}}$ be an arbitrary configuration. We realize simultaneously two Markov chains $(X_n^{x})_{n\ge n_0}$ and $X^{\mathrm {env}}=(X^{\mathrm {env}}_n)_{n\ge n_0}$ evolving respectively according to ${\tau _{\varepsilon }}$ and ${\tau _{\varepsilon }^{\mathrm {env}}}$ . First, we initialize by setting $X^{x}_{n_0} := x$ , and $X^{\mathrm {env}}_{n_0}:= \ldots {\mathop {?}} {\mathop {?}} {\mathop {?}} {\mathop {?}} {\mathop {?}} \ldots $ (the configuration with only symbols ${\mathop {?}} $ ). Later times of the processes are computed recursively through the formula

(17) $$ \begin{align} X^{x}_{n+1}(i):= {\varphi_\varepsilon}(X^{x}_n(i),X^{x}_n(i+1),U_{n+1}(i)) \end{align} $$

and

(18) $$ \begin{align} X^{\mathrm{env}}_{n+1}(i):= {\varphi_\varepsilon^{\mathrm{env}}}(X^{\mathrm{env}}_n(i),X^{\mathrm{env}}_n(i+1),U_{n+1}(i)). \end{align} $$

3.3.4 Coupling with the percolation model

We return now to the oriented site percolation model we defined in §3.3.1 from the same random variables $(U_n(i))_{n,i\in \mathbb {Z}}$ . We just set the parameter p to be equal to $1-\tilde {\varepsilon }$ . In this way, the site $(n,i)$ is open if and only if $U_n(i)\in I_{\mathop {?}} $ (see Figure 5).

The way we constructed the three processes together yields interesting properties which we describe in the two following lemmas.

Lemma 3.20. For each $n\ge n_0+1$ , $i\in \mathbb {Z}$ , we have $X^{\mathrm {env}}_n(i)={\mathop {?}} $ if and only if in the percolation model there exists $j\in \mathbb {Z}$ such that $(n_0+1,j)$ leads to $(n,i)$ .

Proof We prove the lemma by induction on $n\ge n_0+1$ . Since $X_{n_0}^{\mathrm {env}}(i)={\mathop {?}} $ , we have $X^{\mathrm {env}}_{n_0+1}(i)={\mathop {?}} $ if and only if $U_{n_0+1}(i)\in I_{\mathop {?}} $ , that is, if and only if the site $(n_0+1,i)$ is open, which proves the result for $n=n_0+1$ . Now assume that the result is true up to level $n\ge n_0+1$ . By construction, $X^{\mathrm {env}}_{n+1}(i)={\mathop {?}} $ if and only if the two following conditions hold:

  • $X^{\mathrm {env}}_{n}(i)={\mathop {?}} $ or $X^{\mathrm {env}}_{n}(i+1)={\mathop {?}} $ , which by assumption means that there exists $j\in \mathbb {Z}$ such that $(n_0+1,j)$ leads to $(n,i)$ or to $(n,i+1)$ ;

  • $U_{n+1}(i)\in I_{\mathop {?}} $ , which means that the site $(n+1,i)$ is open.

Therefore $X^{\mathrm {env}}_{n+1}(i)={\mathop {?}} $ if and only if there exists $j\in \mathbb {Z}$ such that $(n_0+1,j)$ leads to $(n+1,i)$ , which ends the proof.

The next lemma is a key result: it shows that wherever the process $X^{\mathrm {env}}$ displays symbols different from ${\mathop {?}} $ , the process $X^x$ displays the same symbols as $X^{\mathrm {env}}$ , regardless of the initial configuration x.

Lemma 3.21. For each $n\ge n_0$ , $i\in \mathbb {Z}$ and $x\in \mathbb {A}^{\mathbb {Z}}$ , if $X^{\mathrm {env}}_{n}(i) \neq {\mathop {?}}$ then $X_n^{x}(i) = X^{\mathrm {env}}_{n}(i)$ .

Proof Again we prove the result by induction on $n\ge n_0$ . For $n=n_0$ this is obvious since $X^{\mathrm {env}}_{n_0}(i)={\mathop {?}} $ for all $i\in \mathbb {Z}$ . Now assume that the property holds up to some level $n\ge n_0$ , and let us consider some site $(n+1,i)$ . There are two cases.

  • First case: $X^{\mathrm {env}}_{n}(i)={\mathop {?}} $ or $X^{\mathrm {env}}_{n}(i+1)={\mathop {?}} $ . Then the only way to have $X^{\mathrm {env}}_{n+1}(i) \neq {\mathop {?}} $ is if $U_{n+1}(i)\in I_a$ for some $a\in \mathbb {A}$ . By construction of the updating functions, this implies that $X^{x}_{n+1}(i)=a=X^{\mathrm {env}}_{n+1}(i)$ .

  • Second case: $X^{\mathrm {env}}_{n}(i)\in \mathbb {A}$ and $X^{\mathrm {env}}_{n}(i+1)\in \mathbb {A}$ . Then by assumption this implies that we have both $X^{x}_{n}(i)=X^{\mathrm {env}}_{n}(i)$ and $X^{x}_{n}(i+1)=X^{\mathrm {env}}_{n}(i+1)$ . By construction of the updating functions, we deduce on the one hand that $X^{\mathrm {env}}_{n+1}(i)\in \mathbb {A}$ , and on the other hand that $X^{x}_{n+1}(i)=X^{\mathrm {env}}_{n+1}(i)$ .

In both cases the property also holds for the site $(n+1,i)$ .

3.3.5 Construction of the process $(X_n)$ generating ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$

We will now let $n_0$ vary in the preceding construction, and to make the dependence on $n_0$ more explicit we now denote $(X^{n_0,x}_n)_{n\ge n_0}$ and $(X^{n_0, \mathrm {env}}_n)_{n\ge n_0}$ the processes respectively initialized at the configuration x and at the configuration $\ldots {\mathop {?}} {\mathop {?}} {\mathop {?}} {\mathop {?}} \ldots ,$ and defined inductively by (17) and (18).

Since the initial configuration of the envelope processes is $\sigma $ -invariant, the action of the shifts $\sigma _{n,i}$ on the envelope processes yields the following formula, valid for all $n_0\le n$ , $i\in \mathbb {Z}$ :

(19) $$ \begin{align} X^{n_0, \mathrm{env}}_n(i) = X^{n_0-n, \mathrm{env}}_0(0)\circ\sigma_{n,i}. \end{align} $$

By a similar argument to the proof of Lemma (3.21), we get the following relations between all those processes.

Lemma 3.22. For each $n\geq n_0$ , $i\in \mathbb {Z}$ , if $X^{n_0, \mathrm {env}}_{n}(i) \neq {\mathop {?}} $ , then for all $n_1\le n_0$ we have $X^{n_1, \mathrm {env}}_{n}(i) = X^{n_0, \mathrm {env}}_{n}(i)$ .

The next proposition will be used to define the process $(X_n)_{n\in \mathbb {Z}}$ , evolving according to ${\tau _{\varepsilon }}$ , under the assumption that $\varepsilon $ is not too close to 0. We recall that $p_c$ stands for the critical value for the oriented site percolation process, defined in Theorem 3.18.

Proposition 3.23. Assume that

(20) $$ \begin{align} (1-p_c) \frac{|\mathbb{A}|-1}{|\mathbb{A}|} < \varepsilon <\frac{|\mathbb{A}|-1}{|\mathbb{A}|},\quad \mathit{that is,}0<1-\tilde{\varepsilon}<p_c. \end{align} $$

Then, with probability $1$ , for all $n,i\in \mathbb {Z}$ , there exists a random variable $N_0(n,i)\le n$ , taking values in $\mathbb {Z}$ , and satisfying the following properties.

  • $N_0(n,i)$ is measurable with respect to $\mathscr {G}_n=\boldsymbol {\Sigma }(U_m(j):\ m\le n,\ j\in \mathbb {Z})$ .

  • For each integer $n_1\le n_0\le N_0(n,i)$ , we have

    $$ \begin{align*} X_n^{n_1, \mathrm{env}}(i) = X_n^{n_0, \mathrm{env}}(i) \neq {\mathop{?}}. \end{align*} $$
  • For each $n,i\in \mathbb {Z}$ , $N_0(n,i)=N_0(0,0)\circ \sigma _{n,i}$ .

Proof Recall that the right-hand-side inequality, $\varepsilon < ({|\mathbb {A}|-1})/{|\mathbb {A}|}$ , is just the reformulation of condition (2). The left-hand-side inequality, $(1-p_c) (({|\mathbb {A}|-1})/{|\mathbb {A}|}) < \varepsilon $ , is equivalent to $p<p_c$ , where $p=1-\tilde {\varepsilon }=1- ({ \varepsilon |\mathbb {A}|}/({|\mathbb {A}|-1}))$ is the parameter of the percolation process constructed in the coupling.

Consider first the case $(n,i)=(0,0)$ . For $p<p_c$ , Corollary 3.19 ensures that the random variable $N_0=N_0(0,0)=\min \{m\le 0: O_m\text { does not hold}\}$ is a.s. well defined and $\mathscr {G}_0$ -measurable. Then Lemma 3.20 shows that, for each $n_0\le N_0(0,0)$ , $X_0^{n_0, \mathrm {env}}(0) \neq {\mathop {?}} $ , and Lemma 3.22 gives

$$ \begin{align*}n_1\le n_0\le N_0(0,0) \Longrightarrow X_0^{n_1, \mathrm{env}}(0) = X_0^{n_0, \mathrm{env}}(0).\end{align*} $$

Then we can generalize this result to any site $(n,i)\in \mathbb {Z}\times \mathbb {Z}$ , setting $N_0(n,i):=N_0\circ \sigma _{n,i}$ . Note that $N_0(n,i)$ is the greatest integer $n_0\le n$ such that there is no $j\in \mathbb {Z}$ with $(n_0,j)$ leading to $(n,i)$ .

We assume now that $\varepsilon $ satisfies (20), and we explain how to construct the process $(X_n)$ satisfying conditions (a)– (e) (see page 21). With the above proposition, we can a.s. define, for each $n,i\in \mathbb {Z}$ ,

(21) $$ \begin{align} X_n(i) := X_n^{N_0(n,i), \mathrm{env}} (i) = \lim_{n_0\to -\infty} X_n^{n_0, \mathrm{env}} (i). \end{align} $$

For each $n\le n_0\in \mathbb {Z}$ , $X_n^{n_0, \mathrm {env}}$ is measurable with respect to $\mathscr {G}_n=\boldsymbol {\Sigma }(U_m(j))_{m\le n, j\in \mathbb {Z}}$ , so the same holds for $X_n$ and we have condition (d).

Considering an integer $n_0\le \min \{ N_0(n,i), N_0(n,i+1), N_0(n+1,i) \}$ , we have simultaneously $X_n(i)=X_n^{n_0, \mathrm {env}} (i)$ , $X_n(i+1)=X_n^{n_0, \mathrm {env}} (i+1)$ and $X_{n+1}(i)=X_{n+1}^{n_0, \mathrm {env}} (i)$ . As the process $(X_n^{n_0, \mathrm {env}})$ satisfies the induction relation (18) at $(n+1,i)$ , it is therefore the same for $(X_n)$ . This proves that the process $(X_n)_{n\in \mathbb {Z}}$ actually evolves according to the PCA ${\tau _{\varepsilon }}$ , and we have conditions (c) and (e). Moreover, from (19) we get. for each $n,i\in \mathbb {Z}$ .

$$ \begin{align*} X_n(i) = X_0(0)\circ\sigma_{n,i}. \end{align*} $$

This gives on the one side that each row $X_n$ is a $\sigma $ -process, so we have condition (a), and on the other side we also get $X_n=X_0\circ \sigma _{n,0}$ . The rows $X_n$ therefore all follow the same law, which has to be $\mu $ by Corollary 2.17, and we have condition (b). This concludes the proof of Theorem 3.17.

Remark 3.24. As already pointed out in the sketch of the proof, the argument relies on inequality (16), hence the result remains true for any PCA satisfying such inequality. In particular, the method can be applied to any probabilistic cellular automaton on $\mathbb {A}^{\mathbb {Z}}$ which is a random perturbation of a deterministic cellular automaton by addition of a random error $\xi $ , provided the law of each $\xi _n(i)$ is bounded below by $\tilde {\varepsilon }\,\mathcal {U}_{\mathbb {A}}$ .

We can even note that, in (16), the uniform measure $\mathcal {U}_{\mathbb {A}}$ does not play a particular role and can be replaced by any fixed probability measure on $\mathbb {A}$ (we just have to adapt the length of the subintervals $(I_a)_{a\in \mathbb {A}}$ accordingly). Therefore the argument also applies to any PCA $\gamma $ on $\mathbb {A}^{\mathbb {Z}}$ for which the conditional law of $X_{n+1}(i)$ is of the form

$$ \begin{align*} \mathbb{P}(X_{n+1}(i)=a \,|\, \boldsymbol{\Sigma}(X_m:m\le n)) = f_\gamma(a\,|\,X_n(i),X_n(i+1)), \end{align*} $$

and the local transition rules $f_\gamma $ satisfy

$$ \begin{align*} \text{for all } a,b,c\in\mathbb{A},\quad f_\gamma(a\,|\,bc) \ge \tilde{\varepsilon}\rho(a), \end{align*} $$

for some fixed probability measure $\rho $ on $\mathbb {A}$ (here no algebraic assumption on $\mathbb {A}$ is needed).

3.3.6 Questions about the factor filtration $({\mathscr {F}_{}^{{\tau _{\varepsilon }}}},\sigma )$

When $\varepsilon $ satisfies (20), we prove the dynamical standardness of $({\mathscr {F}_{}^{{\tau _{\varepsilon }}}},\sigma )$ by immersing it into a factor filtration $(\mathscr {G},\sigma )$ dynamically of product type. The way we construct $\mathscr {G}$ , it is clear that it carries more information than ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$ , so that ${\mathscr {F}_{}^{{\tau _{\varepsilon }}}}$ is a strict subfiltration of $\mathscr {G}$ . Thus we may ask whether, in this case, $({\mathscr {F}_{}^{{\tau _{\varepsilon }}}},\sigma )$ is itself a factor filtration dynamically of product type.

Another natural question is of course whether the dynamical standardness of $({\mathscr {F}_{}^{{\tau _{\varepsilon }}}},\sigma )$ persists as $\varepsilon $ gets close to 0, which prevents the percolation argument from applying. One could already ask whether, for $\varepsilon $ close to 0, the criterion of dynamical I-cosiness is satisfied for the factor filtration $({\mathscr {F}_{}^{{\tau _{\varepsilon }}}},\sigma )$ . We do not have the answer to this question, but we provide below a somewhat naive argument suggesting that $({\mathscr {F}_{}^{{\tau _{\varepsilon }}}},\sigma )$ might not be dynamically I-cosy when $\varepsilon $ is close enough to 0.

We consider the case where the group $\mathbb {A}$ is $\mathbb {Z}/2\mathbb {Z}$ . In an attempt to establish dynamical I-cosiness, we try to realize a real-time dynamical coupling of $({\mathscr {F}_{}^{{\tau _{\varepsilon }}}},\sigma )$ , independent in the distant past, and for which the two copies $X'$ and $X"$ of the process $X=(X_n)$ satisfy $X^{\prime }_0(0)=X^{\prime \prime }_0(0)$ with probability close to 1. We want to construct this coupling in a dynamical system $(\Omega ,\mathbb {P},T)$ , in which for some $n_0<0$ we already have two independent copies $(X^{\prime }_n)_{n\le n_0}$ and $(X^{\prime \prime }_n)_{n\le n_0}$ of $(X_n)_{n\le n_0}$ (of course, each $X^{\prime }_n$ and each $X^{\prime \prime }_n$ is supposed to be a T-process of law $\mu $ ). We assume that we also have an independent family $(U_n)_{n\ge n_0+1}=(U_n(i):i\in \mathbb {Z})_{n\ge n_0+1}$ of T-processes, each $U_n$ being of law $\lambda ^{\otimes \mathbb {Z}}$ ( $\lambda $ is the Lebesgue measure on $[0,1]$ ), and the $U_n$ being independent. We want to use this family to construct error processes $(\xi ^{\prime }_n)_{n\ge n_0+1}$ and $(\xi ^{\prime \prime }_n)_{n\ge n_0+1}$ , and then complete the processes $X'$ and $X"$ inductively via the usual relations

$$ \begin{align*} X^{\prime}_{n}=\tau X^{\prime}_{n-1}+\xi^{\prime}_n\quad\text{and}\quad X^{\prime\prime}_{n}=\tau X^{\prime\prime}_{n-1}+\xi^{\prime\prime}_n\quad(n\ge n_0+1). \end{align*} $$

As in §2.4, we still want to define the error processes in the form $\xi ^{\prime }_n(i)=g^{\prime }_{n,i}(U_n(i))$ and $\xi ^{\prime \prime }_n(i)=g^{\prime \prime }_{n,i}(U_n(i))$ , where $g^{\prime }_{n,i}$ and $g^{\prime \prime }_{n,i}$ are random updating functions, but we point out the essential difference with the construction in §2.4: we now want the law of the coupling to be invariant with respect to the left shift of coordinates. This means that these updating functions are still allowed to depend on the realizations of the processes up to time $n-1$ , but in an equivariant way. In other words, they can be chosen according to what we see around the site $(n,i)$ , but the rule for the choice should be independent of i.

We also introduce the auxiliary process $Z:=X'-X"$ , which satisfies the same relation (10) for $n\ge n_0+1$ . To get dynamical I-cosiness, we have to realize the coupling in such a way that

$$ \begin{align*} \mathbb{P}(Z_0(i)=0){\xrightarrow[{n_0}\to{-\infty}]{}} 1 \end{align*} $$

(note that the probability on the left-hand side does not depend on i).

Here is a natural idea to get what we are aiming at: we wish to encourage as much as possible the appearance of $0$ s in the Z process. For this, let us observe that we have, for each $n,i\in \mathbb {Z}$ ,

(22) $$ \begin{align} 0 \leq \mathbb{P}({Z}_{n}(i) \neq \tau {Z}_{n-1}(i)) & = \mathbb{P}(\xi^{\prime}_n(i) \neq \xi^{\prime\prime}_n(i)) \nonumber\\& \leq \mathbb{P}(\xi^{\prime}_n(i) \neq 0 \text{ or } \xi^{\prime\prime}_n(i)\neq 0) \nonumber\\ & \leq \mathbb{P}(\xi^{\prime}_n(i) \neq 0) + \mathbb{P}(\xi^{\prime\prime}_n(i)\neq 0) \nonumber\\ & \leq 2 \varepsilon. \end{align} $$

Moreover, we have the possibility of making the inequality on the left-hand side an equality, by deciding to take $g^{\prime }_{n,i}=g^{\prime \prime }_{n,i}$ (we call this Option 0). On the other hand, we can also choose to turn the right-hand-side inequalities into equalities, by choosing $g^{\prime }_{n,i}$ and $g^{\prime \prime }_{n,i}$ in such a way that the subsets $\{u\in [0,1]:g^{\prime }_{n,i}(u)=1\}$ and $\{u\in [0,1]:g^{\prime \prime }_{n,i}(u)=1\}$ are disjoint (we call this Option 1). So a natural dynamical coupling is obtained by the following strategy. Knowing the processes up to time $n-1$ , we compute $\tau Z_{n-1}(i)$ . If $\tau Z_{n-1}(i)=0$ , we decide to keep this 0 by choosing Option 0. If $\tau Z_{n-1}(i)=1$ , we maximize the probability that ${Z}_{n}(i)$ will still be 0 by choosing Option 1.

Applying systematically this strategy, we are left with a process $(Z_n)_{n\ge n_0}$ taking values in $\mathbb {Z}/2\mathbb {Z}$ , which evolves according to the following PCA rule: for each $n,i$ , if $\tau Z_{n-1}(i)=0$ then $Z_n(i)=0$ , otherwise $Z_n(i)=0$ with probability $2\varepsilon $ . Observe that the initial configuration $Z_{n_0}$ has law $\mathcal {U}_{\mathbb {Z}/2\mathbb {Z}}^{\otimes \mathbb {Z}}$ , so that the initial density of 1s is $1/2$ . The question now is whether those 1s can survive over the long term.

It turns out that this type of evolution falls into the domain of a family of PCAs studied by Bramson and Neuhauser [Reference Bramson and Neuhauser1], who proved that if $\varepsilon $ is small enough, then symbols 1 survive. More precisely, we can state the following consequence of their result.

Theorem 3.25. (Bramson and Neuhauser)

In the coupling described above, for $\varepsilon>0$ small enough, there exists $\rho =\rho (\varepsilon )>0$ such that

$$ \begin{align*} \liminf_{n_0\to-\infty} \mathbb{P}(Z_0(0)=1) \geq \rho. \end{align*} $$

This shows that this attempt to get dynamical I-cosiness for $({\mathscr {F}_{}^{{\tau _{\varepsilon }}}},\sigma )$ fails if $\varepsilon $ is small.

3.4 A partial converse to Theorem 3.10

The question whether dynamical I-cosiness of a factor filtration implies its dynamical standardness (as in the static case) is so far open. However, the construction given in §3.3 (for $\varepsilon $ not too close to 0) provides an example where we are able to derive dynamical standardness from a particular form of dynamical I-cosiness. We want now to formalize this kind of situation, which can be viewed as the dynamic counterpart of the so-called Rosenblatt self-joining criterion in [Reference Laurent11].

In a dynamical system $(\Omega ,\mathbb {P},T)$ , consider a factor filtration $(\mathscr {F},T)$ generated by a family $X=(X_n)_{n\le 0}$ of T-processes, each $X_n(i)$ taking its values in a fixed finite set $\mathbb {A}$ . We also assume the existence of a factor filtration $(\mathscr {G},T)$ , dynamically of product type, and generated by a family of independent T-processes $(U_n)_{n\le 0}$ satisfying, for each $n\le 0$ ,

  • $U_n$ is independent of $\mathscr {F}_{n-1}\vee \mathscr {G}_{n-1}$ ,

  • $\mathscr {F}_n\subset \mathscr {F}_{n-1}\vee \boldsymbol {\Sigma }(U_n)$ .

(The two conditions above mean that $(U_n)_{n\le 0}$ is a superinnovation of $\mathscr {F}$ , according to [Reference Laurent11].) In particular, there exists a measurable map $\phi _n$ such that

$$ \begin{align*} X_n = \phi_n(U_n,X_{n-1},X_{n-2},\ldots). \end{align*} $$

We can also assume without loss of generality that there exists in $(\Omega ,\mathbb {P},T)$ a copy $X'$ of X, whose coordinates $X^{\prime }_n$ are also T-processes, and which is independent of $\mathscr {G}_0\vee \mathscr {F}_0$ (if not, we place ourselves on the product $(\Omega \times \Omega ,\mathbb {P}\otimes \mathbb {P},T\times T)$ as in the proof of Lemma 3.8). Now we build a family of dynamical real-time couplings of $(\mathscr {F},T)$ in the following way: for a fixed $n_0\le 0$ , define $\bar X^{n_0}={(\bar X_n^{n_0})}_{n\le 0}$ by

  • $\bar X_n^{n_0}=X^{\prime }_n$ for $n\le n_0$ ,

  • then inductively, for $n_0+1\le n\le 0$ , $\bar X_n^{n_0}:=\phi _n(U_n,\bar X_{n-1}^{n_0},\bar X_{n-2}^{n_0},\ldots ).$

(The important point here is that, for $n\ge n_0+1$ , the same random variable $U_n$ is involved in the inductive relations giving $X_n$ and $\bar X_n^{n_0}$ .) Then the filtration $\bar {\mathscr {F}}^{n_0}$ generated by $\bar X^{n_0}$ is dynamically isomorphic to $\mathscr {F}$ , and the pair $((\mathscr {F},T),(\bar {\mathscr {F}}^{n_0},T))$ is a dynamical real-time coupling of $(\mathscr {F},T)$ , which is $n_0$ independent.

Definition 3.26. (Simple dynamical I-cosiness)

If, in the above construction, we have, for each integer $m\le 0$ ,

(23) $$ \begin{align} \mathbb{P}(\bar X^{n_0}_m(0) = X_m(0)){\xrightarrow[{n_0}\to{-\infty}]{}} 1, \end{align} $$

then we say that the factor filtration $(\mathscr {F},T)$ is simply dynamically I-cosy.

Clearly, simple dynamical I-cosiness implies dynamical I-cosiness. Observe also that, when $\varepsilon $ satisfies (20), the results of §3.3 prove that the factor filtration $({\mathscr {F}_{}^{{\tau _{\varepsilon }}}},\sigma )$ is simply dynamically I-cosy.

The following theorem generalizes the conclusion of §3.3.

Theorem 3.27. If the factor filtration $(\mathscr {F},T)$ is simply dynamically I-cosy, then it is dynamically standard.

Proof We fix some integer $m\le 0$ . For each $n_0\le m-1$ , we denote by $U_{n_0+1}^m$ the collection of T-processes $U_{n_0+1},\ldots ,U_m$ . We have

$$ \begin{align*} \mathbb{P}(\bar X^{n_0}_m(0) = X_m(0)) & = \sum_{a\in\mathbb{A}} \mathbb{P}(\bar X^{n_0}_m(0) = X_m(0) = a) \\ & = \sum_{a\in\mathbb{A}} \mathbb{E}_{\mathbb{P}}[ \mathbb{P}(\bar X^{n_0}_m(0) = X_m(0) = a \, | \, U_{n_0+1}^m) ]. \end{align*} $$

But $\bar X^{n_0}_m(0)$ and $X_m(0)$ are independent conditionally on $U_{n_0+1}^m$ : indeed, once we know $U_{n_0+1}^m$ , $\bar X^{n_0}_m(0)$ only depends on $\bar X_{n_0}$ while $X_m(0)$ only depends on $X_{n_0}$ . We thus get

$$ \begin{align*} \mathbb{P}(\bar X^{n_0}_m(0) = X_m(0)) & = \sum_{a\in\mathbb{A}} \mathbb{E}_{\mathbb{P}}[ \mathbb{P}(\bar X^{n_0}_m(0) = a\, | \, U_{n_0+1}^m) \mathbb{P}(X_m(0) = a \, | \, U_{n_0+1}^m) ]\\ & = \sum_{a\in\mathbb{A}} \mathbb{E}_{\mathbb{P}}[ \mathbb{P}(X_m(0) = a \, | \, U_{n_0+1}^m)^2 ]\\ & =\mathbb{E}_{\mathbb{P}}\bigg[ \sum_{a\in\mathbb{A}} \mathbb{P}(X_m(0) = a \, | \, U_{n_0+1}^m)^2 \bigg]. \end{align*} $$

Let $M_{n_0}$ be the random variable measurable with respect to $\boldsymbol {\Sigma }(U_{n_0+1}^m)$ and defined by

$$ \begin{align*} M_{n_0} := \arg\max_{a\in\mathbb{A}} \mathbb{P}(X_m(0) = a \, | \, U_{n_0+1}^m). \end{align*} $$

(If the maximum is reached for several symbols in $\mathbb {A}$ , we take the first of them with respect to some previously chosen total order on $\mathbb {A}$ .) We get

$$ \begin{align*} \mathbb{P}(\bar X^{n_0}_m(0) = X_m(0)) & \le \mathbb{E}_{\mathbb{P}}\bigg[ \sum_{a\in\mathbb{A}} \mathbb{P}(X_m(0) = M_{n_0} \, | \, U_{n_0+1}^m)\mathbb{P}(X_m(0) = a \, | \, U_{n_0+1}^m) \bigg]\\ & = \mathbb{E}_{\mathbb{P}}\bigg[ \mathbb{P}(X_m(0) = M_{n_0} \, | \, U_{n_0+1}^m) \underbrace{\sum_{a\in\mathbb{A}} \mathbb{P}(X_m(0) = a \, | \, U_{n_0+1}^m)}_{=1} \bigg]\\ & = \mathbb{P}(X_m(0) = M_{n_0} ). \end{align*} $$

But we assumed that $(\mathscr {F},T)$ is simply dynamically I-cosy. So we have (23), which yields

$$ \begin{align*} \mathbb{P}(X_m(0) = M_{n_0} ) {\xrightarrow[{n_0}\to{-\infty}]{}} 1. \end{align*} $$

For each integer $k\ge 1$ , we can thus find $n_k\le m-1$ such that

$$ \begin{align*} \mathbb{P}(X_m(0) \neq M_{n_k} ) < 2^{-k}. \end{align*} $$

By Borel–Cantelli we get $X_m(0)=\lim _{k\to \infty }M_{n_k}$ ( $\mathbb {P}$ -a.s.). Moreover, $M_{n_k}$ is always $\mathscr {G}_m$ -measurable. Therefore $X_m(0)$ is measurable with respect to $\mathscr {G}_{m}$ , and since the sigma-algebras are factor sigma-algebras, the whole T-process $X_m$ is $\mathscr {G}_{m}$ -measurable.

This already proves that $\mathscr {F}$ is a subfiltration of the natural filtration $\mathscr {G}$ of the process $(U_n)_{n\le 0}$ , hence $(U_n)_{n\le 0}$ is a generating superinnovation of $\mathscr {F}$ . By [Reference Laurent11], it follows that $\mathscr {F}$ is immersed in $\mathscr {G}$ . Since the factor filtration $(\mathscr {G},T)$ is dynamically of product type, we get the dynamical standardness of $(\mathscr {F},T)$ .

Acknowledgements

The authors wish to thank Irène Marcovici for fruitful conversations which inspired the subject of this paper, and Séverin Benzoni, Christophe Leuridan, Alejandro Maass and an anonymous referee for their careful reading, relevant questions and valuable comments. Research partially supported by the Regional Research Project Moustic (Normandie).

References

Bramson, M. and Neuhauser, C.. Survival of one-dimensional cellular automata under random perturbations. Ann. Probab. 22(1) (1994), 244263.CrossRefGoogle Scholar
Bressaud, X., Maass, A., Martinez, S. and San Martin, J.. Stationary processes whose filtrations are standard. Ann. Probab. 34(4) (2006), 15891600.CrossRefGoogle Scholar
Ceillier, G. and Leuridan, C.. Sufficient conditions for the filtration of a stationary processes to be standard. Probab. Theory Related Fields 167(3–4) (2017), 979999.CrossRefGoogle Scholar
de la Rue, T.. An introduction to joinings in ergodic theory . Discrete Contin. Dyn. Syst. 15(1) (2006), 121142.CrossRefGoogle Scholar
Dubins, L., Feldman, J., Smorodinsky, M. and Tsirelson, B.. Decreasing sequences of $\sigma$ -fields and a measure change for Brownian motion. I. Ann. Probab. 24(2) (1996), 882904.CrossRefGoogle Scholar
Émery, M. and Schachermayer, W.. On Vershik’s standardness criterion and Tsirelson’s notion of cosiness . Séminaire de Probabilités XXXV. Ed. J. Azéma, M. Émery, M. Ledoux and M.Yor. Springer, Berlin, 2001, pp. 265305.CrossRefGoogle Scholar
Furstenberg, H.. Disjointness in ergodic theory, minimal sets, and a problem in Diophantine approximation . Math. Syst. Theory 1 (1967), 149.CrossRefGoogle Scholar
Hoffman, C.. A zero entropy $\mathrm{T}$ such that the $\left[\mathrm{T},\mathrm{Id}\right]$ endomorphism is nonstandard. Proc. Am. Math. Soc. 128(1) (2000), 183188.CrossRefGoogle Scholar
Hoffman, C. and Rudolph, D.. A dyadic endomorphism which is Bernoulli but not standard. Israel J. Math. 130 (2002), 365379.CrossRefGoogle Scholar
Lanthier, P.. Aspects ergodiques et algébriques des automates cellulaires. PhD Thesis, Université de Rouen Normandie, 2020.Google Scholar
Laurent, S.. On standardness and I-cosiness . Séminaire de Probabilités XLIII, Poitiers, France, Juin 2009. Ed. C. Donati-Martin, A. Lejay and A.Rouault. Springer, Berlin, 2011, pp. 127186.Google Scholar
Liggett, T. M.. Survival of discrete time growth models, with applications to oriented percolation. Ann. Appl. Probab. 5(3) (1995), 613636.CrossRefGoogle Scholar
Marcovici, I.. Ergodicity of noisy cellular automata: the coupling method and beyond . Pursuit of the Universal. Proc. 12th Conf. on Computability in Europe, CiE 2016 (Paris, France, June 27–July 1, 2016). Ed. Beckmann, A., Bienvenu, L. and Jonoska, N.. Springer, Cham, 2016, pp. 153163.Google Scholar
Smorodinsky, M.. Processes with no standard extension . Israel J. Math. 107 (1998), 327331.CrossRefGoogle Scholar
Tsirelson, B.. Triple points: from non-Brownian filtrations to harmonic measures . Geom. Funct. Anal. 7(6) (1997), 10961142.CrossRefGoogle Scholar
Vershik, A. M.. Decreasing sequences of measurable partitions and their applications. Sov. Math. Dokl. 11 (1970), 10071011.Google Scholar
Vershik, A. M.. The theory of decreasing sequences of measurable partitions . St. Petersburg Math. J. 6(4) (1994), 168.Google Scholar
Figure 0

Figure 1 Illustration of Lemma 2.8: the variables marked with a circle can be computed from the ones marked with a square.

Figure 1

Figure 2 Choice of $g^{\prime }_{n,i}$ and $g^{\prime \prime }_{n,i}$ when we apply the strategy $S_a$.

Figure 2

Figure 3 The choice of the local strategies to construct the real-time coupling.

Figure 3

Figure 4 The oriented site percolation model.

Figure 4

Figure 5 Values of the updating functions and coupling with the percolation model.