Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

1 Definition of a Gibbs Random Field

The notion of Gibbs random fields was formalized by mathematicians relatively recently. Before that, these fields were known in physics, particularly in statistical physics and quantum field theory. Later, it was understood that Gibbs fields play an important role in many applications of probability theory. In this section we define the Gibbs fields and discuss some of their properties.

We shall deal with random fields with a finite state space X defined over \({\mathbb{Z}}^{d}\). The realizations of the field will be denoted by \(\omega = ({\omega }_{k},k \in {\mathbb{Z}}^{d})\), where \({\omega }_{k}\) is the value of the field at the site k.

Let V and W be two finite subsets of \({\mathbb{Z}}^{d}\) such that \(V \subset W\) and \(\mathrm{dist}(V, {\mathbb{Z}}^{d}\setminus W) > R\) for a given positive constant R. We can consider the following conditional probabilities:

$$\mathrm{P}({\omega }_{k} = {i}_{k},k \in V \vert {\omega }_{k} = {i}_{k},k \in W\setminus V ),\ \ \mathrm{where}\ {i}_{k} \in X\ \ \mathrm{for}\ k \in W.$$

Definition 22.1

A random field is called a Gibbs field with memory R if, for any finite sets V and W as above, these conditional probabilities (whenever they are defined) depend only on those of the values \({i}_{k}\) for which \(\mathrm{dist}(k,V ) \leq R\).

Note that the Gibbs fields can be viewed as generalizations of Markov chains. Indeed, consider a Markov chain on a finite state space. The realizations of the Markov chain will be denoted by \(\omega = ({\omega }_{k},k \in \mathbb{Z})\). Let \({k}_{1},{k}_{2},{l}_{1}\) and \({l}_{2}\) be integers such that \({k}_{1} < {l}_{1} \leq {l}_{2} < {k}_{2}\). Consider the conditional probabilities

$$f({i}_{{k}_{1}},\ldots, {i}_{{l}_{1}-1},{i}_{{l}_{2}+1},\ldots, {i}_{{k}_{2}}) =$$
$$\mathrm{P}({\omega }_{{l}_{1}}\! =\! {i}_{{l}_{1}},\ldots, {\omega }_{{l}_{2}}\! =\! {i}_{{l}_{2}}\vert {\omega }_{{k}_{1}}\! =\! {i}_{{k}_{1}},\ldots, {\omega }_{{l}_{1}-1}\! =\! {i}_{{l}_{1}-1},{\omega }_{{l}_{2}+1}\! =\! {i}_{{l}_{2}+1},\ldots, {\omega }_{{k}_{2}}\! =\! {i}_{{k}_{2}})$$

with \({i}_{{l}_{1}}\),…,\({i}_{{l}_{2}}\) fixed. It is easy to check that whenever f is defined, it depends only on \({i}_{{l}_{1}-1}\) and \({i}_{{l}_{2}+1}\) (see Problem 12, Chap. 5). Thus, a Markov chain is a Gibbs field with \(d = 1\) and \(R = 1\).

Let us introduce the notion of the interaction energy. Let \({N}_{d,R}\) be the number of points of \({\mathbb{Z}}^{d}\) that belong to the closed ball of radius R centered at the origin. Let U be a real-valued function defined on \({X}^{{N}_{d,R}}\). As arguments of U we shall always take the values of the field in a ball of radius R centered at one of the points of \({\mathbb{Z}}^{d}\). We shall use the notation \(U({\omega }_{k};{\omega }_{k^{\prime}},0 < \vert k^{\prime} - k\vert \leq R)\) for the value of U on a realization ω in the ball centered at k and call U the interaction energy with radius R.

For a finite set \(V \subset {\mathbb{Z}}^{d}\), its R-neighborhood will be denoted by \({V }^{R}\),

$${V }^{R} =\{ k :\mathrm{ dist}(V,k) \leq R\}.$$

Definition 22.2

A Gibbs field with memory 2R is said to correspond to the interaction energy U if

$$\quad\mathrm{P}({\omega }_{k} = {i}_{k},k \in V \vert {\omega }_{k} = {i}_{k},k \in {V }^{2R}\setminus V )$$
$$= \frac{\exp (-{\sum \nolimits }_{k\in {V }^{R}}U({i}_{k};{i}_{k^{\prime}},0 < \vert k^{\prime} - k\vert \leq R))} {Z({i}_{k},k \in {V }^{2R}\setminus V )}, $$
(22.1)

where \(Z = Z({i}_{k},k \in {V }^{2R}\setminus V )\) is the normalization constant, which is called the partition function,

$$Z({i}_{k},k \in {V }^{2R}\setminus V ) = \sum\limits_{\{{i}_{k},k\in V \}}\exp (-\sum\limits_{k\in{V}^{R}}U({i}_{k};{i}_{k^{\prime}},0 < \vert k^{\prime} - k\vert\leq R)).$$

The equality (22.1) for the conditional probabilities is sometimes called the Dobrushin-Lanford-Ruelle or, simply, DLR equation after three mathematicians who introduced the general notion of a Gibbs random field. The minus sign is adopted from statistical physics.

Let \(\Omega (V )\) be the set of configurations \(({\omega }_{k},k \in V )\). The sum

$$\sum\limits_{k\in {V }^{R}}U({\omega }_{k};{\omega }_{k^{\prime}},0 < \vert k^{\prime} - k\vert \leq R)$$

is called the energy of the configuration \(\omega \in \Omega (V )\). It is defined as soon as we have the boundary conditions \({\omega }_{k}\), \(k \in {V }^{2R}\setminus V\).

Theorem 22.3

For any interaction energy U with finite radius, there exists at least one Gibbs field corresponding to U.

Proof

Take a sequence of cubes V i centered at the origin with sides of length 2i. Fix arbitrary boundary conditions \(\{{\omega }_{k},k \in {V }_{i}^{2R}\setminus {V }_{i}\}\), for example \({\omega }_{k}\,=\,x\) for all \(k \in {V }_{i}^{2R}\setminus {V }_{i}\), where x is a fixed element of X, and consider the probability distribution \(\mathrm{{P}}_{{V }_{i}}(\cdot \vert {\omega }_{k},k \in {V }_{i}^{2R}\setminus {V }_{i})\) on the finite set \(\Omega ({V }_{i})\) given by (22.1) (with V i instead of V ).

Fix \({V }_{j}\). For \(i > j\), the probability distribution \(\mathrm{{P}}_{{V }_{i}}(\cdot \vert {\omega }_{k},k \in {V }_{i}^{2R}\setminus {V }_{i})\) induces a probability distribution on the set \(\Omega ({V }_{j})\). The space of such probability distributions is tight. (The set \(\Omega ({V }_{j})\) is finite, and we can consider an arbitrary metric on it. The property of tightness does not depend on the particular metric.)

Take a subsequence \(\{{j}_{s}^{(1)}\}\) such that the induced probability distributions on \(\Omega ({V }_{1})\) converge to a limit \(\mathrm{{Q}}^{(1)}\). Then find a subsequence \(\{{j}_{s}^{(2)}\} \subseteq \{ {j}_{s}^{(1)}\}\) such that the induced probability distributions on the space \(\Omega ({V }_{2})\) converge to a limit \(\mathrm{{Q}}^{(2)}\). Since \(\{{j}_{s}^{(2)}\} \subset \{ {j}_{s}^{(1)}\}\), the probability distribution induced by \(\mathrm{{Q}}^{(2)}\) on the space \(\Omega ({V }_{1})\) coincides with \(\mathrm{{Q}}^{(1)}\). Arguing in the same way, we can find a subsequence \(\{{j}_{s}^{(m)}\} \subseteq \{ {j}_{s}^{(m-1)}\}\), for any \(m \geq 1\), such that the probability distributions induced by \(\mathrm{{P}}_{{V }_{j(m)}}(\cdot \vert {\omega }_{k},k\notin {V }_{j(m)})\) on \(\Omega ({V }_{m})\) converge to a limit, which we denote by \(\mathrm{{Q}}^{(m)}\). Since \(\{{j}_{s}^{(m)}\} \subseteq \{ {j}_{s}^{(m-1)}\}\), the probability distribution on \(\Omega ({V }_{m-1})\) induced by \(\mathrm{{Q}}^{(m)}\) coincides with \(\mathrm{{Q}}^{(m-1)}\).

Then, for the sequence of probability distributions

$$\mathrm{{P}}_{{V }_{{ j}_{m}^{(m)}}}(\cdot \vert {\omega }_{k},k \in {V }_{{j}_{m}^{(m)}}^{2R}\setminus {V }_{{ j}_{m}^{(m)}})$$

corresponding to the diagonal subsequence \(\{{j}_{m}^{(m)}\}\), we have the following property:

For each m, the restrictions of the probability distributions to the set \(\Omega ({V }_{m})\) converge to a limit \(\mathrm{{Q}}^{(m)}\), and the probability distribution induced by \(\mathrm{{Q}}^{(m)}\) on \(\Omega ({V }_{m-1})\) coincides with \(\mathrm{{Q}}^{(m-1)}\). The last property is a version of the Consistency Conditions, and by the Kolmogorov Consistency Theorem, there exists a probability distribution \(\mathrm{Q}\) defined on the natural σ-algebra of subsets of the space Ω of all possible configurations \(\{{\omega }_{k},k \in {\mathbb{Z}}^{d}\}\) whose restriction to each \(\Omega ({V }_{m})\) coincides with \(\mathrm{{Q}}^{(m)}\) for any \(m \geq 1\).

It remains to prove that Q is generated by a Gibbs random field corresponding to U. Let V be a finite subset of \({\mathbb{Z}}^{d}\), W a finite subset of \({\mathbb{Z}}^{d}\) such that \({V }^{2R} \subseteq W\), and let the values \({\omega }_{k} = {i}_{k}\) be fixed for \(k \in W\setminus V\). We need to consider the conditional probabilities

$$q =\mathrm{ Q}\{{\omega }_{k} = {i}_{k},k \in V \vert {\omega }_{k} = {i}_{k},k \in W\setminus V \}.$$

In fact, it is more convenient to deal with the ratio of the conditional probabilities corresponding to two different configurations, \({\omega }_{k} =\bar{ {i}_{k}}\) and \({\omega }_{k} =\bar{\bar{ {i}_{k}}}\), \(k \in V\), which is equal to

$$\begin{array}{rcl}{ q}_{1}& =& \frac{\mathrm{Q}\{{\omega }_{k} =\bar{ {i}_{k}},k \in V \vert {\omega }_{k} = {i}_{k},k \in W\setminus V \}} {\mathrm{Q}\{{\omega }_{k} =\bar{\bar{ {i}_{k}}},k \in V \vert {\omega }_{k} = {i}_{k},k \in W\setminus V \}} \\ & =& \frac{\mathrm{Q}\{{\omega }_{k} =\bar{ {i}_{k}},k \in V,{\omega }_{k} = {i}_{k},k \in W\setminus V \}} {\mathrm{Q}\{{\omega }_{k} =\bar{\bar{ {i}_{k}}},k \in V,{\omega }_{k} = {i}_{k},k \in W\setminus V \}}.\end{array}$$
(22.2)

It follows from our construction that the probabilities \(\mathrm{Q}\) in this ratio are the limits found with the help of probability distributions \(\mathrm{{P}}_{{V }_{{ j}_{m}^{(m)}}}(\cdot \vert {\omega }_{k},k \in {V }_{{j}_{m}^{(m)}}^{2R}\setminus {V }_{{ j}_{m}^{(m)}})\). We can express the numerator in (22.2) as follows:

$$\begin{array}{rcl} & & \qquad \qquad \qquad \mathrm{Q}\{{\omega}_{k} =\bar{ {i}_{k}},k \in V,{\omega }_{k} = {i}_{k},k \in W\backslash V \} \\& =& \lim \limits_{m\rightarrow \infty }\mathrm{{P}}_{{V }_{{j}_{m}^{(m)}}}({\omega }_{k} =\bar{ {i}_{k}},k \in V,{\omega }_{k} ={i}_{k},k \in W\setminus V \vert {\omega }_{k} = {i}_{k},k \in {V}_{{j}_{m}^{(m)}}^{2R}\setminus {V }_{{ j}_{m}^{(m)}}) \\ & =& \lim\limits_{m\rightarrow \infty }\frac{\sum \limits_{\{{i}_{k},k\in {V}_{{ j}_{m}^{(m)}}\setminus W\}}\;\exp (-\sum _{k\in {V }_{{j}_{m}^{(m)}}^{R}}U({i}_{k};{i}_{{k}^{^{\prime}}},0 < \vert{k}^{^{\prime}}- k\vert \leq R))} {Z({i}_{k},k \in {V}_{{j}_{m}^{(m)}}^{2R}\setminus {V }_{{j}_{m}^{(m)}})}.\end{array}$$

A similar expression for the denominator in (22.2) is also valid. The difference between the expressions for the numerator and the denominator is that, in the first case, \({i}_{k} =\bar{ {i}_{k}}\) for \(k \in V\), while in the second, \({i}_{k} =\bar{\bar{ {i}_{k}}}\) for kV.

Therefore, the corresponding expressions \(U({i}_{k};{i}_{{k}^{^{\prime}}},\vert {k}^{^{\prime}}- k\vert \leq R)\) for k such that dist \((k,V ) > R\) coincide in both cases, and

$${q}_{1} = \frac{{r}_{1}} {{r}_{2}},$$

where

$${r}_{1} =\exp (-\sum\limits_{k\in {V }^{R}}U({i}_{k};{i}_{k^{\prime}},0 < \vert k^{\prime} - k\vert \leq R)),\ \ \ {i}_{k} =\bar{ {i}_{k}}\ \ \mathrm{for}\ k \in V,$$

while

$${r}_{2} =\exp (-\sum\limits_{k\in {V }^{R}}U({i}_{k};{i}_{k^{\prime}},0 < \vert k^{\prime} - k\vert \leq R)),\ \ \ {i}_{k} =\bar{\bar{ {i}_{k}}}\ \ \mathrm{for}\ k \in V.$$

This is the required expression for q1, which implies that \(\mathrm{Q}\) is a Gibbs field.

2 An Example of a Phase Transition

Theorem 22.3 is an analogue of the theorem on the existence of stationary distributions for finite Markov chains. In the ergodic case, this distribution is unique. In the case of multi-dimensional time, however, under very general conditions there can be different random fields corresponding to the same function U. The related theory is connected to the theory of phase transitions in statistical physics.

If \(X =\{ -1,1\}\), \(R = 1\) and \(U({i}_{0};{i}_{k},\vert k\vert = 1) = \pm \beta {\sum \nolimits }_{\vert k\vert =1}{({i}_{0} - {i}_{k})}^{2}\), the corresponding Gibbs field is called the Ising model with inverse temperature β (and zero magnetic field). The plus sign corresponds to the so-called ferromagnetic model; the minus sign corresponds to the so-called anti-ferromagnetic model. Again, the terminology comes from statistical mechanics. We shall consider here only the case of the ferromagnetic Ising model and prove the following theorem.

Theorem 22.4

Consider the following interaction energy over \({\mathbb{Z}}^{2}\) :

$$U({\omega }_{0};{\omega }_{k},\vert k\vert = 1) = \beta \sum_{\vert k\vert =1}{({\omega }_{0} - {\omega }_{k})}^{2}.$$

If β is sufficiently large, there exist at least two different Gibbs fields corresponding to U.

Proof

As before, we consider the increasing sequence of squares V i and plus-minus boundary conditions, i.e., either \({\omega }_{k} \equiv +1\), \(k\notin {V }_{i}\), or \({\omega }_{k} \equiv -1\), \(k\notin {V }_{i}\). The corresponding probability distributions on \(\Omega ({V }_{i})\) will be denoted by \(\mathrm{{P}}_{{V }_{i}}^{+}\) and \(\mathrm{{P}}_{{V }_{i}}^{-}\) respectively. We shall show that \(\mathrm{{P}}_{{V }_{i}}^{+}({\omega }_{0} = +1)\) \(\geq 1 - \epsilon (\beta )\), \(\mathrm{{P}}_{{V }_{i}}^{-}({\omega }_{0} = -1) \geq 1 - \epsilon (\beta )\), \(\epsilon (\beta ) \rightarrow 0\) as \(\beta \rightarrow \infty \). In other words, the Ising model displays strong memory of the boundary conditions for arbitrarily large i. Sometimes this kind of memory is called the long-range order. It is clear that the limiting distributions constructed with the help of the sequences \(\mathrm{{P}}_{{V }_{i}}^{+}\) and \(\mathrm{{P}}_{{V }_{i}}^{-}\) are different, which constitutes the statement of the theorem.

We shall consider only \(\mathrm{{P}}_{{V }_{i}}^{+}\), since the case of \(\mathrm{{P}}_{{V }_{i}}^{-}\) is similar. We shall show that a typical configuration with respect to \(\mathrm{{P}}_{{V }_{i}}^{+}\) looks like a “sea” of + 1’s surrounding small “islands” of − 1’s, and the probability that the origin belongs to this “sea” tends to 1 as \(\beta \rightarrow \infty \) uniformly in i.

Take an arbitrary configuration \(\omega \in \Omega ({V }_{i})\). For each \(k \in {V }_{i}\) such that \({\omega }_{k} = -1\) we construct a unit square centered at k with sides parallel to the coordinate axes, and then we slightly round off the corners of the square.

The union of these squares with rounded corners is denoted by \(I(\omega )\). The boundary of \(I(\omega )\) is denoted by \(B(\omega )\). It consists of those edges of the squares where \(\omega \) takes different values on different sides of the edge. A connected component of \(B(\omega )\) is called a contour.

It is clear that each contour is a closed non-self-intersecting curve. If \(B(\omega )\) does not have a contour containing the origin inside the domain it bounds, then \({\omega }_{0} = +1\).

Given a contour \(\Gamma \), we shall denote the domain it bounds by \(\mathrm{int}(\Gamma )\). Let a contour \(\Gamma \) be such that the origin is contained inside \(\mathrm{int}(\Gamma )\). The number of such contours of length n does not exceed \(n{3}^{n-1}\).

Indeed, since the origin is inside \(\mathrm{int}(\Gamma )\), the contour \(\Gamma \) intersects the semi-axis \(\{{z}_{1} = 0\} \cap \{ {z}_{2} < 0\}\). Of all the points belonging to \(\Gamma \cap \{ {z}_{1} = 0\}\), let us select that with the smallest z2 coordinate and call it the starting point of the contour. Since the contour has length n, there are no more than n choices for its starting point. Once the starting point of the contour is fixed, the edge of \(\Gamma \) containing it is also fixed. It is the horizontal segment centered at the starting point of the contour. Counting from the segment connected to the right end-point of this edge, there are no more than three choices for each next edge, since the contour is not self-intersecting. Therefore, there are no more than \(n{3}^{n-1}\) contours in total.

Lemma 22.5 (Peierls Inequality)

Let Γ be a closed curve of length n. Then,

$$\mathrm{{P}}_{{V }_{i}}^{+}(\{\omega \in \Omega ({V }_{ i}) : \Gamma \subseteq B(\omega )\}) \leq {e}^{-8\beta n}.$$

We shall prove the Peierls Inequality after completing the proof of Theorem 22.4.

Due to the Peierls Inequality, the probability \(\mathrm{{P}}_{{V }_{i}}^{+}\) that there is at least one contour \(\Gamma \) with the origin inside \(\mathrm{int}(\Gamma )\), is estimated from above by

$$\sum\limits_{n=4}^{\infty }n{3}^{n-1}{e}^{-8\beta n},$$

which tends to zero as \(\beta \rightarrow \infty \). Therefore, the probability of \({\omega }_{0} = -1\) tends to zero as \(\beta \rightarrow \infty \). Note that this convergence is uniform in i.

Proof of the Peierls Inequality. For each configuration \(\omega \in \Omega ({V }_{i})\), we can construct a new configuration \(\omega ^{\prime} \in \Omega ({V }_{i})\), where

$${\omega }_{k}^{\prime} = -{\omega }_{k}\ \ \ \mathrm{if}\ \ k \in \mathrm{ int}(\Gamma ),$$
$${\omega }_{k}^{\prime} = {\omega }_{k}\ \ \ \mathrm{if}\ \ k\notin \mathrm{int}(\Gamma ).$$

For a given Γ, the correspondence \(\omega \leftrightarrow \omega ^{\prime}\) is one-to-one.

Let \(\omega \in \Omega ({V }_{i})\) be such that \(\Gamma \subseteq \mathcal{B}(\omega )\). Consider the ratio

$$\frac{\mathrm{{P}}_{{V }_{i}}^{+}(\omega )} {\mathrm{{P}}_{{V }_{i}}^{+}(\omega ^{\prime})} = \frac{\exp (-\beta {\sum \nolimits }_{k:\mathrm{dist}(k,{V }_{i})\leq 1}{ \sum \nolimits }_{k^{\prime}:\vert k^{\prime}-k\vert =1}{({\omega }_{k} - {\omega }_{k^{\prime}})}^{2})} {\exp (-\beta {\sum \nolimits }_{k:\mathrm{dist}(k,{V }_{i})\leq 1}{ \sum \nolimits }_{k^{\prime}:\vert k^{\prime}-k\vert =1}{({\omega }_{k}^{\prime} - {\omega }_{k^{\prime}}^{\prime})}^{2})}.$$

Note that all the terms in the above ratio cancel out, except those in which k and k′ are adjacent and lie on the opposite sides of the contour Γ. For those terms, \(\vert {\omega }_{k} - {\omega }_{k^{\prime}}\vert = 2\), while \(\vert {\omega }_{k}^{\prime} - {\omega }_{k^{\prime}}^{\prime}\vert = 0\). The number of such terms is equal to 2n (one term for each side of each of the edges of Γ). Therefore,

$$\mathrm{{P}}_{{V }_{i}}^{+}(\omega ) = {e}^{-8\beta n}\mathrm{{P}}_{{ V }_{i}}^{+}(\omega ^{\prime}).$$

By taking the sum over all \(\omega \in \Omega ({V }_{i})\) such that \(\Gamma \subseteq \mathcal{B}(\omega )\), we obtain the statement of the lemma.

One can show that the Gibbs field is unique if β is sufficiently small. The proof of this statement will not be discussed here.

The most difficult problem is to analyze Gibbs fields in neighborhoods of those values \({\beta }_{cr}\) where the number of Gibbs fields changes. This problem is related to the so-called critical percolation problem and to conformal quantum field theory.