1 Introduction

Sequential rationality is a key plausibility requirement in extensive games: It requires that players behave optimally at each point of the game given the other players’ behavior and available information. If information is perfect, requiring equilibrium strategies to be sequentially rational is equivalent to subgame perfection (Selten 1965). If information is imperfect, requiring sequential rationality requires specifying players’ beliefs about previous play at each decision node. On the path of play, Bayes’ rule pins down beliefs. Off the path of play, Bayes’ rule is not well defined, raising the question of which restrictions on beliefs are plausible and/or desirable. Selten (1975) addressed the question by posing the possibility that players may make mistakes with small probabilities. He defined (trembling-hand) perfect equilibria as limits of equilibria of versions of the game perturbed according to a sequence of vanishing trembles. In each perturbed game, Bayes’ rule can always be applied since all instances of the game occur with positive probability on the path of play.

While requiring perfection is a powerful and sensible way to refine the set of Nash equilibria, it is difficult to use in practice. The reason is that obtaining and working with the right tremble sequences and the corresponding sequences of equilibria to prove or disprove that a strategy profile is a perfect equilibrium is often cumbersome and many times not feasible. To address this problem, Kreps and Wilson (1982) proposed a weaker but easier-to-use equilibrium concept, sequential equilibrium, where beliefs are obtained as the limit of fully mixed sequences of strategy profiles, but sequential rationality is only required in the limit.Footnote 1 Still, while using sequential equilibria is indeed simpler than using perfect equilibria, it suffers from a similar problem: proving or disproving that a given assessment is consistent requires proving or disproving that it is not supported by any sequence of strategy profiles, which is often difficult to do. This has led to the use of other, weaker equilibrium concepts, such as different versions of perfect Bayesian equilibrium, first defined in Fudenberg and Tirole (1991).

This paper provides a simple characterization of assessment consistency. We show that an assessment \((\sigma ,\mu )\) is consistent if and only if it is supported by a power sequence of strategy profiles, that is, a sequence of strategy profiles \((\sigma _n)_n\) for which, for each action \(a\), there is a pair \((x(a),y(a))\in \mathbb R_{++}\times {\mathbb {R}}_{+}\) satisfying that \(\sigma _n(a)=x(a)\,n^{-y(a)}\) for all \(n\) when \(y(a)>0\), and \(\sigma _n(a)\rightarrow x(a)\) when \(y(a)=0\).Footnote 2 We argue that since obtaining assessments from power sequences of strategy profiles is straightforward, our result greatly simplifies proving or disproving that a given assessment is consistent.

The proof of the result is involved. To understand its difficulties, take a consistent belief system \((\sigma ,\mu )\), meaning that some sequence of fully-mixed strategy profiles \((\sigma _n)_n\) supports it. The probabilities with which actions are played along the sequence may not be power functions of the index and may differ significantly across actions. The crucial step in the proof is showing that there is a strictly increasing sequence \(({\hat{n}}_n\in {\mathbb {N}})_n\) and a \(K\)-dimensional sequence \((q^1_n,\ldots ,q^K_n)_n\), for some \(K\in {\mathbb {N}}\), with the properties that, for all \(a\in A\),

$$\begin{aligned} \lim _{n\rightarrow \infty }\frac{\sigma _{{\hat{n}}_n}(a)}{\alpha ^0(a)\,\prod \nolimits _{k=1}^K(q^k_{n})^{\alpha ^{k}(a)}} =1 \end{aligned}$$

for unique values \(\alpha (a)\in {\mathbb {R}}^{K+1}\), and that \(\lim _{n\rightarrow \infty }(q^k_{n})^\gamma /q^{k+1}_{n}=0\) for all \(k\) and \(\gamma \in {\mathbb {R}}\). The sequence \((q^1_n,\ldots ,q^K_n)_n\) contains the different types of convergence to 0 across different actions. The value \(\alpha (a)\) can then be used to obtain a pair \((x(a),y(a))\) such that the corresponding power sequence of strategy profiles supports \((\sigma ,\mu )\).

The class of power sequences of strategy profiles is particularly convenient to use. We show that the relative likelihoods of different histories can be straightforwardly obtained from the values that the maps \(x\) and \(y\) assign to their actions. Conversely, the consistency of an assessment is characterized by simple equations involving the values that \(x\) and \(y\) take for the actions in different histories. We also show that any consistent assessment is supported by a power sequence of strategy profiles with integer exponents, further simplifying the analysis of sequential equilibria.

Literature review Since the seminal paper of Kreps and Wilson (1982), belief consistency has been seen as a natural requirement in extensive games. Still, as working with (high-dimensional) sequences of strategy profiles is not feasible in many applications, other equilibrium concepts have been favored, most notably (weak/strong) perfect Bayesian equilibrium (PBE) as defined in Fudenberg and Tirole (1991), where off-path beliefs are updated using Bayes’ rule “whenever possible”. PBE is often too permissive, so ad-hoc restrictions on off-path belief updating are often imposed, such as the “no signaling what you do not know” and “never dissuaded once convinced” conditions (see Osborne and Rubinstein 1994). Alternative restrictions on belief updating off the path of play have been used in Cramton (1985), Rubinstein (1985), Bagwell (1990), and Harrington (1993). The current paper takes a different direction by identifying a simple set of sequences of strategy profiles that is sufficient to characterize consistency. In separate work (Dilmé 2023a), we use the current characterization to provide a full characterization of sequential equilibria in terms of lexicographic numbers.

The rest of the paper is organized as follows. Section 2 provides the notation for extensive form games and consistent assessments. Section 3 states and proves our main result. Section 4 provides techniques to prove or disprove the consistency of an assessment and illustrates them through examples. An “Appendix” contains the omitted proofs.

2 Belief consistency in extensive-form games

We now provide the definitions and corresponding notation for an extensive-form game.

An (finite) extensive-form game \(G\equiv \langle A,H,{\mathcal {I}},N,\pi ,u\rangle\) has the following components. A finite set of actions \(A\). A finite set of histories \(H\) containing finite sequences of actions satisfying that, for all \((a_j)_{j=1}^J\in H\) with \(J>0\), we have \((a_j)_{j=1}^{J-1}\in H\); the set of terminal histories is denoted \(T\). An information partition \({\mathcal {I}}\) of the set of non-terminal histories satisfying that there is a partition \(\{A^I|I\in {\mathcal {I}}\}\) of \(A\) with the property that, for each \(I\in {\mathcal {I}}\) and \(h,h'\in H\), we have (1) \((h,a)\in H\) for some \(a\in A^I\) if and only if \(h\in I\), and (2) if \(h\in I\) and \(h'>h\) then \(h'\notin I\).Footnote 3 A finite set of players \(N\not \ni 0\). A player assignment \(\iota :{\mathcal {I}}\rightarrow N\cup \{0\}\) assigning each information set to a player or to nature. A strategy by nature \(\pi :\cup _{I\in \iota ^{-1}(\{0\})}\,A^I\rightarrow (0,1]\) satisfying \(\sum _{a\in A^{I}}\pi (a)=1\) for each \(I\in \iota ^{-1}(\{0\})\). For each player \(i\in N\), a (von Neumann–Morgenstern) payoff function \(u_i:T\rightarrow {\mathbb {R}}\).

A strategy profile is a map \(\sigma :A\rightarrow [0,1]\) such that \(\sum _{a\in A^I}\sigma (a)=1\) for all \(I\in {\mathcal {I}}\) (i.e., it is a probability distribution for each set of actions available at each information set) and \(\sigma (a)=\pi (a)\) for all \(a\) played by nature (i.e., nature plays according to \(\pi\)). We let \(\Sigma\) be the set of strategy profiles.

2.1 Consistent assessments

As explained in the introduction, requiring that a player behaves sequentially rationally requires assigning beliefs over the information she does not know at the time she makes a decision. Such beliefs are modeled using a belief system, which is a map \(\mu :H\backslash T\rightarrow [0,1]\) satisfying that \(\sum _{h\in I}\mu (h)=1\) for all \(I\in {\mathcal {I}}\). Kreps and Wilson (1982) define an assessment as a pair \((\sigma ,\mu )\), where \(\sigma\) is a strategy profile and \(\mu\) is a belief system. The usual interpretation is that for each information set \(I\), action \(a\in A^I\) and history \(h\in I\), \(\sigma (a)\) is the probability with which player \(\iota (I)\) (i.e., the player taking action at this node of the game) chooses \(a\) while the value \(\mu (h)\) is the posterior belief that this player assigns to \(h\) (both probabilities are conditional on \(I\) being reached).

Kreps and Wilson (1982) pointed out that not all assessments are plausible. Fix, for instance, a strategy profile \(\sigma\) and an on-path information set \(I\). If a rational player assumes that players use \(\sigma\), she should use Bayes’ rule to assign the probability of \(h\in I\) conditional on \(I\) having been reached. If \(\sigma\) is a full-support strategy profile (i.e., \(\sigma (a)>0\) for all \(a\in A\)), Bayes’ rule pins down a unique assessment, where \(\mu\) is given by

$$\begin{aligned} \mu (h|\sigma )=\frac{\sigma (h)}{\sigma (I)} \end{aligned}$$
(2.1)

for all information sets \(I\in {\mathcal {I}}\) and histories \(h\in I\), where abusing notation \(\sigma (\cdot )\) indicates the probability with which a history or an information set occurs under \(\sigma\). Instead, if there is an information set \(I\) that is not reached under \(\sigma\), both the numerator and the denominator of the right side of equation (2.1) are 0, thus the Bayes’ rule cannot be applied.

Selten (1975) proposed studying perturbed versions of the game with “slight mistakes” (in the form of trembles), where all information sets occur on the path of play. Following a similar approach, Kreps and Wilson (1982) introduced the concept of consistent assessment as one that is approximated by a sequence of fully-mixed strategy profiles. Formally, \((\sigma ,\mu )\) is a consistent assessment if it is supported by some fully-mixed sequence \((\sigma _n)_n\), that is, a sequence satisfying \(\sigma _n(a)>0\) and

$$\begin{aligned} (\sigma (a),\mu (h)) =\lim _{n\rightarrow \infty }(\sigma _n(a),\mu (h|\sigma _n)) \end{aligned}$$
(2.2)

for all \(a\in A\) and \(h\in H\backslash T\). Consistency is a crucial requirement of Kreps and Wilson ’s definition of sequential equilibrium, which, additionally, requires that \(\sigma\) is sequentially rational given \(\mu\).Footnote 4

Different reasons make using sequential equilibria desirable. First, consistency provides a systematic and powerful way to discipline off-path beliefs, hence avoiding using ad-hoc requirements often made in perfect Bayesian equilibria (see the Literature Review). Second, Blume and Zame (1994) showed that the set of sequential equilibria coincides with the set of limits of Nash equilibria of perturbed games where both actions and payoffs are perturbed. Hence, they are the predictions of an external observer who cannot perfectly assess the small likelihood of trembles or the exact payoffs of the players. Third, because perfect equilibrium outcomes are sequential equilibrium outcomes (perfection requires exact equilibria along the sequence) and also sequentially stable outcomes (sequential stability requires exact stability for all trembles; see Dilmé 2023b), ruling out that an outcome is sequential permits ruling out its perfection and sequential stability. Additionally, if one can prove that there is a unique sequential outcome, then such outcome is the unique perfect outcome and the unique sequentially stable outcome (and also the unique outcome of a proper equilibrium, see Myerson 1978, and the unique quasi-perfect equilibrium, see van Damme 1984).Footnote 5

3 Main result

3.1 Power sequences of strategy profiles

In this section, we introduce a simple and natural class of sequences of strategy profiles. These are sequences of strategy profiles where actions that are played with vanishing probability decrease according to power functions of the sequence’s index.

Definition 3.1

A \((\sigma _n)_n\) is a power sequence of strategy profiles if there exists a pair of maps \(x:A\rightarrow \mathbb R_{++}\) and \(y:A\rightarrow {\mathbb {R}}_{+}\) such that, for every \(a\in A\) and \(n\in {\mathbb {N}}\) large enough,Footnote 6

$$\begin{aligned} \sigma _n(a)= {\left\{ \begin{array}{ll} x(a)\,n^{-y(a)} &{} \text {if }y(a)>0,\\ M_n(a)\,x(a) &{} \text {if }y(a)=0, \end{array}\right. } \end{aligned}$$
(3.1)

where \(M_n(a)\) is a factor satisfying that \(\lim _{n\rightarrow \infty }M_n(a)=1\).

In a power sequence of strategy profiles, the probability with which an action \(a\) is played along the sequence is characterized by two parameters. The first is a “rate” \(y(a)\in \mathbb R_+\) at which the sequence tends to 0. The action \(a\) is played with probability 0 in the limit if and only if \(y(a)>0\). Also, if \(y(a)>y(a')\) then \(a\) is played with a vanishing probability relative to \(a'\). The second parameter is a multiplying factor \(x(a)\in \mathbb R_{++}\). This factor determines the relative likelihood of actions with the same rate. If \(y(a)=0\), then \(x(a)\) is the asymptotic probability with which action \(a\) is played; hence, it must be that \(\sum _{a\in A^I|y(a)=0}\,x(a)=1\) for all \(I\). Note that if \(n\) is large enough, then \(\sigma _n\) is a fully-mixed strategy profile.

Power sequences of strategy profiles are simple and often used to provide examples of consistent assessments. The reason is that the asymptotic likelihood of each action is determined by only two numbers; hence, obtaining relative likelihoods is easy. As we shall see, even though the set of power sequences of strategy profiles is small relative to the set of all sequences, it is rich enough to support all consistent assessments. Furthermore, we show in Sect. 4 that the limit relative likelihoods of histories can be straightforwardly computed from the values that \(x\) and \(y\) take on their actions. It follows that a power sequence of strategy profiles has a well-defined limit belief system; that is, it supports a consistent assessment.

3.2 Main result

Our main result establishes that any consistent assessment is supported by a power sequence of strategy profiles.

Theorem 3.1

An assessment is consistent if and only if it is supported by a power sequence of strategy profiles.

Theorem 3.1 is proven in Sect. 3.3. It establishes that the set of power sequences of strategy profiles is rich enough to support all consistent assessments. As we will see in Sect. 4, our result simplifies working with consistent assessments and hence eases the use of sequential equilibria.

We now provide an idea of the challenges that proving Theorem 3.1 poses. Since the “if” part of the statement is trivial (assessments supported by power sequences of strategy profiles are consistent), we focus on the “only if” part of the result.

Proving the “only if” part of Theorem 3.1 is involved. To see why, we fix a consistent assessment \((\sigma ,\mu )\) and a sequence \(({{\hat{\sigma }}}_n)_n\) supporting it. Note that the rate at which \({{\hat{\sigma }}}_n(a)\) converges to \(\sigma (a)\) may be very different for different actions \(a\in A\), and not necessarily through a power sequence of the form (3.1). The complication of the proof lies in that each action typically belongs to different histories belonging to different information sets; hence, the choice of \((x(a),y(a))\) simultaneously affects the likelihood of all histories containing \(a\). Guaranteeing that a power sequence of the form (3.1) supports \((\sigma ,\mu )\) involves simultaneously satisfying a potentially large set of interconnected constraints without a clear structure, some of them in the form of inequalities. Our proof provides a method to obtain values for \(\{(x(a),y(a))|a\in A\}\) and some sequence \(({{\hat{n}}}_n)\) satisfying that, for all \(h,h'\in H\),

$$\begin{aligned} \lim _{n\rightarrow \infty } \frac{{{\hat{\sigma }}}_{{{\hat{n}}}_n}(h)}{\hat{\sigma }_{\hat{n}_n}(h')} =\lim _{n\rightarrow \infty } \frac{\prod \nolimits _{j=1}^J\hat{\sigma }_{\hat{n}_n}(a_j)}{\prod \nolimits _{j=1}^{J'}\hat{\sigma }_{\hat{n}_n}(a'_j)} =\lim _{n\rightarrow \infty } \frac{\prod \nolimits _{j=1}^J\sigma _{n}(a_j)}{\prod \nolimits _{j=1}^{J'}\sigma _{n}(a'_j)} = \lim _{n\rightarrow \infty } \frac{\sigma _{n}(h)}{\sigma _{n}(h')}, \end{aligned}$$

where \(\sigma _n\) is a sequence of the form (3.1). We are then able to ensure that the asymptotic relative likelihoods of all pairs of histories under \((\hat{\sigma }_{\hat{n}_n})_n\) and \((\sigma _n)_n\) coincide.

Fig. 1
figure 1

Game tree 1

To shed further light on the problem, consider the game in Fig. 1 (the game in the figure is assumed to continue after each depicted terminal history). Consider the depicted assessment, where the arrows indicate the actions played with positive probability. Consider a fully mixed sequence of strategy profiles \((\hat{\sigma }_n)_n\) such that

$$\begin{aligned} \hat{\sigma }_n(r_1)=\log (n+3)^{-1},\ \ \hat{\sigma }_n(r_2)=e^{-n}+e^{-n^3}\,\sin (n),\ \ \text {and}\ \ \hat{\sigma }_n(r_3)=3\log (n+4)^{-1}\,e^{-n}. \end{aligned}$$
(3.2)

This implies that

$$\begin{aligned}&\alpha =\lim _{n\rightarrow \infty } \frac{(1-\hat{\sigma }_n(r_1))\,(1-\hat{\sigma }_n(r_2))\,\hat{\sigma }_n(r_3)}{(1-\hat{\sigma }_n(r_1))\,(1-\hat{\sigma }_n(r_2))\,\hat{\sigma }_n(r_3) +(1-\hat{\sigma }_n(r_1))\,\hat{\sigma }_n(r_2)\,(1-\hat{\sigma }_n(r_3))}=1 \ \ \text {and}\\&\gamma =\lim _{n\rightarrow \infty } \frac{\hat{\sigma }_n(r_2)\,\hat{\sigma }_n(r_3)}{\hat{\sigma }_n(r_2)\,\hat{\sigma }_n(r_3)+\hat{\sigma }_n(r_1)} =\frac{1}{4}. \end{aligned}$$

Hence, for a power sequence of strategy profiles of the form (3.1) to exist, it must be that

$$\begin{aligned} 0<y(r_3)<y(r_2),\ \ 0<y(r_2)+y(r_3)=y(r_1),\ \ \text {and}\ \ x(r_1) =x(r_2)\,x(r_3)/3. \end{aligned}$$
(3.3)

The first equation says that \(r_3\) is played infinitely more likely than \(r_2\) as \(n\rightarrow 0\), but both probabilities tend to 0. The second equation says that the likelihood of history \((r_1)\) vanishes at the same rate that history \((r_2,r_3)\). The third equation says that the asymptotic relative likelihood of \((r_1)\) relative to \((r_2,r_3)\) is 1/3. While it is easy to find \((x,y):A\rightarrow {\mathbb {R}}_{++}\times {\mathbb {R}}_{+}\) satisfying the constraints in (3.3) in this simple example, it is difficult to obtain a systematic procedure to prove the existence of solutions in general. The proof of Theorem 3.1 provides a constructive procedure that gives a power sequence of strategy profiles supporting any given consistent assessment of an arbitrary extensive-form game.

3.3 Proof of Theorem 3.1

We here present a proof of the “only if” part of the statement of Theorem 3.1, since the “if” part is trivial, as we explained before. A reader interested in the use of Theorem 3.1 to work with consistent assessments can directly move to Sect. 4.

Proof of Theorem 3.1

We begin with a useful lemma, proven in the “Appendix”.

Lemma 3.1

(Representation) Let \(\hat{A}=\{a_1,\ldots ,a_J\}\) be a finite set, and \((\sigma _n:\hat{A}\rightarrow (0,1])_n\) be a sequence. There are a strictly increasing sequence \(({\hat{n}}_n\in \mathbb N)_n\) and a sequence \(((q^1_n,\ldots ,q^K_n)\in \mathbb R_{++}^K)_n\), for some \(K\in \{0,\ldots ,J\}\), such that

  1. 1.

    \(\lim _{n\rightarrow \infty } q^1_{{\hat{n}}_n}=0\) and \(\lim _{n\rightarrow \infty }q^{k}_{{\hat{n}}_n}/(q^{k-1}_{{\hat{n}}_n})^{\gamma }=0\) for all \(\gamma \in {\mathbb {R}}\) and \(k=2,\ldots ,K\).

  2. 2.

    For each \(a\in A\) there is a unique \(\alpha (a)\equiv (\alpha ^1(a),\ldots ,\alpha ^K(a))\in \mathbb R^{K}\) such that

    $$\begin{aligned} \lim _{n\rightarrow \infty }\frac{\sigma _{{\hat{n}}_n}(a)}{\prod \nolimits _{k=1}^{K} (q_{{\hat{n}}_n}^k)^{\alpha ^k(a)}}\in \mathbb R_{++}. \end{aligned}$$
  3. 3.

    There are two sequences \((\hat{\jmath }_k)_{k=1}^K\) with \(1\le \hat{\jmath }_1<\cdots <\hat{\jmath }_{K}\le J\) and \((\hat{k}_k)_{k=1}^K\) with \(\hat{k}_k\ne \hat{k}_{k'}\) for all \(k,k'\), such that, for all \(k\in \{1,\ldots ,K\}\), we have \((1)\,\alpha ^{\hat{k}_{k'}}(\hat{a}_j)=0\) for all \(k'>k\) and \(j<\hat{\jmath }_k\), and \((2)\,\alpha ^{\hat{k}_k}(\hat{a}_{\hat{\jmath }_k})\ne 0\).

Lemma 3.1 shows that any finite collection of sequences on (0, 1] can be represented as the product of the powers of a basis of sequences. Its proof constructively provides such sequences by proceeding iteratively over the set of actions. In each iteration j, we expand the basis of sequences only if (an appropriate subsequence of) \(\sigma _n(a_j)\) tends to 0 at a rate that cannot be expressed as the product of powers of the current basis.Footnote 7 The case \(K=0\) corresponds to the case where \(({\hat{n}}_n)_n\) is such that \(\lim _{n\rightarrow \infty }\sigma _{{\hat{n}}_n}(a_j)>0\) for all \(j=1,\ldots ,J\).

Fix a consistent assessment \((\sigma ,\mu )\) supported by some sequence \((\sigma _n)_n\). Let \(({\hat{n}}_n)_n\) and \((q^1_n,\ldots ,q^K_n)_n\) be two sequences with the properties in the statement of Lemma 3.1 applied to \((\sigma _n)_n\) and \(A\). We then have that, for each action \(a\in A\), there exist unique \(\alpha ^0(a)\in {\mathbb {R}}_{++}\) and \(\alpha (a)\in \mathbb R^K\) such that

$$\begin{aligned} \lim _{n\rightarrow \infty }\frac{\sigma _{{\hat{n}}_n}(a)}{\alpha ^0(a)\,\prod \nolimits _{k=1}^{K} (q_{{\hat{n}}_n}^k)^{\alpha ^k(a)}}=1. \end{aligned}$$

Note that \(\sigma (a)>0\) if and only if \(\alpha ^k(a)=0\) for all \(k=1,\ldots ,K\), in which case \(\sigma (a)=\alpha ^0(a)\). Similarly, it follows that, for each history \(h=(a_1,\ldots ,a_J)\in H\), we have

$$\begin{aligned} \lim _{n\rightarrow \infty }\frac{\sigma _{{\hat{n}}_n}(h)}{(\prod \nolimits _{j=1}^J\alpha ^0(a_j))\,\prod \nolimits _{k=1}^{K} (q_{{\hat{n}}_n}^k)^{\alpha ^k(h)}}=1, \end{aligned}$$

where \(\alpha (h):=\sum _{j=1}^J\alpha (a_j)\).

We define the reversed lexicographic order \(\succ\) on \(\mathbb R^K\) as follows. For each pair \(\alpha (h),\alpha (h')\in {\mathbb {R}}^K\), \(\alpha (h)\succ \alpha (h')\) indicates that there is some \(k\in \{1,\ldots ,K\}\) such that \(\alpha ^{k'}(h)=\alpha ^{k'}(h')\) for all \(k'=k+1,\ldots ,K\) and \(\alpha ^k(h)>\alpha ^k(h')\). By the first property of the statement of Lemma 3.1, we have that if \(\alpha (h)\succ \alpha (h')\) then \(\lim _{n\rightarrow \infty }\sigma _{{\hat{n}}_n}(h)/\sigma _{{\hat{n}}_n}(h')=0\).Footnote 8 Our goal is to use \(\alpha\) and \(\alpha ^0\) to construct the maps \(x\) and \(y\) such that the power sequence of strategy profiles in Eq. (3.1) supports \((\sigma ,\mu )\).

For each \(M\in {\mathbb {R}}_{++}\), we define

$$\begin{aligned} (q_{M,n}^1,\ldots ,q_{M,n}^K) :=\big ((q^1_n)^{1/M},(q^2_n)^{1/M^2}, \ldots , (q^K_n)^{1/M^K}\big ) \end{aligned}$$

and

$$\begin{aligned} (\alpha _M^1(a),\alpha _M^2(a),\ldots ,\alpha _M^K(a)) :=\big (M^1\,\alpha ^1(a), M^2\,\alpha ^1(a),\ldots , M^K\,\alpha ^K(a)\big ), \end{aligned}$$

and note that \(\prod _{k=1}^{K} (q_{M,n}^k)^{\alpha ^k_M(a)}=\prod _{k=1}^{K} (q_{n}^k)^{\alpha ^k(a)}\). We also define the pair of maps \((x_M,y_M):A\rightarrow {\mathbb {R}}_{++}\times \mathbb R_{+}\) assigning, to each action \(a\in A\), the value

$$\begin{aligned} (x_M(a),y_M(a)) :=\left( \textstyle {\alpha ^0(a), \sum \limits _{k=1}^K\alpha _{M}^k(a) }\right) . \end{aligned}$$
(3.4)

It is then clear that, if M is large enough, then \(y_M(a)>y_M(a')\) if and only if \(\alpha (a)\succ \alpha (a')\), and also that for all histories \(h,h'\in H\),

$$\begin{aligned} \lim _{n\rightarrow \infty } \frac{\sigma _{{\hat{n}}_n}(h)}{\sigma _{{\hat{n}}_n}(h')}={\left\{ \begin{array}{ll} x_M(h)/x_M(h') &{} \text {if }y_M(h)=y_M(h'),\\ 0 &{} \text {if }y_M(h)<y_M(h'),\\ \infty &{} \text {if }y_M(h)>y_M(h'). \end{array}\right. } \end{aligned}$$
(3.5)

In particular, if M is large enough, we have \(\sigma (a)=0\) if and only if \(y_M(a)=0\), and therefore \(\sum _{a\in A^I|y(a)=0}x_M(a)=1\) (recall that \(\sigma (a)=\alpha ^0(a)\) when \(\sigma (a)>0\)). In conclusion, if M is large enough, then \((\sigma ,\mu )\) is supported by a power sequence of strategy profiles satisfying Eq. (3.1) with \((x,y):=(x_M,y_M)\), so the proof of Theorem 3.1 is complete. \(\square\)

4 Considerations

In this section, we illustrate how Theorem 3.1 can be used to obtain consistent assessments easily. Section 4.1 introduces a result that is useful in finding consistent assessments. Section 4.2 shows that the exponents in a power sequence of strategy profiles can be chosen to be integers. Section 4.3 examines alternative classes of sequences of strategy profiles with similar properties. Section 4.4 provides some examples illustrating how our results can be used to find consistent assessments.

4.1 Constructing consistent assessments

We begin with a useful corollary of Theorem 3.1.

Corollary 4.1

Fix a sequence of power strategy profiles \((\sigma _n)_n\) as in Eq. (3.1) for some pair \((x,y)\). Then, for any history \(h\equiv (a_j)_{j=1}^J\in H\),

$$\begin{aligned} \lim _{n\rightarrow \infty }\overbrace{\frac{\sigma _n(h)}{ (\prod \limits _{j=1}^Jx(a_j))\,n^{-\sum \nolimits _{j=1}^Jy(a_j)}}}^{=:(*)}=1. \end{aligned}$$
(4.1)

Corollary 4.1 establishes that the asymptotic likelihood of a history \(h\) can be easily obtained from the values that \(x\) and \(y\) take on its actions. This is clear if \(h\) is only composed of actions with vanishing probabilities; in this case, the term \((*)\) in (4.1) is exactly 1 for all \(n\). Otherwise, each action with a non-vanishing probability is typically not a power function of \(n\), but the sum of polynomial functions of \(n\). These additional terms typically make the expressions of the probability of a history \(h\) under \(\sigma _n\) long and complicated. Still, given that \(\lim _{n\rightarrow \infty }M_n(a)=1\) (i.e., \(\sigma _n(a)\rightarrow x(a)\)), the belief system can be computed “as if” the probabilities of all actions were of the form \(x(a)\,n^{-y(a)}\), including those with \(y(a)=0\).

Generating consistent assessments is easy using Corollary 4.1: First, choose a pair of maps \((x,y):A\rightarrow {\mathbb {R}}_{++}\times {\mathbb {R}}_{+}\) with the condition that \(\sum _{a\in A^I|y(a)=0}\,x(a)=1\) for all \(I\). For each action \(a\) define \(\sigma (a):=x(a)\) if \(y(a)=0\) and \(\sigma (a):=0\) otherwise. Next, for each history \(h\equiv (a_j)_{j=1}^J\), compute

$$\begin{aligned} {\textstyle x(h):=\prod \limits _{j=1}^Jx(a_j)\ \ \text {and}\ \ y(h):=\sum \limits _{j=1}^Jy(a_j).} \end{aligned}$$
(4.2)

Then, for each information set \(I\), only histories \(h\in I\) with the lowest value \(y(h)\) receive a positive probability under \(\mu\), and the relative probability of two such histories \(h,h'\in I\) is \(x(h)/x(h')\).

Verifying that a given assessment \((\sigma ,\mu )\) is consistent (or proving it is not) is more difficult but can often be done using Corollary 4.1 as follows. First, set \(x(a):=\sigma (a)\) and \(y(a):=0\) for all \(a\in A\) such that \(\sigma (a)>0\). Second, for the rest of the actions, look for values \(x(a),y(a)>0\) so that for all histories \(h,h',h''\in H\) in the same information set with \(\mu (h),\mu (h')>\mu (h'')=0\) we have \(y(h)=y(h')<y(h'')\) and \(x(h)/x(h')=\mu (h)/\mu (h')\).

An important observation is that obtaining the values \(\{x(a)|a\in A\}\) and \(\{y(a)|a\in A\}\) can be done separately. Indeed, to obtain maps \(x\) and \(y\) supporting a given assessment, the first step consists in obtaining \(y\), which determines the relative rates at which the probability of different histories vanish. The second step consists in finding \(x\) so that the implied relative probabilities of histories with a positive probability coincide with those implied by the belief system.Footnote 9 For example, it is easy to see that if a consistent assessment \((\sigma ,\mu )\) is such that \(\mu (h)\in \{0,1\}\) for all \(h\in H\backslash T\), then there is a power sequence of strategy profiles with \(x(a)=1\) for all actions satisfying that \(\sigma (a)=0\). The examples below illustrate this procedure.

Finally, we state that power sequences of strategy profiles can be rescaled while supporting the same assessment.

Proposition 4.1

Fix \((x,y):A\rightarrow {\mathbb {R}}_{++}\times {\mathbb {R}}_{+}\) such that \(\sum _{a\in A^{I}|y(a)=0}\,x(a)=1\) for all \(I\in {\mathcal {I}}\). Let \(K_1,K_2\in {\mathbb {R}}_{++}\) be two constants. Define

$$\begin{aligned} ({{\tilde{x}}}(a),{{\tilde{y}}}(a)) :=(K_2^{y(a)}\,x(a) ,K_1\,y(a)) \ \ \ \hbox { for all}\ a\in A. \end{aligned}$$

Then, the power sequences of strategy profiles defined by \((x,y)\) and \(({{\tilde{x}}},{{\tilde{y}}})\) support the same assessment.

The intuition for Proposition 4.1 is the following. First, note that two histories \(h\equiv (a_j)_{j=1}^J\) and \(h'\equiv (a'_{j'})_{j'=1}^{J'}\) satisfy \({{\tilde{y}}}(h)<{{\tilde{y}}}(h')\) if and only if they also satisfy \(y(h)<y(h')\). Also, if \({{\tilde{y}}}(h)={{\tilde{y}}}(h')\) then \(y(h)=y(h')\), and

$$\begin{aligned} \frac{{{\tilde{x}}}(h)}{{{\tilde{x}}}(h')}= \frac{\prod \nolimits _{j=1}^{J}{{\tilde{x}}}(a_j)}{\prod \nolimits _{j'=1}^{J'}{{\tilde{x}}}(a'_{j'})} =\frac{K_2^{y(h)}\,\prod \nolimits _{j=1}^{J}x(a_j)}{K_2^{y(h')}\,\prod \nolimits _{j'=1}^{J'}x(a'_{j'})} =\frac{\prod \nolimits _{j=1}^{J}x(a_j)}{\prod \nolimits _{j'=1}^{J'}x(a'_{j'})} =\frac{x(h)}{x(h')}. \end{aligned}$$

Hence, the parameters associated with actions played with vanishing probabilities can be rescaled while supporting the same assessment (notice that \(({{\tilde{x}}}(a),{{\tilde{y}}}(a))=(x(a),y(a))\) when \(y(a)=0\)).

4.2 Integer exponents

Our definition of power sequences of strategy profiles (Definition 3.1) allows the rate of convergence of each action’s probability to be any non-negative real number. In practice, nevertheless, integer exponents are often used in examples. In this section, we show that all assessments are supported by power sequences of strategy profiles with integer exponents; hence, the assumption of integer exponents can be made without loss of generality.

Definition 4.1

We say that \((\sigma _n)_n\) is an integer power sequence of strategy profiles if it can be written as in (3.1) for some functions \(x:A\rightarrow \mathbb R_{++}\) and \(y:A\rightarrow {\mathbb {Z}}_+\).

The class of integer power sequences of strategy profiles is significantly smaller than the class of power sequences of strategy profiles. Still, as the following proposition states, it is large enough to support all consistent assessments.

Proposition 4.2

An assessment is consistent if and only if some integer power sequence of strategy profiles supports it.

Proposition 4.2 establishes that any consistent assessment is supported by an integer power sequence of strategy profiles; that is, the rates of convergence to 0 of different actions can be chosen to be non-negative integers. There are various reasons for doing so. One is pedagogical, as integers make it easier to categorize different convergence rates. Another is a potential technical advantage since looking for integer solutions to systems of equations is often numerically easier. In many cases, nevertheless, not restricting to integer exponents is easier as it gives more flexibility to find or construct consistent assessments.

Remark 4.1

Our characterization of sequential equilibria using monomials with integer powers is reminiscent of Blume and Zame (1994)’s characterization of sequential equilibria using algebraic geometry. They show that the graph of the equilibrium correspondence for sequential equilibria is a semi-algebraic set; that is, it is defined by finite systems of polynomial inequalities (with integer powers). While we would find a thorough exploration of the connection between these two approaches interesting, it lies outside the boundaries of the current paper’s focus.

4.3 Other convenient sequences

Power sequences of strategy profiles are easy and convenient to use. They are determined by two maps (\(x\) and \(y\) in Definition 3.1 with the only condition that \(\sum _{a\in A^{I}|y(a)=0}\,x(a)=1\)), and they permit easily computing relative likelihoods of histories (recall Corollary 4.1). In this section, we present other classes of sequences of strategy profiles that are convenient to use.

We begin by noticing three desirable properties of the class of power sequences of strategy profiles. First, a single parameter determines the relative limit likelihood of actions: \(\lim _{n\rightarrow \infty }\sigma _n(a)/\sigma _n(a')=0\) if and only if \(y(a)>y(a')\). Second, another parameter determines the relative probability of the actions with the same convergence rate: \(\lim _{n\rightarrow \infty }\sigma _n(a)/\sigma _n(a')=x(a)/x(a')\) whenever \(y(a)=y(a')\). Finally, the procedure to obtain the rate of convergence and relative probability for histories from their actions involves simple operations (addition and multiplication, see Eq. (4.2)).

We now show that other classes of sequences of strategy profiles share similar properties. We begin noticing that, for any function \(f:{\mathbb {N}}\rightarrow {\mathbb {R}}_{++}\) satisfying that \(\lim _{n\rightarrow \infty }f(n)=0\) and any consistent assessment \((\sigma ,\mu )\), there exists a pair \((x,y):A\rightarrow \mathbb R_{++}\times {\mathbb {R}}_{+}\) such that \((\sigma ,\mu )\) is supported by

$$\begin{aligned} \sigma _n(a):= {\left\{ \begin{array}{ll} x(a)\,f(n)^{-y(a)} &{} \text {if }y(a)>0,\\ M_n(a)\,x(a) &{} \text {if }y(a)=0, \end{array}\right. } \end{aligned}$$
(4.3)

where \(M_n(a)\) is a factor satisfying that \(\lim _{n\rightarrow \infty }M_n(a)=1\). The convergence rates of histories can be computed using Eq. (4.2). So, the properties of a class of sequences of strategy profiles are the same as the properties of power sequences of strategy profiles, so they support the same assessment (for a given pair \((x,y)\)) independently of the value of f.

A slightly different formulation is the following. It is easy to see that, for each \(f:{\mathbb {N}}\rightarrow {\mathbb {R}}_+\) such that \(\lim _{n\rightarrow \infty }f(n)=+\infty\) and for each consistent assessment \((\sigma ,\mu )\), there is a pair of maps \((\hat{x},\hat{y}):A\rightarrow {\mathbb {R}}_{++}\times (0,1]\) (note that the co-domain of \(\hat{y}\) is different from before) such that \((\sigma ,\mu )\) is supported by a sequence \((\hat{\sigma }_n)_n\) satisfying that, if \(n\) is large enough,

$$\begin{aligned} \hat{\sigma }_n(a)= {\left\{ \begin{array}{ll} \hat{x}(a)\,\hat{y}(a)^{-f(n)} &{} \text {if }\hat{y}(a)>0,\\ {\hat{M}}_n(a)\,\hat{x}(a) &{} \text {if } \hat{y}(a)=1, \end{array}\right. } \end{aligned}$$
(4.4)

where again \({\hat{M}}_n(a)\) is a factor satisfying that \(\lim _{n\rightarrow \infty }{\hat{M}}_n(a)=1\). In this case, an action is played with vanishing probability if and only if \(\hat{y}(a)<1\). Now, the converge rate of a history \(h\equiv (a_j)_{j=1}^J\) is given by

$$\begin{aligned} \textstyle {\hat{x}(h) :=\prod \limits _{j=1}^J\hat{x}(a_j)}\ \ \text {and}\ \ \textstyle {\hat{y}(h) :=\prod \limits _{j=1}^J\hat{y}(a_j)} \end{aligned}$$

instead of by Eq. (4.2). In this case, it is easy to see that the assessment supported by \((\hat{\sigma }_n)_n\) for \((\hat{x},\hat{y})\) is the same as the assessment supported by \((\sigma _n)_n\) satisfying (3.1) with \((x,y):=(\hat{x},-\log (\hat{y}))\).

4.4 Examples

Fig. 2
figure 2

Game tree 2

Example 4.1

We first illustrate how consistent assessments can be easily generated using power sequences of strategy profiles. Consider Fig. 2, which represents part of a tree of a game where all information sets belong to different players (the game is assumed to continue after each depicted terminal history). Assign, for example, \((x(r_1),y(r_1)):=(1,4)\), \((x(r_2),y(r_2)):=(2,2)\), \((x(r_3),y(r_3)):=(2,1)\) and \((x(r_4),y(r_4)):=(1,1)\) (note that, necessarily, all other actions get assigned (1, 0)). For example, note that

$$\begin{aligned}&(x(r_1),y(r_1))=(1,4)\ \ \text {and}\\&(x(l_1,r_2,r_3,r_4), y(l_1,r_2,r_3,r_4))=(1\cdot 2\cdot 2\cdot 1,0+2+1+1)=(4,4). \end{aligned}$$

Hence, since \(y(r_1)=y(l_1,r_2,r_3,r_4)=4\), we have that \((1-\gamma )/\gamma =x(r_1)/x(l_1,r_2,r_3,r_4)=1/4\). It is then easy to see that this implies that \(\alpha =1\), \(\beta =1/3\), and \(\gamma =4/5\); hence, such assessment is consistent.

Example 4.2

The example illustrates how a power sequence of strategy profiles can be obtained from a given assessment. Consider Fig. 2 again. Assume that, under an assessment, all \(\{r_j\}_{j=1}^4\) are played with probability 0 (hence the rest of the actions are played with probability 1) and \(\alpha ,\beta ,\gamma \in (0,1)\). Set \(x(r_2):=1\) and \(y(r_2):=1\) (note that, by Proposition 4.1, the values of \(x\) and \(y\) assigned to one of the actions played with probability zero can be chosen freely). That \(\alpha \in (0,1)\) implies \(y(r_3)=1\) and \(x(r_3)=\alpha /(1-\alpha )\). Simiarly, that \(\beta \in (0,1)\) implies \(y(r_4)=1\) and \(x(r_4)=\alpha \,\beta /((1-\alpha )\,(1-\beta ))\). Finally, that \(\gamma \in (0,1)\) implies \(y(r_1)=3\) and \(x(r_1)=\alpha ^2\,\beta \,(1-\gamma )/((1-\alpha )^2\,(1-\beta )\,\gamma )\).

Example 4.3

Fig. 3 reproduces the game depicted in Figure 235.1 in Osborne and Rubinstein (1994) (for graphical clarity, we omit the continuation game after the dashed information sets). In their Example 234.3, they argue through a lengthy argument that the assessment where only actions with an arrow are played with positive probability and with the belief system given in the picture is not consistent (their argument further relies on their assumption that \(i_1\), \(i_2\), and \(i_3\) are played with the same probability).

Fig. 3
figure 3

Game tree 3

We now argue that their result follows trivially using Theorem 3.1 (without the need to assume that \(i_1\), \(i_2\), and \(i_3\) are played with the same probability). If the assessment in Fig. 3 would be consistent, then there would exist functions \((x,y):A\rightarrow \mathbb R_{++}\times {\mathbb {R}}_{+}\) satisfying

$$\begin{aligned} y(c_2)+y(l_2)<\min \{y(c_1),y(l_3)\} \ \ \text {and}\ \ y(c_1)+y(m_1) <\min \{y(c_2),y(m_3)\}. \end{aligned}$$

The first equation requires that the exponent of \(n\) for the history \((i_2,c_2,l_2)\) is strictly smaller (in absolute value) than those of the other histories in the same information set (note that \(y(i_2)=0\)), while the second equation requires that the exponent of \(n\) for the history \((i_1,c_1,m_1)\) is strictly lower than those of the other histories of the same information set. The contradiction is clear: the left equation requires \(y(c_2)<y(c_1)\), while the second one requires \(y(c_1)<y(c_2)\).