1 Introduction

Despite being inconsistent with subjective expected utility, the uncertainty averse pattern of preferences suggested by Ellsberg (1961) has been defended normatively (Gilboa et al. 2012; Gilboa 2014) and descriptively (Camerer and Weber 1992; Fox and Tversky 1995; Halevy 2007). Many new theories have emerged to accommodate such behaviour (Gilboa and Marinacci 2013). An important class of uncertainty averse preferences is the variational preferences of Maccheroni et al. (2006), MMR hereforth. Variational preferences generalise the maxmin expected utility (MEU) model of Gilboa and Schmeidler (1989). The main theorem of this paper characterises a fully subjective version of variational preferences. The special case of purely subjective MEU preferences is also characterised.

In this paper, we consider preferences over acts with two stages of subjective uncertainty. As with Savage (1954), we employ a rich set of states and allow the outcome set to be arbitrary. Ellsberg-type preferences, in our case, refers to being more uncertainty averse towards first-stage uncertainty than second-stage uncertainty. That is, we assume source dependence of uncertainty attitudes (Heath and Tversky 1991; Fox and Tversky 1995; Chow and Sarin 2001; Wakker 2001; Abdellaoui et al. 2011), also called issue preference (Ergin and Gul 2009; Strzalecki 2011). Indeed, we will assume that, while second-stage uncertainty can be subjective, the decision maker displays no particular aversion to such uncertainty. The two-stage framework is used to develop a subjective notion of averaging events and averaging acts. Ellsberg’s example corresponds to a decision maker who prefers first-stage uncertainty to be averaged out. We call such a preference second-stage uncertainty aversion. It will be shown that, when combined with other known axioms, second-stage uncertainty aversion characterises purely subjective variational preferences.

To the best of my knowledge, there has been no purely subjective foundation for variational preferences. For MEU, there has been more success, although a foundation at the full level of generality of Savage (1954) remains elusive. Chateauneuf (1991) assumed a single stage of subjective uncertainty and derived MEU with linear utility for final outcomes. MMR’s axioms could be similarly reinterpreted. Recent derivations of MEU have used a single stage of purely subjective uncertainty and assumed topological conditions on the final outcome set (Casadesus-Masanell et al. 2000; Ghirardato et al. 2003; Alon and Schmeidler 2014). A particular benefit of this approach is that, unlike the approach taken in this paper, it allows for finite state spaces. The topological richness of the outcome set facilitates the derivation of cardinal utility and separation of decision weights from utility, yielding a biseparable representation (Ghirardato and Marinacci (2001)). Variational preferences are not necessarily biseparable; hence, such approaches have not yielded new foundations for variational preferences. This paper advances the literature by characterising purely subjective variational preferences in a different, special case of the Savage (1954) framework.

2 Ellsberg and multiple stages of uncertainty

The acts considered in Ellsberg (1961) can be naturally viewed as having two separate stages of uncertainty (Segal 1987; Klibanoff et al. 2005; Nau 2006; Halevy 2007; Chew and Sagi 2008; Ergin and Gul 2009; Seo 2009; Machina 2011). A simple version of Ellsberg’s three-colour problem imagines a ball will be chosen from an urn containing three numbered balls. Ball 1 is red, and the remaining balls are either yellow or black. In the first stage of uncertainty, the composition of the urn, a colour for each remaining ball, is determined. In the second stage, the number of the ball is determined. Hence, the outcome of choices in this problem is determined by the realisation of two states, s and t, where \(s \in \{YY, YB, BY, BB \}\) and \(t\in \{1,2,3\}\). Consider the following choices:

Choice One.

$$\begin{aligned}&f \text{ pays } \pounds 100 \text{ only } \text{ if } \text{ red }. \quad v's \quad g \text{ pays } \pounds 100 \text{ only } \text{ if } \text{ yellow }.\\&\quad f = \left\{ \begin{array}{ll} \pounds 100 &{} \text{ if } t = 1 \\ \pounds 0 &{} \text{ if } t \ne 1 \\ \end{array}\right. \quad v's \quad g = \left\{ \begin{array}{ll} \left\{ \begin{array}{ll} \pounds 0 &{} \text{ if } t = 1 \\ \pounds 100 &{} \text{ if } t \ne 1 \\ \end{array}\right. &{} \text{ if } s=YY \\ \left\{ \begin{array}{ll} \pounds 0 &{} \text{ if } t \in \{1,3\} \\ \pounds 100 &{} \text{ if } t \notin \{1,3\} \\ \end{array}\right. &{} \text{ if } s = YB \\ \left\{ \begin{array}{ll} \pounds 0 &{} \text{ if } t \in \{1,2\} \\ \pounds 100 &{} \text{ if } t \notin \{1,2\} \\ \end{array}\right. &{} \text{ if } s =BY \\ \pounds 0 &{} \text{ if } s= BB \\ \end{array}\right. \end{aligned}$$

Choice Two.

$$\begin{aligned}&{\tilde{f}} \text{ pays } \pounds 100 \text{ only } \text{ if } \text{ a } \text{ red } \text{ or } \text{ black }. \quad v's \quad \tilde{g} \text{ pays } \pounds 100 \text{ only } \text{ if } \text{ a } \text{ yellow } \text{ or } \text{ black }. \\&\quad \tilde{f} = \left\{ \begin{array}{ll} \left\{ \begin{array}{ll} \pounds 100 &{} \text{ if } t = 1 \\ \pounds 0 &{} \text{ if } t\ne 1 \\ \end{array}\right. &{} \text{ if } s=YY \\ \left\{ \begin{array}{ll} \pounds 100 &{} \text{ if } t \in \{1,3\} \\ \pounds 0 &{} \text{ if } t \notin \{1,3\} \\ \end{array}\right. &{} \text{ if } s=YB \\ \left\{ \begin{array}{ll} \pounds 100 &{} \text{ if } t \in \{1,2\} \\ \pounds 0 &{} \text{ if } t \notin \{1,2\} \\ \end{array}\right. &{} \text{ if } s=BY \\ \pounds 100 &{} \text{ if } s=BB \end{array}\right. \quad v's \quad \tilde{g} = \left\{ \begin{array}{ll} \pounds 0 &{} \text{ if } t = 1 \\ \pounds 100 &{} \text{ if } t \ne 1 \\ \end{array}\right. \quad \quad \quad \end{aligned}$$

The modal pattern of preferences seems to be f preferred to g and \(\tilde{g}\) preferred to \(\tilde{f}\). The above presentation makes it clear that the commonly preferred acts, f and \(\tilde{g}\), have no first-stage uncertainty. For some reason, the acts with uncertainty about the second-stage uncertainty are disliked. In the following section, the two-stage framework is formalised so that, in Sect. 6, this aversion to uncertainty about second-stage uncertainty can be precisely defined. When combined with known axioms, this will provide the key axiom for variational preferences in this framework.

3 The two-stage framework

Let \({\mathcal {S}}\) denote a finite set of first-stage states with \(\sigma \)-algebra \({\mathcal {E}}_{\mathcal {S}}\). Let \({\mathcal {T}}\) denote a set of second-stage states with \(\sigma \)-algebra \({\mathcal {E}}_{{\mathcal {T}}}\). Let \({\mathcal {X}}\) denote a finite set of outcomes.Footnote 1 We impose richness on the second-stage state space \({\mathcal {T}}\) in what follows via a structural condition (second-stage solvability). Let \({\mathcal {E}}_{{\mathcal {S}}}\otimes {\mathcal {E}}_{{\mathcal {T}}} = \sigma \{A\times B{:} A\in {\mathcal {E}}_{{\mathcal {S}}}, B\in {\mathcal {E}}_{{\mathcal {T}}}\}\). An act is a function \(f{:}{\mathcal {S}}\times {\mathcal {T}}\rightarrow {\mathcal {X}}\) that is measurable with respect to \({\mathcal {E}}_{{\mathcal {S}}}\otimes {\mathcal {E}}_{{\mathcal {T}}}\). The set of acts is \({\mathcal {A}}\) and contains all such functions. The decision maker is modelled using a preference relation \(\succcurlyeq \), a binary relation defined over \({\mathcal {A}}\).

If \(A\in {\mathcal {E}}_{{\mathcal {S}}}\) is a first-stage event, then \(f_A g\) refers to the act such that \(f_A g (s,\cdot ) = f(s,\cdot )\) if \(s\in A\) and \(f_A g (s,\cdot )= g(s,\cdot )\) otherwise. If \(B\in {\mathcal {E}}_{{\mathcal {T}}}\) is a second-stage event, then \(f_B g\) refers to the act such that \(f_B g (\cdot , s) = f(\cdot , t)\) if \(t\in B\) and \(f_B g (\cdot , t) = g(\cdot , t)\) otherwise. An act f is constant if there is \(x\in {\mathcal {X}}\) such that, for all \(s\in {\mathcal {S}}\) and all \(t\in {\mathcal {T}}\), \(f(s,t)=x\). We abuse notation and write x to mean the constant act yielding x. An event \(A\in {\mathcal {E}}_i\), \(i\in \{{\mathcal {S}},{\mathcal {T}}\}\), is null if \(f_A h \sim g_A h\) for all \(f,g,h\in {\mathcal {A}}\), otherwise it is nonnull.

Given an act \(f{:}{\mathcal {S}}\times {\mathcal {T}}\rightarrow {\mathcal {X}}\) and first-stage state \(s\in {\mathcal {S}}\), we define the second-stage act \(f_s{:}{\mathcal {T}}\rightarrow {\mathcal {X}}\) such that \(f_s(t)=f(s,t)\) for all \(t\in {\mathcal {T}}\). An act f is second-stage constant if \(f_s\) is constant for all \(s\in {\mathcal {S}}\). An act f is first-stage constant if, for all \(s,\tilde{s}\in {\mathcal {S}}\), the second-stage acts \(f_s\) and \(f_{\tilde{s}}\) coincide. Preferences \(\succcurlyeq \) over first-stage constant acts naturally induce preferences over second-stage acts, that we also denote \(\succcurlyeq \), as follows: \(f_s\succcurlyeq g_s\) if and only if \(h\succcurlyeq \tilde{h}\), where h and \(\tilde{h}\) are first-stage constant acts with \(h_s = f_s\) and \(\tilde{h}_s=g_s\) for all \(s\in {\mathcal {S}}\).

Subjective expected utility evaluates acts as follows:

$$\begin{aligned} f \mapsto \int _{{\mathcal {S}}} \int _{{\mathcal {T}}}u(f(s,t))\mathrm{d}p(s,t) \end{aligned}$$

where p is a joint probability measure over \({\mathcal {E}}_{{\mathcal {S}}}\otimes {\mathcal {E}}_{{\mathcal {T}}}\) and \(u{:}{\mathcal {X}}\rightarrow {\mathbb {R}}\) is a strictly \(\succcurlyeq \)-increasing utility function.

Denote by \({\mathcal {P}}_{{\mathcal {S}}}\) the set of all finitely additive probability measures over the first-stage state space \({\mathcal {S}}\). A function \(c{:}{\mathcal {P}}_{{\mathcal {S}}}\rightarrow [0,\infty ]\) is grounded if its infimum value is zero. Preferences \(\succcurlyeq \) have a subjective variational representation if acts are evaluated as follows:

$$\begin{aligned} f \mapsto \min _{p\in {\mathcal {P}}_{{\mathcal {S}}}}\bigg ( \int _{{\mathcal {S}}} \int _{{\mathcal {T}}} u(f(s,t))\mathrm{d}p(s)\mathrm{d}q(t) + c(p) \bigg ) \end{aligned}$$

where q is a probability measure over second-stage events \({\mathcal {E}}_{{\mathcal {T}}}\), \(u{:}{\mathcal {X}}\rightarrow {\mathbb {R}}\) is a strictly \(\succcurlyeq \)-increasing utility function and \(c{:}{\mathcal {P}}_{{\mathcal {S}}}\rightarrow [0,\infty ]\) is a grounded, convex and lower semicontinuous function. The function c is called the ambiguity index. In this paper, the probability measure q is convex valued: for all \(\alpha \in [0, 1]\), there exists \(A\in {\mathcal {E}}_{{\mathcal {T}}}\) such that \(q(A)=\alpha \).

4 Variational preferences in the AA framework

The Anscombe–Aumann (AA) framework is the special case of our framework where \({\mathcal {T}}\) is already equipped with a convex-valued probability measure \(q{:}{\mathcal {E}}_{{\mathcal {T}}}\rightarrow [0,1]\). Then, second-stage acts induce, via the probability measure q on \({\mathcal {T}}\), lotteries on \({\mathcal {X}}\). Let \(\varDelta ({\mathcal {X}})\) be the set of lotteries over \({\mathcal {X}}\). Suppose the act f takes values in \(\{x_1,\ldots ,x_n\}\). Then, the act f can be identified with a function \(F{:}{\mathcal {S}}\rightarrow \varDelta ({\mathcal {X}})\) such that:

$$\begin{aligned} F(s) = (q(f_s^{-1}(x_1)), x_1; \ldots ; q(f_s^{-1}(x_n)), x_n) \quad \text{ for } \text{ all } s\in {\mathcal {S}}. \end{aligned}$$

We call F an AA act, the set of which is \({\mathcal {A}}^*\). Preferences \(\succcurlyeq \) over \({\mathcal {A}}\) naturally induce preferences over \({\mathcal {A}}^*\). For all \(F,G\in {\mathcal {A}}^*\), and \(\alpha \in [0,1]\), define \(\alpha F + (1-\alpha )G\) as the AA act that yields the well-defined lottery \(\alpha F(s) + (1-\alpha )G(s)\), for all \(s\in {\mathcal {S}}\). The set of AA acts \({\mathcal {A}}^*\) can then be taken to be a convex set. The key axioms for variational preferences, weak certainty independence and uncertainty aversion, and the continuity axiom used, all make explicit use of the convexity of the set of acts.

Axiom MMR1 (Weak ordering) Preferences \(\succcurlyeq \) over \({\mathcal {A}}^*\) are complete and transitive.

Axiom MMR2 (Weak certainty independence) For all \(F,G\in {\mathcal {A}}^*\), lotteries \(P,Q\in \varDelta ({\mathcal {X}})\), and \(\alpha \in (0,1)\): \(\alpha F + (1-\alpha )P \succcurlyeq \alpha G + (1-\alpha )P\) only if \(\alpha F + (1-\alpha )Q \succcurlyeq \alpha G + (1-\alpha )Q\).

Axiom MMR3 (Mixture continuity) For all \(F,G,H\in {\mathcal {A}}^*\), \(\{ \alpha \in [0,1]: \alpha F + (1-\alpha )G \succcurlyeq H \}\) and \(\{ \alpha \in [0,1]: H\succcurlyeq \alpha F + (1-\alpha )G \}\) are closed subsets of the [0, 1] interval.

Axiom MMR4 (Monotonicity) \(F(s)\succcurlyeq G(s)\) for all \(s\in {\mathcal {S}}\) only if \(F\succcurlyeq G\).

Axiom MMR5 (Uncertainty aversion) For all \(F,G\in {\mathcal {A}}^*\) and \(\alpha \in (0,1)\): \(F \sim G\) only if \(\alpha F + (1-\alpha )G \succcurlyeq F\).

Axiom MMR6 (Nondegeneracy) \(F\succ G\) for some \(F,G\in {\mathcal {A}}^*\).

Preferences have a variational representation in the AA framework if there exists a nonconstant affine function \(U{:}\varDelta ({\mathcal {X}})\rightarrow {\mathbb {R}}\) and a grounded, convex and lower semicontinuous function \(c{:}{\mathcal {P}}_{{\mathcal {S}}}\rightarrow [0,\infty ]\) such that, for all \(F,G\in {\mathcal {A}}^*\):

$$\begin{aligned} F \succcurlyeq G \quad \Leftrightarrow \quad \min _{p\in {\mathcal {P}}_{{\mathcal {S}}}}\bigg ( \int _{{\mathcal {S}}} U(F)dp + c(p) \bigg ) \geqslant \min _{p\in {\mathcal {P}}_{{\mathcal {S}}}}\bigg ( \int _{{\mathcal {S}}} U(G)dp + c(p) \bigg ). \end{aligned}$$

The following theorem characterises variational preferences in the AA framework:

Theorem 1

(Maccheroni et al. 2006) Preferences \(\succcurlyeq \) over Anscombe–Aumann acts \({\mathcal {A}}^*\) satisfy axioms MMR1-6 (weak ordering, weak certainty independence, mixture continuity, monotonicity, uncertainty aversion and nondegeneracy) if and only if they admit a variational representation. In such a representation, utility U is cardinal and the ambiguity index c is a ratio scale.

5 Second-stage subjective expected utility

Consider the following axioms for preferences \(\succcurlyeq \) over \({\mathcal {A}}\):

Axiom 1 (Weak ordering) Preferences \(\succcurlyeq \) over \({\mathcal {A}}\) are complete and transitive.

Axiom 2 (Monotonicity) \(f(s,t)\succcurlyeq g(s,t)\) for all \(s\in {\mathcal {S}}\) and \(t\in {\mathcal {T}}\) only if \(f\succcurlyeq g\).

Axiom 3 (Nondegeneracy) There exist constant acts such that \(x\succ y\).

Axiom 4 (Weak comparative probability) For all events \(A,B\in {\mathcal {E}}_{{\mathcal {T}}}\) and outcomes \(x,\tilde{x},y,\tilde{y}\in {\mathcal {X}}\), with \(x\succ y\) and \(\tilde{x}\succ \tilde{y}\), \(x_A y\succcurlyeq x_B y\) only if \(\tilde{x}_A \tilde{y} \succcurlyeq \tilde{x}_B \tilde{y}\).

Axiom 5 (Archimedeanity) If \(S = A^1, A^2, \ldots ,\) is a sequence of nonnull second-stage events such that \(x_{A^1} f \succ x_{A^1} g\) and \(x_{A^i} f \sim x_{A^{i+1}}g\) for all \(i=1,2,\ldots ,\) then S is finite.

Apart from the fact that we have adapted these axioms to two-stage uncertainty, axioms 1–5 are well known. Rather than use a second-stage version of Savage’s P6, small event continuity, we use an Archimedean axiom and a solvability assumption, as follows:

Structural assumption (Second-stage solvability) For all \(f\in {\mathcal {A}}\) and \(x,y\in {\mathcal {X}}\), with \(x \succcurlyeq f \succcurlyeq y\), there exists \(A\in {\mathcal {E}}_{{\mathcal {T}}}\) such that \(f \sim {x}_A y\).

In this way, nonnecessary conditions are separated from the axiom set. Subjective variational preferences necessarily satisfy axioms 1–5. The following axiom, the second-stage sure-thing principle plays a similar role here to that played by the weak certainty independence axiom in the AA framework. Strzalecki (2011) first introduced this axiom.Footnote 2

Axiom 6 (Second-stage sure-thing principle) For all acts \(f,g\in {\mathcal {A}}\), first-stage constant acts \(h,\tilde{h}\in {\mathcal {A}}\), and second-stage events \(E\in {\mathcal {E}}_{{\mathcal {T}}}{:} f_E h \succcurlyeq g_E h\) only if \(f_E \tilde{h} \succcurlyeq g_E \tilde{h}\).

For \(c\in {\mathbb {R}}\), let \({\mathbf {c}}=(c,\ldots ,c)\in {\mathbb {R}}^n\). A function \(\phi {:}{\mathbb {R}}^n\rightarrow {\mathbb {R}}\) is vertically invariant if for all \(x\in {\mathbb {R}}^n\), \(\alpha \in [0,1]\), \(\phi (\alpha x + (1-\alpha ){\mathbf {c}})=\phi (\alpha x) + (1-\alpha )c\). For finite \({\mathcal {S}}=\{s_1,\ldots ,s_n\}\), preferences are represented by invariant second-stage expected utility if acts are evaluated as follows:

$$\begin{aligned} f \mapsto \phi \bigg ( \int _{{\mathcal {T}}} u(f(s_1,t))dq(t), \ldots ,\int _{{\mathcal {T}}} u(f(s_n,t))dq(t) \bigg ) \end{aligned}$$

where \(\phi {:}{\mathbb {R}}^n\rightarrow {\mathbb {R}}\) is monotonic, mixture continuous and vertically invariant, q is a convex-valued probability measure over \({\mathcal {E}}_{{\mathcal {T}}}\) and \(u{:}{\mathcal {X}}\rightarrow {\mathbb {R}}\) is a strictly \(\succcurlyeq \)-increasing utility function. This class of preferences encompasses subjective variational preferences. Axioms 1–6 characterise invariant second-stage expected utility:

Theorem 2

Assume second-stage solvability. Then, preferences \(\succcurlyeq \) over \({\mathcal {A}}\) satisfy axioms 1–6 (weak ordering, monotonicity, nondegeneracy, weak comparative probability, Archimedeanity, the second-stage sure-thing principle) if and only if they admit an invariant second-stage expected utility representation. In such a representation, utility u is cardinal and the probability measure q is unique.

6 Averaging events and acts

The uncertainty aversion axiom captures a preference for hedging subjective uncertainty. If \(f\sim g\) holds, the mixture \(\frac{1}{2}f+\frac{1}{2}g\) smooths out the subjective uncertainty of each act. Here, we construct a similar idea using only subjective uncertainty. The approach taken here is admittedly less straightforward. We will use the notion of a second-stage event average.

A second-stage event average of two events A and C will, intuitively, correspond to an event B that evenly mixes the probabilities of A and C. To visualise the idea, consider the following example. Imagine a container of different coloured rice (or similar fine material). There is red, yellow and possibly other coloured rice. Bets will be placed on the colour of a grain selected from the mixed bucket. Event A is “red” and event C is “yellow”. Suppose that:

$$\begin{aligned} \left\{ \begin{array}{ll} \pounds 100 &{} \text{ if } s\in A \\ \pounds 0 &{} \text{ if } s\notin A \\ \end{array}\right. \succ \left\{ \begin{array}{ll} \pounds 100 &{} \text{ if } s\in C \\ \pounds 0 &{} \text{ if } s\notin C\\ \end{array}\right. \end{aligned}$$

The interpretation is that the decision believes there are more red grains in the mix than yellow. Abdellaoui and Wakker (2005) refer to A being revealed as more likely than C in a basic sense. Now suppose we can scoop out a subset E of red grains and add the same amount of yellow grains back. Then, “red” is the event \(A\setminus E\) and “yellow” is the event \(C\cup E\). Suppose that, perhaps by iterating the procedure, we elicit the following indifference:

$$\begin{aligned} \left\{ \begin{array}{ll} \pounds 100 &{} \text{ if } s\in A\setminus E \\ \pounds 0 &{} \text{ if } s\notin A \setminus E \\ \end{array}\right. \sim \left\{ \begin{array}{ll} \pounds 100 &{} \text{ if } s\in C\cup E \\ \pounds 0 &{} \text{ if } s\notin C\cup E\\ \end{array}\right. \end{aligned}$$

Then, \(A\setminus E\) is revealed equally likely to \(C\cup E\), in a basic sense. We call any event B that is also revealed equally likely to \(A\setminus E\) and \(C\cup E\) an event average of A and C. This a subjective definition of averaging events. Applying this idea in the two-stage framework leads to the following:

Definition (Second-stage event average) Given outcomes \(x,y\in {\mathcal {X}}\) with \(x\succ y\) and second-stage events \(A,B,C\in {\mathcal {E}}_{{\mathcal {T}}}\), B is a second-stage event average of A and C if \(x_A y \succcurlyeq x_C y\) and there exists \(E\subset A\setminus C\) such that \(x_B y\sim x_{A \setminus E} y \sim x_{C \cup E} y\).

We also call B a second-stage event average of A and C if it is a second-stage event average of C and A. The following lemma further justifies this terminology:

Lemma 3

For subjective variational preferences, \(B\in {\mathcal {E}}_{{\mathcal {T}}}\) is a second-stage event average of A and B only if \(q(B)=\frac{1}{2}q(A)+\frac{1}{2}q(C)\).

To see this, suppose that B a second-stage event average of A and C and that preferences admit a subjective variational preferences representation. Then, \(x_B y \sim x_{A \setminus E} y\) holds if and only if:

$$\begin{aligned} q(B)u(x) + (1-q(B))u(y) = q(A {\setminus } E)u(x) + (1-q(A \setminus E))u(y), \end{aligned}$$

and \(x_B y \sim x_{C \cup E} y\) holds if and only if,

$$\begin{aligned} q(B)u(x) + (1-q(B))u(y) = q(C \cup E)u(x) + (1-q(C \cup E))u(y). \end{aligned}$$

Given \(u(x)>u(y)\) and \(E\subset A{\setminus } C\), these are jointly equivalent to:

$$\begin{aligned} q(B) = q(A \setminus E) = q(C \cup E) \Leftrightarrow q(B) = q(A) - q(E) = q(C) + q(E) \\ \Leftrightarrow q(A) - q(B) = q(B) - q(C) \Leftrightarrow q(B)=\frac{1}{2}q(A)+\frac{1}{2}q(C). \end{aligned}$$

Using the notion of a second-stage event average, a notion of averaging acts can be constructed. For two acts f and g, in each first-stage state s, replace the second-stage acts \(f_s\) and \(g_s\) with equivalent binary acts \(x_{A(s)} y\) and \(x_{C(s)}y\), then construct an act with second-stage acts \((x_{B(s)}y)_s\) such that each B(s) is a second-stage event average of A(s) and C(s):

Definition (Second-stage act average) Given acts \(f,g,h\in {\mathcal {A}}\) and outcomes \(x,y\in {\mathcal {X}}\) with \(x\succ y\) and \(x\succcurlyeq j \succcurlyeq y\) for \(j=f,g,h\) and, for all \(s\in {\mathcal {S}}\):

$$\begin{aligned} f_s \sim (x_{A(s)} y)_s \quad \text{ and } \quad g_s \sim (x_{B(s)} y)_s \quad \text{ and } \quad h_s \sim (x_{C(s)} y)_s, \end{aligned}$$

where \(A(s),B(s),C(s)\in {\mathcal {E}}_{{\mathcal {T}}}\), g is a second-stage act average of f and h if, for all \(s\in {\mathcal {S}}\), B(s) is a second-stage event average of A(s) and C(s).

This definition of averaging acts is purely subjective. It would be problematic if it depended on the particular x and y outcomes, but our axiom set rules out such a dependence. Implementing the definition could, depending on the acts under consideration require the elicitation a large number of indifferences. The following lemma justifies the term second-stage act average:

Lemma 4

For subjective variational preferences, g is a second-stage act average of f and h if and only if, for all \(s\in {\mathcal {S}}\),

$$\begin{aligned} \int _{{\mathcal {T}}} u(g(s,t))\mathrm{d}q(t) = \frac{1}{2}\int _{{\mathcal {T}}} u(f(s,t))\mathrm{d}q(t) + \frac{1}{2}\int _{{\mathcal {T}}} u(h(s,t))\mathrm{d}q(t). \end{aligned}$$

The following axiom uses the above notion of second-stage act averages to define a subjective version of the uncertainty aversion axiom:

Axiom 7 (Second-stage uncertainty aversion) For all acts \(f,g,h\in {\mathcal {A}}\), with g a second-stage act average of f and h, \(f\sim h\) only if \(g\succcurlyeq f\).

Let us reconsider the Ellsberg three-colour problem, interpreted in the two-stage framework. There are three numbered balls in an urn. Ball 1 is red and balls 2 an 3 are known only to be yellow or black. The first stage of uncertainty concerns the unknown colours \(\{YY, YB, BY, BB \}\), and in the second stage, a number is chosen from \(\{1, 2, 3\}\). Let \(x\succ y\). The act g that pays x only if a yellow ball is drawn is typically considered worse than the act f that pays x only if a red ball is drawn. Intuitively, one of the three balls is red, while it is possible that none of the other balls are yellow. The same reasoning suggests that f would be preferred to an act \(g^*\) that pays x only if a black ball is drawn. Further, appealing to the symmetry of their descriptions \(g\sim g^*\) is reasonable, and we assume this holds:

$$\begin{aligned} g = \left\{ \begin{array}{ll} \left\{ \begin{array}{ll} x &{} \text{ if } s_2 \in \{2,3\} \\ y &{} \text{ if } s_2 = 1 \\ \end{array}\right. &{} \text{ if } s_1=YY \\ \left\{ \begin{array}{ll} x &{} \text{ if } s_2 =2 \\ y &{} \text{ if } s_2 \in \{1,3\} \\ \end{array}\right. &{} \text{ if } s_1 = YB \\ \left\{ \begin{array}{ll} x &{} \text{ if } s_2 = 3 \\ y &{} \text{ if } s_2 \in \{1,2\} \\ \end{array}\right. &{} \text{ if } s_1 =BY \\ y \phantom {\begin{array}{l} x \\ y \end{array}}&{} \text{ if } s_1= BB \\ \end{array}\right. \sim g^* = \left\{ \begin{array}{ll} y \phantom {\begin{array}{l} x \\ y \end{array}}&{} \text{ if } s_1=YY \\ \left\{ \begin{array}{ll} x &{} \text{ if } s_2 =3 \\ y &{} \text{ if } s_2 \in \{1,2\} \\ \end{array}\right. &{} \text{ if } s_1 = YB \\ \left\{ \begin{array}{ll} x &{} \text{ if } s_2 =2 \\ y &{} \text{ if } s_2 \in \{1,3\} \\ \end{array}\right. &{} \text{ if } s_1 =BY \\ \left\{ \begin{array}{ll} x &{} \text{ if } s_2 \in \{2,3\} \\ y &{} \text{ if } s_2 = 1 \\ \end{array}\right. &{} \text{ if } s_1= BB \\ \end{array}\right. \end{aligned}$$

We appeal to symmetry again and assume:

$$\begin{aligned} x_{\{1\}}y \sim x_{\{2\}}y \sim x_{\{3\}}y \quad \text{ and } \quad x_{\{1,2\}}y \sim x_{\{1,3\}}y \sim x_{\{2,3\}}y. \end{aligned}$$

Hence, \(\{1\}\) is a second-stage event average of \(\{2,3\}\) and \(\emptyset \). Also, \(\{1\}\) is a second-stage event average of \(\{2\}\) and \(\{3\}\). It follows that \(x_{\{1\}}y = f\) is a second-stage act average of g and \(g^*\); therefore, \(f\succ g\) is consistent with second-stage uncertainty aversion. A similar argument can be used to argue that \(\tilde{g}\succ \tilde{f}\) is simultaneously consistent with second-stage uncertainty aversion. Second-stage uncertainty averse preferences are consistent with Ellberg’s example. We now state the main theorem of this paper:

Theorem 5

Assume second-stage solvability. Then, preferences \(\succcurlyeq \) over \({\mathcal {A}}\) satisfy axioms 1–7 (weak ordering, monotonicity, nondegeneracy, weak comparative probability, Archimedeanity, the second-stage sure-thing principle and second-stage uncertainty aversion) if and only if they admit a subjective variational representation. In such a representation, utility u is cardinal, the ambiguity index c is a ratio scale, and the probability measure q is unique.

Finally, consider the following restriction of axiom 7:

Axiom 7 \(^{*}\) (Second-stage uncertainty neutrality) For all acts \(f,g,h\in {\mathcal {A}}\), with g a second-stage act average of f and h, \(f\sim h\) only if \(g\sim f\).

We have the following:

Theorem 6

If, in statement 1 of Theorem, axiom A7 (second-stage uncertainty aversion) is replaced with axiom A7\(^*\) (second-stage uncertainty neutrality), then subjective expected utility holds.

7 Second-order risk aversion

Strzalecki (2011), following Ergin and Gul (2009), referred to the following axiom as second-order risk aversion:Footnote 3

Axiom 7 \(^{**}\) (Second-order risk aversion) For all second-stage constant acts \(f,g\in {\mathcal {A}}\), and second-stage events \(A\in {\mathcal {E}}_{{\mathcal {T}}}\), \(f\sim g\) only if \(f_A g\succcurlyeq f\).

Second-order risk aversion considers acts f and g that may have first-stage uncertainty, but no second-stage uncertainty. Considering the act \(f_A g\), second-stage uncertainty is introduced, but the first-stage uncertainty is somewhat smoothed, with the overall change being preferable. Second-order risk aversion seems a more elegant axiom than second-stage uncertainty aversion. The following theorem implies that if \({\mathcal {X}}\) is suitably solvable, then axiom 7 can be replaced with axiom 7\(^{**}\) in Theorem 5:

Theorem 7

Assume preferences \(\succcurlyeq \) satisfy axioms 1–6 (weak ordering, monotonicity, nondegeneracy, weak comparative probability, Archimedeanity, and the second-stage sure-thing principle), second-stage solvability holds, and all acts have certainty equivalents. Then, axiom 7 (second-stage uncertainty aversion) holds if and only if axiom 7\(^{**}\) (second-order risk aversion) holds.

Preferences \(\succcurlyeq \) have a second-order expected utility representation (Ergin and Gul 2009) if acts are evaluated as follows:

$$\begin{aligned} f \mapsto \int _{{\mathcal {S}}} \psi \bigg ( \int _{{\mathcal {T}}} u(f(s,t))\mathrm{d}q(t) \bigg )\mathrm{d}p(s) \end{aligned}$$

where p is a probability measure over first-stage events \({\mathcal {E}}_{{\mathcal {S}}}\), q is a probability measure over second-stage events \({\mathcal {E}}_{{\mathcal {T}}}\), \(u{:}{\mathcal {X}}\rightarrow {\mathbb {R}}\) is a strictly \(\succcurlyeq \)-increasing utility function and \(\psi {:}{\mathbb {R}}\rightarrow {\mathbb {R}}\) is strictly increasing and continuous. Theorem 2 of Strzalecki (2011) gives the equivalence of concave \(\psi \) and second-order risk aversion. Our theorem 7 shows the equivalence of concavity of the \(\phi \) obtained in Theorem 2 and second-order risk aversion.

8 Maxmin expected utility

Preferences \(\succcurlyeq \) have a subjective maxmin expected utility (MEU) representation if acts are evaluated as follows:

$$\begin{aligned} f \mapsto \min _{p\in C}\bigg (\int _{{\mathcal {S}}} \int _{{\mathcal {T}}} u(f(s,t))\mathrm{d}p(s)\mathrm{d}q(t)\bigg ) \end{aligned}$$

where q is a convex-valued probability measure over second-stage events \({\mathcal {E}}_{{\mathcal {T}}}\), \(u{:}{\mathcal {X}}\rightarrow {\mathbb {R}}\) is a strictly \(\succcurlyeq \)-increasing utility function and \(C\subseteq {\mathcal {P}}_{{\mathcal {S}}}\) is a convex set of probability measures. They are a special case of variational preferences, seen by taking c that is zero on C and infinite outside of C. Hence, the axioms of Theorem 5 are necessary but not sufficient for MEU. An additional axiom is required, as follows:

Axiom 8 (Second-stage constant independence) For all acts \(f,\tilde{f},g,\tilde{g}\in {\mathcal {A}}\), and outcome \(x\in {\mathcal {X}}\), with \(\tilde{f}\) a second-stage act average of f and x and \(\tilde{g}\) a second-stage act average of g and x, \(f\sim g\) only if \(\tilde{f}\sim \tilde{g}\).

We now state the following foundation for purely subjective maxmin expected utility:

Theorem 8

Assume second-stage solvability. Then, preferences \(\succcurlyeq \) over \({\mathcal {A}}\) satisfy axioms 1–8 (weak ordering, monotonicity, nondegeneracy, weak comparative probability, Archimedeanity, the second-stage sure-thing principle, second-stage uncertainty aversion and second-stage constant independence) if and only if they admit a subjective maxmin expected utility representation. In such a representation, utility u is cardinal, the set of priors C is unique, and the probability measure q is unique.

9 Discussion of frameworks of uncertainty

The main results of this paper have been to characterise existing ambiguity models in a framework different to their original axiomatisations. This section discusses the frameworks commonly employed in modelling choice under uncertainty, to justify and clarify the contribution of this paper. Choice under uncertainty is now a vast field with various frameworks employed. Let us restrict attention to frameworks that employ, as primitives, a set of states \({\mathcal {S}}\), with \(\sigma \)-algebra of events \({\mathcal {E}}_{{\mathcal {S}}}\), and a set of outcomes \({\mathcal {X}}\). States and outcomes are primitives in the sense that, once these objects are understood, all other constructions can be made only in terms of these objects. For example, an act is a function from \({\mathcal {S}}\) to \({\mathcal {X}}\), the set of acts is \({\mathcal {A}}\subseteq {\mathcal {S}}^{{\mathcal {X}}}\), and a preference relation is a subset \(\succcurlyeq \subseteq {\mathcal {S}}^{{\mathcal {X}}}\times {\mathcal {S}}^{{\mathcal {X}}}\). The following frameworks are well known:

The savage framework Acts are measurable functions from \({\mathcal {S}}\) to \({\mathcal {X}}\). \({\mathcal {S}}\) has a rich structure (solvability). \({\mathcal {X}}\) is an arbitrary set.

The Wakker framework Acts are measurable functions from \({\mathcal {S}}\) to \({\mathcal {X}}\). \({\mathcal {S}}\) is an arbitrary set. \({\mathcal {X}}\) has a rich structure (topologically connectedness or algebraic solvability).

The Anscombe–Aumann framework Acts are measurable functions from \({\mathcal {S}}\) to \(\varDelta ({\mathcal {X}})\). \({\mathcal {S}}\) is an arbitrary set. \({\mathcal {X}}\) is an arbitrary set. \(\varDelta ({\mathcal {X}})\) is the set of lotteries over \({\mathcal {X}}\).

The trade-off between the Savage and Wakker frameworks is quite clear. The Savage framework retains complete flexibility with the outcome set \({\mathcal {X}}\); hence, theories developed in that framework are applicable for monetary outcomes, health outcomes, nondivisible goods, and so on. This price of this flexibility with \({\mathcal {X}}\) is the requirement that \({\mathcal {S}}\) has a rich structure.Footnote 4 Conversely, in the Wakker framework, the richness is bourne by \({\mathcal {X}}\), allowing for complete flexibility of the state space. Hence, the Wakker framework allows for monetary outcomes \(({\mathcal {X}}={\mathbb {R}})\) and finitely many states, highly relevant for economic theory. Of course, requiring richness of \({\mathcal {X}}\) rules out applications to nondivisible goods, and so on.

In a sense, the Anscombe–Aumann (AA) framework retains the nicest properties of both the Savage and Wakker frameworks. The AA framework allows both \({\mathcal {S}}\) and \({\mathcal {X}}\) to be arbitrary. However, the axiomatic derivation of expected utility, with cardinal utility and a unique subjective probability measure, requires additional structure. The AA framework’s additional structure comes in two parts:

  1. 1.

    There are two stages of uncertainty.

  2. 2.

    The second stage of uncertainty is characterised by objective probabilities.

The additional structure imposed is very convenient.Footnote 5 Axioms and proofs developed in the AA framework are more transparent. Clearly, however, this additional structure has some costs involved. Fishburn (1970:167) refers to the AA framework as one that employs extraneous probabilities. Ther term “extraneous” seems, at least to this author, to be carefully chosen. If one is describing the assumption that some objective probabilities are assumed, the more usual term would be “exogenous”. Fishburn’s use of the term “extraneous” seems to refer to both the fact the AA framework takes probabilites as exogenously given and the opinion that such objects are “surplus to requirement”. Regarding the use of two stages, while there are many occasions where multiple sources of uncertainty are present, it is a restrictive assumption. Trautmann and Wakker (2015) and others have criticised the use of backward induction implicit in the two-stage evaluation commonly used in the AA framework. Inspired in part by these concerns, further frameworks have been developed. Sarin and Wakker (1997), for example, cleverly avoided the use of two-stage acts. Instead of uncertainty followed by risk, they assume uncertainty alongside risk; hence, acts are either uncertain or are risky lotteries.

The Sarin and Wakker framework Acts are either measurable functions from \({\mathcal {S}}\) to \({\mathcal {X}}\), or elements of \(\varDelta ({\mathcal {X}})\). \({\mathcal {S}}\) is an arbitrary set. \({\mathcal {X}}\) is an arbitrary set.

Ergin and Gul (2009), and this paper, retain the two-stage feature of the AA framework, and all of its criticisms, but dispense with the requirement that the second stage of uncertainty is quantified by objective probabilities:

The Ergin and Gul framework Acts are measurable functions from \({\mathcal {S}}\times {\mathcal {T}}\) to \({\mathcal {X}}\). \({\mathcal {S}}\) and \({\mathcal {T}}\) have rich structures (solvability). \({\mathcal {X}}\) is an arbitrary set.

The specific two-stage formulation of the Ellsberg paradox used in Sect. 2 is an example of the Ergin and Gul framework. In this paper, we have assumed that both stages are subject to subjective uncertainty, but only the first stage is subject to ambiguity. That is, subjective uncertainty that cannot necessarily be quantified by a subjective probability measure. Uncertainty aversion here refers to the dislike of subjective uncertainty that the decision maker finds too “ambiguous” to confidently assign any unique probability. The distinction between first and second stage is made a priori, rather than revealed by preferences. This limitation has been discussed by Klibanoff, Marinacci and Mukerji (2005: 1872–1875). A decision maker, when presented with the Ellsberg choices, may well entertain any belief about the composition, have second-order beliefs about such beliefs, have beliefs about the experiment in general, and so on. We have imposed a very specific set of assumptions. In certain applications, the distinction can be quite clear, in particular if the decision maker has certain competences. For example, one might reasonably expect that an expert political commentator is more confident in his beliefs about the next US presidential election than his beliefs about the outcome of the Kentucky Derby. Of course, if the decision maker is a horse racing pundit, the converse is more reasonable. This connection between competence and ambiguity is supported by experimental evidence (Heath and Tversky 1991; Fox and Tversky 1995). Nevertheless, this remains a limitation of the approach used here.

There is an ever-present trade-off involved in choosing a framework to model uncertainty. In terms of applicability, realism and simplicity, no one framework seems to be dominant in all aspects. Worse still, whichever framework is chosen, there are always assumed objects that simply must be regarded as hypothetical. For example, both the existence of constant acts for all outcomes, and the assumption that preferences are defined over all conceivable acts, even those counterfactual to any plausible reality, are highly questionable. One might claim that if some aspects of each framework are necessarily hypothetical, then surely it does not matter if a hypothetical randomisation device is also used. Given this, it is reasonable to ask, at this point, does the choice of framework really matter? After all, subjective expected utility can be derived in all of the above frameworks, and this involves constructing objects with the same properties as a randomisation device. When we know it is not really necessary, the use of a hypothetical randomisation device that greatly simplifies the axiomatic work seems to be quite acceptable. For ambiguity models, further reassurance has been provided by Ghirardato et al. (2003), who show how many results developed with a hypothetical randomisation device can be recovered in the Wakker framework. However, their approach, based on a weak biseparability assumption, does not apply to variational preferences. Although a hypothetical randomisation device can be suitably defended, it would be concerning if a model simply could not be derived without such a device. This paper has addressed this concern for the variational preferences model.

10 Closing comments

This paper characterises variational preferences in a purely subjective framework. The key assumption was that, as in the Anscombe and Aumann (1963) approach, there are two stages of uncertainty. Even without the linear structure delivered by a randomisation device, multiple stages of uncertainty seem to be useful for utility measurement.Footnote 6

Savage’s expected utility axioms were extended to this framework, and axioms introduced: the second-stage sure-thing principle, second-stage uncertainty aversion, and second-stage constant independence. It is not claimed here that this paper fully solves the open problem reported by Alon and Schmeidler (2014: 397). Progress is made here only by imposing additional conditions on the Savage (1954) framework.