1 Introduction

Abreu et al. [1, 2] (henceforth, APS) provide set-valued dynamic programming techniques for solving repeated games. They give conditions such that the set of sequential equilibrium payoffs in a repeated game is a fixed point of an operator analogous to the Bellman operator in dynamic programming. This operator is monotone on a space of compact sets and is the basis of an iterative procedure for calculating the equilibrium payoff set. The approach of APS can be extended to cover a large class of dynamic games that arise naturally in industrial organization, macroeconomics, and public finance.Footnote 1 In the dynamic case, the object of interest is a value correspondence that maps physical state variables to sets of equilibrium payoffs. In this setting, numerical implementation of the APS approach requires approximation of (candidate value) correspondences. Our paper provides such a scheme that is applicable to games with uncountable endogenous states and integrates it into an APS-type algorithm for computing equilibrium value correspondences.

As a first step, following Cronshaw and Luenberger [8] and Judd et al. [12]’s analyses of repeated games, we convexify a dynamic game by introducing a public randomization device. We approximate candidate value correspondences with simple (convex valued) step correspondences. We describe how to construct sequences of such correspondences that contain or are contained by the approximated candidate value correspondence. Integration of these approximation procedures into an APS-type algorithm permits computation of bounding inner and outer approximations to the equilibrium payoff correspondence. Since the latter lies between these approximations, the difference between them provides us with an error bound for gauging the accuracy of the procedure.Footnote 2 This approach improves on both discretization and linear interpolation methods that have been used elsewhere.Footnote 3 The former is non-parsimonious and is vulnerable to the curse of dimensionality, the latter delivers neither an inner nor an outer approximation and lacks error bounds. To illustrate the value of our methods, we briefly describe an application of them to a bilateral insurance game with storage.

1.1 Literature

Cronshaw and Luenberger [8] provide an algorithm for calculating the payoffs of strongly symmetric equilibria in convexified repeated games. Judd et al. [12] give a numerical implementation of the APS iterative procedure for such games. Their procedure permits computation of all sub-game perfect equilibrium payoffs. Baldauf et al. [5] extend the approach of Judd, Yeltekin, and Conklin to dynamic games with a finite number of (endogenous) states. Our procedure can accommodate an uncountable number of endogenous states. Many alternative numerical methods exist for characterizing the equilibria of dynamic games. These alternatives provide direct characterization of specific sub-game perfect equilibria in certain settings. Examples include Herings and Peeters [11] who provide a homotopy method for computing stationary equilibria in dynamic games and Balbus et al. [4] who show how to calculate extremal Markov equilibria in stochastic games with complementarities. We mention also Feng et al. [9] who use a similar approach to us to characterize equilibria in economies with distortions and taxes.

2 Recursive Dynamic Games and the APS Algorithm

In this section, we lay out a class of recursive dynamic games and describe the theoretic application of the APS algorithm to this class.

2.1 The Game

A group of infinitely lived agents \(i=1,\ldots ,I\) play a dynamic game. Agents enter successive periods \(t=0,1,2,\ldots \) with an inherited pair of state variables \((s_t,k_t)\in S\times K\), where \(S\) is finite and \(K\subset {\mathbb {R}}^{N}\) is compact.Footnote 4 The former is a shock and evolves according to a Markov chain with transition \(\Pi \); the latter is an endogenous state whose evolution is defined below. On entering period \(t\), the agents simultaneously select a profile of actions \(a_t=\{a^i_t\}_{i=1}^I\), with each agent \(i\) choosing his or her action \(a^i_t\) from a compact set \(A^i(s_t,k_t)\subset {\mathbb {R}}^D\). Let \(A=\times _{i=1}^I A^i\) denote the feasibility correspondence for actions. The inherited state \(k_t\) and the action profile \(a_t\) induce a successor state \(k_{t+1}=f(s_t,k_t,a_t)\), with each \(f(s,\cdot ,\cdot ):\text { Graph }A(s)\rightarrow K\) a continuous function.Footnote 5

Each agent has a per-period utility function \( u^{i}:\) Graph \( A\rightarrow {\mathbb {R}}\), with each \(u^i(s,\cdot ,\cdot ):\) Graph \(A(s)\rightarrow {\mathbb {R}}\) a continuous function. Let \(\overline{V}:S\times K\rightrightarrows {\mathbb {R}}^{I}\) denote the feasible payoff correspondence:

$$\begin{aligned} \overline{V}(s,k)=\{w\in {\mathbb {R}}^{I}:w^{i}=u^{i}(s,k,a),a\in A(s,k),i=1,\cdots I\}. \end{aligned}$$

The graph of \(\overline{V}\) is clearly compact.

At \(t=0\), the history of the game \(h_0\) is just the initial aggregate state: \((s_0,k_0)\). For \(t>0\), a history, \(h_{t}\), is a sequence \(\{s_0,k_0,\{a_{\tau },s_{\tau +1},k_{\tau +1}\}_{\tau =0}^{t-1}\}\), with each \(k_{\tau +1}=f(s_\tau ,k_{\tau },a_{\tau })\) and \(a_{\tau }\in A(s_\tau ,k_{\tau })\). Let \(H_t\) denote the set of feasible \(t\) period histories from a given \((s_0,k_0)\). The \(i\)th player values processes for states and actions according to:

$$\begin{aligned} (1-\beta )\sum _{\tau =0}^{\infty }\beta ^{\tau }E[u_{i}(s_\tau ,k_{\tau },a_{\tau })], \end{aligned}$$
(1)

where \(\beta \in (0,1) \) is the common discount factor among agents. A pure strategy for agent \(i\) is a sequence of functions \(\{\sigma ^i_{t}\}_{t=0}^{\infty }\) with \(\sigma ^i _{t}:H_{t}\rightarrow {\mathbb {R}}^D\) and \(\sigma ^i _{t}(h_{t-1},a_{t-1},s_t,k_{t})\in A^{i}(s_t,k_{t})\).

Let \(V^{*}:S\times K\rightrightarrows {\mathbb {R}}^I\) denote the sub-game perfect equilibrium payoff or value correspondence for this game. We assume that \(V^*\) is non-empty-valued. Let \({\fancyscript{V}}\) denote the set of all non-empty-valued correspondences \( V:S\times K\rightrightarrows {\mathbb {R}}^{I}\) such that i) Graph \(V\) is compact and ii) it is contained within Graph \(\overline{V}\). By an elaboration of arguments in APS, \( V^{*}\) is an element of \({\fancyscript{V}}\). Since each element \(V\in {\fancyscript{V}}\) can be identified with its graph, \({\fancyscript{V}}\) can be identified with the collection of non-empty, compact subsets of Graph \(\overline{V}\subset S\times K\times {\mathbb {R}}^{I}\). We furnish \({\fancyscript{V}}\) with the set inclusion order (\( V_{1}\ge (\text {resp.} \le ) V_{2}\) if Graph \(V_{1}\subseteq (\text {resp.} \supseteq )\) Graph \(V_{2}\)) and the metric:

$$\begin{aligned} d_{V}(\text {Graph } V_{1},\text {Graph }V_{2})=\max _S d_H(\text {Graph }V_1(s),\text {Graph }V_2(s)) \end{aligned}$$

where \(d_H\) is the Hausdorff metric

$$\begin{aligned} d_H(\text {Graph }V_1(s),&\text {Graph }V_2(s))=\\&\max \left\{ \max _{x\in \text {Graph } V_{1}(s)}\min _{y\in \text {Graph }V_{2}(s)}\hbox {d}(x,y), \max _{y\in \text {Graph } V_{2}(s)}\min _{x\in \text {Graph}V_{1}(s)}\hbox {d}(x,y)\right\} , \end{aligned}$$

and \(\hbox {d}((k,v),(k^\prime ,v^\prime ))\) is the Euclidean metric on \({\mathbb {R}}^{N+I}\). \(({\fancyscript{V}},\) \(\ge ,\) \(d_{V})\) is a complete, partially ordered metric space.

2.2 Correspondence-valued Dynamic Programming

The recursive approach of APS utilizes a multi-player version of the principle of optimality: each sub-game perfect equilibrium payoff vector is supported by a profile of actions consistent with Nash play in the current period and a vector of continuation payoffs that are themselves sub-game perfect. In analogy with the Bellman operator of dynamic programming, this observation motivates the introduction of the operator \(B:{\fancyscript{V}}\rightarrow {\fancyscript{V}}\):

$$\begin{aligned}&B(V)(s,k)\\&\quad = \left\{ v\left| \begin{array}{l} v^{\prime }(s^{\prime })\in V(s^{\prime },f(s,k,a)),a\in A(s,k),\text { and for each } i\in 1,\cdots ,I,\\ \; v^{i}=(1-\beta )u^{i}(s,k,a)+\beta \sum _{S} v^{i\prime }(s^{\prime })\Pi (s^{\prime }|s)\\ \; v^{i}\ge \max \limits _{A^{i}(s,k)}\min \limits _{\{V(s^{\prime },f(s,k,a^{-i},a^{\prime }))\}} (1-\beta )u^{i}(s,k,a^{-i},a^{\prime })+\beta \sum _{S} \tilde{v}^{i}(s^{\prime })\Pi (s^{\prime }|s) \end{array} \right. \right\} . \end{aligned}$$

A value \(v=\{v^i\}_{i=1}^I\) is in \(B(V)(s,k)\) if two conditions are satisfied. First, there is an action profile \(a\in A(s,k)\) and continuation payoffs \(v^{\prime }(s^{\prime })=\{v^{i\prime }\}(s^{\prime })\in V(s^{\prime },f(s,k,a))\) such that each player \(i\) receives \(v^{i}\) when the current period action is \(a\) and the player’s continuation payoff is \(v^{i\prime }\):

$$\begin{aligned} v^{i}=(1-\beta )u^{i}(s,k,a)+\beta \sum _{S}v^{i\prime }(s^{\prime })\Pi (s^{\prime }|s). \end{aligned}$$
(2)

Second, each player is better off adhering to \(a^{i}\) in the current period and receiving the continuation payoff \( v^{i\prime }\) than deviating and receiving the worst feasible continuation payoff consistent with \(V\):

$$\begin{aligned} v^{i}\ge \max _{A^{i}(s,k)}\min _{\{V(s^{\prime },f(s,k,a^{-i},a^{\prime }))\}} (1-\beta )u_{i}(s,k,a^{-i},a^{\prime })+\beta \sum _{S}\tilde{v}^{i}(s^{\prime })\Pi (s^{\prime }|s). \end{aligned}$$
(3)

Inequality (3) will be referred to as the incentive compatibility constraint.

Remark 1

In some problems, it may be possible to find:

$$\begin{aligned} \underline{v}^{i}(s,k):= \max _{A^{i}(s,k)}\min _{\{V(s^{\prime },f(s,k,a^{-i},a^{\prime }))\}} (1-\beta )u_{i}(s,k,a^{-i},a^{\prime })+\beta \sum _{S}\tilde{v}^{i}(s^{\prime })\Pi (s^{\prime }|s). \end{aligned}$$
(4)

directly. In these cases, this fact may be exploited by substituting \(\underline{v}^i(s,k)\) directly into the incentive compatibility constraint:

$$\begin{aligned} v^{i}\ge \underline{v}^{i}(s,k). \end{aligned}$$
(5)

The modified incentive constraint (5) may then be incorporated into a (modified) \(B\)-operator. We do so in Sect. 4.

For a game with a finite number of states (and compact action sets), Baldauf et al. [5] extend arguments in APS to show that the equilibrium payoff correspondence \(V^*\) is a fixed point of \(B\). In addition, they show that \(B\) is monotone (in the set inclusion order), i.e., \(V_{1}\ge V_{2}\) implies \(B(V_{1})\ge \) \(B(V_{2})\) and a self-map on \({\fancyscript{V}}\), i.e., \(B({\fancyscript{V}})\subset {\fancyscript{V}}\). It follows that for \(j=1,2,\ldots \), \(\overline{V}\ge B(\overline{V})\ge B^j(\overline{V})\ge B^{j+1}(\overline{V})\downarrow \cap _{j=1}^\infty B^j(\overline{V}):=B^\infty (\overline{V})\). Arguments in Baldauf et al. [5] also show that \(B^\infty (\overline{V})=V^*\), thus repeated application of \(B\) to \(\overline{V}\) (and, by similar arguments, any set \(V\) in the interval \(V^*\le V\le \overline{V}\)) generates a sequence of sets that contain \(V^*\) and converge monotonically to it in the Hausdorff metric. Although, Baldauf et al. [5] prove these results for games with a finite number of states, they are readily extended to those with a finite number of exogenous shocks and a compact endogenous state space (in \({\mathbb {R}}^{N}\)), the case considered in this paper. In addition, the monotonicity of \(B\) and the fact that \(V^*\) is a fixed point of \(B\) can also be exploited to show that if \(V\le B(V)\le V^*\), then \(V\le B(V)\le B^j(V)\le B^{j+1}(V)\uparrow \cup _{j=1}^\infty B^j(V):=B^\infty (V)\le V^*\).

2.3 A Recursive Game with Lotteries

To facilitate computation, we “convexify” the preceding game via the introduction of lotteries. Specifically, at the beginning of each period, a lottery is held over a set \(\Omega \). Strategies condition player choices on histories of lottery outcomes as well as histories of states and actions. The lottery probabilities are determined by a profile of functions that map histories into probability distributions on \(\Omega \). These and the lottery outcomes are common knowledge. It is readily shown that this formulation’s equilibrium payoff or value correspondence \(V^{L*}\) is convex valued.Footnote 6 Let \({\fancyscript{V}}^L\) denote the set of all correspondences \( V:S\times K\rightrightarrows {\mathbb {R}}^{I}\) such that (1) Graph \(V\) is compact, (2) Graph \(V\) is convex valued, and (3) Graph \(V\) is contained within Graph \(\overline{V}^L\), where for each \((s,k)\in S\times K\), \(\overline{V}^L(s,k):=\) co \(\overline{V}^L(s,k)\) and co \(X\) denotes the convex hull of a set \(X\). Define the operator \(B^L:{\fancyscript{V}}^L\rightarrow {\fancyscript{V}}^L\) according to: \(B^{L}(V)(s,k)=\) co \(B(V)(s,k)\). It is not hard to show that analogs of the results in the previous subsection go through with \(B^L\) replacing \(B\) and \({\fancyscript{V}}^L\) replacing \({\fancyscript{V}}\). Specifically, (1) \(V^{L*}\) is a fixed point of \(B^{L}\), (2) repeatedly applying \(B^{L}\) to \(\overline{V}^{L}\) induces a sequence of convex-valued correspondences that converge monotonically from above to \(V^{L*}\), and (3) repeatedly applying \(B^L\) to \(V\le B^L(V)\le V^{L*}\) induces a sequence of convex-valued correspondences that converge monotonically from below to \(B^{L\infty }(V)\le V^{L*}\).

3 Approximating Value Correspondences

For a repeated game without state variables, the introduction of lotteries implies that the equilibrium payoff set is convex. Furthermore, (the analog of) \(B^L\) can be restricted to a domain of compact and convex sets. Well-known “outer” and “inner” approximation procedures exist for such sets. Specifically, an outer approximation can be formed from the intersection of a finite number of closed half-spaces that contain the convex set; an inner approximation from the convex hull of a finite number of points contained by the set. The outer approximation provides an upper bound for the set; the inner approximation a lower bound. Together they provide an error bound. Judd et al. [12] integrate such approximations into the monotone iterative scheme described in the previous section and, hence, create practical algorithms for calculating the equilibrium payoff sets of repeated games.

In the setting of dynamic games with state variables, the situation is more complicated. Lotteries ensure that the equilibrium value correspondence \(V^{L*}\) is convex valued (which is useful), but not that the sets Graph \(V^{L*}(s)\) are convex. Consequently, implementation of the theoretic algorithms described in the previous section typically requires the approximation of a sequence of non-convex sets (i.e., the graphs of value correspondences).

To fix language and notation, we introduce some general classes of approximation procedure. We then turn to the specific procedures that we use to represent correspondences.

3.1 Approximation Schemes

We introduce the following general definition of an approximation scheme.

Definition 1

(Approximation scheme). Let \(({\fancyscript{W}},\ge )\) be a collection of non-empty, compact, convex-valued correspondences \(W:S\times K\rightrightarrows {\mathbb {R}}^I\) with common domain \(S\times K\), \(K\subset {\mathbb {R}}^N\) compact, and partially ordered by set inclusion. An approximation rule is a pair \( (F,\hat{\fancyscript{W}})\) with \(\hat{\fancyscript{W}}\subset {\fancyscript{W}}\) and \(F:{\fancyscript{W}}\rightarrow \hat{\fancyscript{W}}\). An approximation scheme is a sequence of approximation rules \( \{F_{j},\hat{\fancyscript{W}}_{j}\}_{j=0}^{\infty }\).

  1. 1.

    An approximation rule \((F,\hat{\fancyscript{W}})\) is finitely parameterized if \(\hat{\fancyscript{W}}=g(T)\), for a set of parameters \(T\subset {\mathbb {R}}^P\) and a one-to-one mapping \(g:T\rightarrow \hat{\fancyscript{W}}\) that associates each parameter value with a unique correspondence. An approximation scheme \(\{F_j,\hat{\fancyscript{W}}_j\}_{j=0}^\infty \) is finitely parameterized if each approximation rule \((F_j,\hat{\fancyscript{W}}_j)\) is finitely parameterized.

  2. 2.

    An approximation rule \((F,\hat{\fancyscript{W}})\) is monotone if for \(W_{1}\), \(W_{2}\) \(\in \fancyscript{W}\), \(W_{1}\) \(\ge \) \(W_{2}\) implies \(F(W_{1})\) \(\ge \) \(F(W_{2})\). A monotone approximation scheme is an approximation scheme composed of monotone approximation rules.

  3. 3.

    An approximation rule \((F,\hat{\fancyscript{W}})\) is outer (resp. inner) if for all \(W\in \fancyscript{W}\), \(F(W)\ge \) \(W\) ( resp. \(W\ge F(W)\)). An outer (inner) approximation scheme is an approximation scheme composed of outer (inner) approximation rules.

  4. 4.

    An approximation scheme \(\{F_{n},\hat{\fancyscript{W}}_{j}\}_{j=0}^{\infty }\) is pointwise (resp. uniformly) convergent if, \( \forall W\in \fancyscript{W}\), the sequence \(F_{j}(W)\) converges to \(W\) pointwise (resp. uniformly) in the metric \(d_V\).

Computation of the convexified equilibrium value correspondence \(V^{L*}\) requires the integration of a monotone iterative scheme with a practical finitely parameterized approximation procedure. Outer and inner approximations that respect the monotonicity of the iterative scheme are especially useful. Identifying \({\fancyscript{W}}\) with \({\fancyscript{V}}^L\), let \(F:\) \({\fancyscript{V}}^L\rightarrow \hat{ {\fancyscript{V}}^L}\) be an approximation rule (with range \(\hat{\fancyscript{V}^L}\)). Then \(\hat{B}=F\circ B^{L}\) maps \( \hat{\fancyscript{V}^L}\) into itself. If \(F\) is monotone, then \(\hat{B}\) is also monotone and if \({\hat{V}}\ge \hat{B}({\hat{V}})\), \({\hat{V}}\in \) \(\hat{\fancyscript{V}^L}\), then \(\{\hat{B}^{j}({\hat{V}})\}_{j=1}^\infty \) is a decreasing and pointwise convergent sequence of correspondences in \(\hat{\fancyscript{V}^L}\). If, in addition, \(F\) is a monotone outer approximation rule and \({\hat{V}}\ge V^{L*}\), then for all \(j\), \(V^{L*}\le F(V^{L*})\) \(=\) \(\hat{B}(V^{L*})\) \(\le \) \(\hat{B}^{j}({\hat{V}})\) \(\le \ldots \) \(\le \hat{B}({\hat{V}})\) \(\le \) \({\hat{V}}\). Hence, the limit of an iteration of \(\hat{B}\) from \({\hat{V}}\) is an outer approximation of \(V^{L*}\). Similarly, if \({\hat{V}}\in \hat{\fancyscript{V}^L}\), \({\hat{V}}\le V^{L*}\), \({\hat{V}}\le \hat{B}({\hat{V}})\) and \(F\) is a monotone inner approximation, the limit of \(\{\hat{B}^{j}({\hat{V}})\}\) is an inner approximation to \(V^{L*}\).

3.2 Block-Specific and Step Correspondences

As a precursor to developing approximation schemes for value correspondences, we introduce certain simple correspondences which we will use to construct approximations. Assume that \(K=\times _{n=1}^N[\underline{k}_{n},\overline{k}_{n}]\subseteq {\mathbb {R}}^{N}\) and that each interval \([\underline{k}_{n},\overline{k}_{n}]\) is subdivided into \(M_{n}\) subintervals \(\{[k_{n,m},k_{n,m+1}]\}_{m=0}^{M_{n}}\), \(\underline{k}_{n}=k_{n,0}<k_{n,1}<\cdots <k_{n,M_n}=\overline{k}_{n}\). The Cartesian product of these subintervals defines a partition \({\mathcal {K}}\) of \(K\), with \({\mathcal {K}}=\{K_{q}\}_{q=1}^{Q}\), \(Q=\prod \nolimits _{n=1}^{N}M_{n}\) and \(K_{q}=\prod \nolimits _{n=1}^{N}[k_{n,m_q},k_{n,m_{q+1}}]\). \(K_{q}\in {\mathcal {K}}\) is called a block.

Definition 2

(Block-specific and step correspondences). Let \({\mathcal {K}}\) be a partition of \(K=\times _{n=1}^{N}[\underline{k}_{n},\overline{k}_{n}]\). A correspondence \(W:K\rightrightarrows {\mathbb {R}}^I \) is block specific if it vanishes (is empty-valued) or is equal to all of \({\mathbb {R}}^I\) outside of a specific block in \({\mathcal {K}}\). \(W\) is a step correspondence if it is constant on the interior of each block in \({\mathcal {K}}\).

The following approximation methods rely on correspondences constructed from finite numbers of block-specific or step correspondences.

3.3 An Outer Approximation Method

This method approximates the underlying correspondence with a collection of convex-valued step correspondences. The procedure is as follows. Let \(V:S\times K\rightrightarrows {\mathbb {R}}^{N}\) be the underlying correspondence. Assume that each \(V(s,\cdot )\) is upper hemi-continuous and non-empty, compact and convex valued. Fix a partition \({\mathcal {K}}=\{K_{q}\}\) of \(K\) and define the family of convex-valued step correspondences \(\omega _{s,q}\) according to:

$$\begin{aligned} \omega _{s,q}(k)={\left\{ \begin{array}{ll} \text {co}\,\cup _{k^{\prime }\in K_{q}}V(s,k^{\prime }) &{} \text {if}\,k\in K_{q}, \\ \emptyset &{} \text {otherwise.} \end{array}\right. } \end{aligned}$$

Each \(\omega _{s,q}\) is completely described by the convex set co \(\cup _{k^{\prime }\in K_{q}}V(s,k^{\prime })\) (and the partition element \(K_q\)). Consequently, it can be approximated by the outer approximation procedure of Judd et al. [12]. More precisely, let \(\{v_{\ell }(s,q)|\ell =1,\ldots ,\overline{\ell }\}\) be a set of \(\overline{\ell }\) points on the boundary of co \(\cup _{k^{\prime }\in K_{q}}V(s,k^{\prime })\); let \(\{z_{1},\ldots ,z_{\overline{\ell }}\}\subset {\mathbb {R}}^{I\overline{\ell }}\) be a corresponding set of normals oriented so that \((v_{\ell }(s,q)-v) \cdot z_{\ell } \ge 0\) for all \(v\in \) co \(\cup _{k^{\prime }\in K_{q}}V(s,k^{\prime })\). The outer approximation \(\omega ^{\mathbb {O}}_{s,q}(k)\) is given by:

$$\begin{aligned} \omega ^{\mathbb {O}}_{s,q}(k) = {\left\{ \begin{array}{ll} \cap _{\ell =1}^{\overline{\ell }}\{v\;|\;z_{\ell }\cdot v\leqq c_{\ell }(s,q)\}, &{} \text {if}\,k \in K_{q}, \\ \emptyset &{} \text {otherwise}. \end{array}\right. } \end{aligned}$$

where for each \(\ell \), \(c_\ell (s,q)=z_{\ell }\cdot v_\ell (s,q)\).Footnote 7

The outer approximation to \(V\) is then defined according to \(V^{\mathbb {O}}(s,k)=\cup _{q}\omega _{s,q}^{\mathbb {O} }(k)\). Using this scheme, an approximating correspondence \(V^{\mathbb {O}}\) is described by a collection \(\{c_{\ell }(s,q)\}\) in \({\mathbb {R}}^{SQI\overline{\ell }}\), with each \(c_{\ell }(s,q)\) determining the position of the \(\ell \)-th bounding hyperplane on the \((s,q)\)-th block.

Let \(\{V^{\mathbb {O}}_{j}\}_{j=0}^{\infty }\) be a sequence of such outer approximations to \(V\). Let \({\mathcal {K}}_{j}\) denote the partition associated with the \(j\)-th approximation and let \(\omega ^{\mathbb {O}} _{s,q,j}\) denote the \((s,q)\)-th block-specific correspondence of the \(j\)-th approximation. Assume that \({\mathcal {K}}_{j}\) is constructed from \(2^{j}\) uniformly distributed blocks and that for a fixed decreasing sequence \(\{\varepsilon _{j}\}\) with \(\lim _{j\rightarrow \infty }\varepsilon _j =0\), the approximations \(\omega ^{\mathbb {O}}_{s,q,j}\) are constructed such that \(d_H (\omega _{s,q,j}^{\mathbb {O}},\omega _{s,q,j})\) \(<\) \(\varepsilon _{j}\) for \(q=1,\cdots ,2^{j} \) (see Appendix 5 for details).

Theorem 1

The sequence \(\{V^{\mathbb {O}}_{j}\}_{j=0}^{\infty }\) is a pointwise convergent, monotone outer approximation scheme. Each \(V^{\mathbb {O}}_{j}(s,\cdot )\) is upper hemi-continuous and non-empty and compact valued.

Proof

Upper hemi-continuity of each \(\omega _{s,q,j}^{\mathbb {O}}\) is immediate. The upper hemi-continuity of \(V^{\mathbb {O}}_{j}(s,\cdot )\) then follows since \(V^{\mathbb {O}}_{j}(s,\cdot )\) is equal to the union of a finite number of upper hemi-continuity correspondences. Each \(V^{\mathbb {O}}_j(s,k)\) is bounded by construction. For the closed valuedness, see Beer [6]. Non-emptiness follows from \(\emptyset \ne V(s,k)\subset V_j^{\mathbb {O}}(s,k)\). Fix \(\varepsilon >0\) and a value \(v\in {\mathbb {R}}^I\). Let \({\mathcal {O}}_{\varepsilon }(X)\) denote the union of all \(\varepsilon \) balls around points in \(X\subset {\mathbb {R}}^I\). By the upper hemi-continuity of \(V(s,\cdot )\), there is a \(\lambda >0\) such that: \(\cup \{V(s,k^{\prime }): \Vert k-k^{\prime }\Vert < \lambda \} \subset {\mathcal {O}}_{\varepsilon }(V(s,k))\). Now, there is a \(j_{1}\) such that for all \(j > j_{1}\), the diameter of each block in the \(j\)-th partition is less than \(\lambda \). Additionally, by construction, there is a \(j_{2}\) such that for all \(j > j_{2}\), \(d_H(\omega ^{\mathbb {O}}_{s,q,j},\omega _{s,q,j})< \varepsilon \), \(\forall q=1,\cdots ,2^{j}\). Fix \(j > \max (j_{1}, j_{2})\) and let \(Q(k)\) denote the set of block indices \(q\) such that \(k\in K_q\in \mathcal K_j\). It follows that\(\cup \{V(s,k^{\prime }):k^{\prime }\in \cup _{q\in Q(k)} K_{q}\} \subset {\mathcal {O}}_{\varepsilon }(V(s,k))\). Since \(V\) is convex valued, it follows that \({\mathcal {O}}_{\varepsilon }(V(s,k))\) is convex. Hence,

$$\begin{aligned} V(s,k)\subset \omega _{s,q(k),j}(k):=\text {co }\cup \{V(s,k^{\prime }):k^{\prime }\in \cup _{q\in Q(k)} K_{q}\} \subset {\mathcal {O}}_{\varepsilon }(V(s,k)) \end{aligned}$$

But since \(j > j_{2}\), for all \(q\in Q(k)\), \(d_H(\omega ^{\mathbb {O}}_{s,q,j},\omega _{s,q,j})< \varepsilon \). Hence, \(V(s,k)\subset V^{\mathbb {O}}_{j}(s,k)=\cup _{q\in Q(k)}\omega ^{\mathbb {O}}_{s,q,j}(k) \subset {\mathcal {O}}_{2\varepsilon }(V(s,k))\), and the approximation is convergent in \(d_V\). Monotonicity of the approximation scheme is immediate from the definition.\(\square \)

In conjunction with the previous result, the following result of Beer [6] shows that a uniformly convergent approximation scheme can be obtained if the correspondence to be approximated is lower hemi-continuous and non-empty and compact valued.

Theorem 2

(Beer [6]). Suppose that each \(V(s,\cdot ):K\rightrightarrows {\mathbb {R}}^I\) is a lower hemi-continuous, compact and non-empty-valued correspondence. Let \(\{V_{j}(s,\cdot )\}_{j=1}^{\infty }\) be a sequence of upper hemi-continuous, compact-valued correspondences such that for each \(k\in K\) and \(j\), \(V(s,k)\subseteq V_{j+1}(s,k)\subseteq V_{j}(s,k)\). If \(\cap _{j=1}^{\infty }V_{j}(s,k)=V(s,k)\) for each \(k\), then \(V_{j}(s,\cdot )\) converges to \(V(s,\cdot )\) uniformly in the Hausdorff metric, \(d_H\).

Under the assumptions made in Sect. 2, the \(B^L\)-operator preserves continuity. Thus, the equilibrium value correspondences \(V^{L*}(s,\cdot )\) can be obtained as the limit of a sequence of continuous correspondences. It is the elements of this sequence that must be approximated. Theorem 2 ensures that there are uniformly convergent approximation schemes available for this purpose.

3.4 The Inner Approximation Step Correspondence Method

This section describes a method for approximating a lower hemi-continuous correspondence with collections of step correspondences whose graphs are contained by the graph of the approximant. Let \(V:S\times K\rightrightarrows {\mathbb {R}}^I\) be the underlying correspondence with each \(V(s,\cdot )\) lower hemi-continuous and with each \(V(s,k)\) convex body, i.e., a compact and convex set with a non-empty interior.Footnote 8 Fix \(s\in S\) and a partition \(\mathcal K\) of \(K\) and define the correspondence \(\omega _{s,q}\) according to:

$$\begin{aligned} \omega _{s,q}(k) = {\left\{ \begin{array}{ll} \cap _{k^{\prime }\in K_{q}}V(s,k^{\prime }) &{} \text {if}\,k \in K_{q},\\ {\mathbb {R}}^I &{} \text {otherwise.} \end{array}\right. } \end{aligned}$$

Thus, each \(\omega _{s,q}\) is completely described by the convex set \(\cap _{k^{\prime }\in K_{q}}V(s,k^{\prime })\) (and the partition element \(K_q\)). Consequently, it can be approximated by the inner approximation procedure of Judd et al. [12]. More precisely, let \(\{v_{\ell }(s,q)|\ell =1,\ldots ,\overline{\ell }\}\) be a set of \(\overline{\ell }\) points in \(\cap _{k^{\prime }\in K_{q}}V(s,k^{\prime })\). The inner approximation \(\omega ^{\mathbb {I}}_{s,q}(k)\) is given by:

$$\begin{aligned} \omega ^{\mathbb {I}}_{s,q}(k) = {\left\{ \begin{array}{ll} \text {co} \ \{v_{\ell }(s,q)|\ell =1,\ldots ,\overline{\ell }\} &{} \text {if}\,k \in K_{q}, \\ {\mathbb {R}}^I &{} \text {otherwise}. \end{array}\right. } \end{aligned}$$

The inner approximation to \(V\) is then defined according to: \(V^{\mathbb {I}}(s,k) = \cap _{q}\omega ^{\mathbb {I}}_{s,q}(k)\). Using this scheme, an approximating correspondence \(V^{\mathbb {I}}\) is described by a collection of points \(\{v_\ell (s,q)\}\) in \({\mathbb {R}}^{SQI\bar{\ell }}\) from which the convex hulls are constructed.

Let \(\{V^{\mathbb {I}}_{j}\}_{j=0}^{\infty }\) denote a sequence of inner approximations to \(V\). Let \({\mathcal {K}}_{j}\) denote the partition associated with the \(j\)-th approximation and define \(\omega ^{\mathbb {I}}_{s,q,j}\) and \(\omega _{s,q,j}\) analogously to \(\omega ^{\mathbb {I}}_{s,q}\) and \(\omega _{s,q}\) above (e.g., \(\omega ^{\mathbb {I}}_{q,j}\) is the \(q\)-th step correspondence of the \(j\)-th approximation). Assume that \({\mathcal {K}}_{j}\) is constructed from \(2^{j}\) uniformly distributed blocks. Additionally, let \(\overline{\ell }_{j}\) denote the number of points used in the construction of the inner approximations \(\{\omega ^{\mathbb {I}}_{s,q,j}\}\). Assume that \(\overline{\ell }_{j+1} > \overline{\ell }_{j}\) and that for a fixed decreasing sequence \(\{\varepsilon _{j}\}\) with \(\lim _{j\rightarrow \infty } \varepsilon _{j} = 0\), each \(\overline{\ell }_j\) is chosen sufficiently large that \(d_H(\omega ^{\mathbb {I}}_{s,q,j},\omega _{s,q,j})\) \( <\) \(\varepsilon _{j}\) for \(q=1,\ldots ,2^j\).

Theorem 3

The sequence \(\{V^{\mathbb {I}}_{j}\}_{j=0}^{\infty }\) is a pointwise convergent, monotone inner approximation scheme. Each \(V^{\mathbb {I}}_{j}(s,\cdot )\) is lower hemi-continuous and compact and convex valued.

The proof is a small adaptation of arguments on p. 15 of Beer [6].

4 Application: A Bilateral Insurance Game

This section considers a problem in which two risk averse agents accumulate a joint fund to hedge endowment risk and neither can commit not to steal from the fund. The problem extends the physical environment of Kocherlakota [13] by permitting the agents to store. The details of the model follow.

An environment is inhabited by two infinitely lived agents with identical preferences. Each period a shock \(s_t\in S\) is drawn from a finite set \(S\) according to a distribution \(\Pi \). For simplicity, shock draws are i.i.d. A function \(\gamma :S\rightarrow {\mathbb {R}}^2_+\), \(\gamma (s)=\{\gamma ^i(s)\}\), gives the endowment of each agent in each shock state. The history of endowments received by each agent is common knowledge. Agents begin the game in a mutual risk-sharing arrangement with \(k_0\ge 0\) in a joint fund. If the arrangement is still active at \(t\), they enter period \(t\) with an amount \(k_t\ge 0\) in the fund. On observing the current endowment realization \(\gamma (s_t)\), each agent transfers \(\tau ^i_t\in \Gamma ^i(k_t)=[-\delta k_t,\gamma ^i(s_t)]\), \(\delta \in [0,1/2]\), to the fund. The agents then choose to remain in the risk-sharing arrangement or walk away from it. If they walk away, any resources left in the fund are destroyed and the agents live in autarky thereafter (though with the ability to accumulate their own fund). While the relationship is active, the fund evolves according to \(k_{t+1}= f(k_{t}+\tau ^{1}_{t}+\tau ^{2}_{t})\), with \(f\) bounded above (and below by zero). If it is dissolved, each agent separately uses the same storage technology \(f\) to accumulate their own fund. The \(i\)-th, \(i=1,2\), agent’s preferences over current consumption are described by utility functions \(u^{i}\), where each \(u^{i}\) is a continuous, strictly concave, strictly increasing function. Each agent discounts the future with discount factor \(\beta \).

At any date \(t\), the agent is free to keep her endowment \(\gamma ^i(s_t)\), withdraw the maximal amount \(\delta k_t\) from the joint fund and live in autarky thereafter. Denote the associated payoff:

$$\begin{aligned} \underline{v}^i(s_t,k_t):=\max _{b \in [0,\gamma ^i(s_t)+\delta k_t]} (1-\beta )u^{i}(\gamma ^i(s_t)+\delta k_t-b)+ \beta v_{aut}^{i}(f(b)), \end{aligned}$$

where \(v_{aut}\) gives the expected payoff to an agent in autarky as a function of her savings. It is not hard to see that \(\underline{v}^i(s_t,k_t)\) gives the worst possible sub-game perfect payoff to agent \(i\) if the relationship is active and in state \((s_t,k_t)\). The payoffs \(\underline{v}^i\) may be calculated directly and substituted into the \(B\)-operator as discussed in Remark 1. The \(B\)-operator so obtained calculates equilibrium payoffs to the agents while the risk-sharing arrangement is active (and uses states sufficient for that purpose). Specifically, it is given by:

$$\begin{aligned} B(V)(s,k)\! =\! \left\{ (v^{1},v^{2})\left| \begin{array}{l} \text {for }i=1,2, \\ \quad v^{i} = \{(1-\beta )u^{i}(\gamma ^i(s)-\tau ^{i}(s)) + \beta \sum _{s^{\prime }\in \mathcal {S}} v^{i^{\prime }}(s^{\prime })\Pi (s^{\prime }), \\ \quad v^{\prime }\in V(f(k+\sum _i \{s^{i}-\tau ^{i}(s)\}),s) \\ \quad (1-\beta )u^{i}(\gamma ^i(s)-\tau ^{i}(s)) + \beta \sum _{s^{\prime }\in \mathcal {S}} v^{i^{\prime }}(s^{\prime })\Pi (s^{\prime }) \ge \underline{v}^i(s,k). \end{array} \right. \right\} \end{aligned}$$

Lotteries may be introduced and the corresponding \(B^L\) operator obtained for the convexified game.

4.1 Calculations

The following parametric forms are assumed for our numerical example: \(u^{i}(c) = \frac{c}{1-\sigma }^{1-\sigma }\) and \(f(k) = \min ( \overline{k},k)\). The preference and technology parameters are set to: \(\beta =0.8\), \(\sigma =1\), \(\overline{k}=4\), \(\gamma ^i(S)=\{0.1, 0.5\}\), \(\Pi =\{0.5, 0.5\}\), \(\delta =0.5\). Calculation using either the outer or inner approximation methods yields very similar results. Figure 1a below illustrates the outer-approximated value correspondence for the case of high endowment shocks for both agents. Figure 1b displays the cross section of the correspondence in Fig. 1a for a sample of \(k\) levels.

Fig. 1
figure 1

Value correspondence from the insurance game. a Value correspondence graph b cross section

The worst equilibrium payoff function for each agent is concave in capital and is independent of the payoff received by the other agent. This functional form is, of course, inherited from the autarkic value function. As the capital level increases, autarky becomes more attractive and the set of equilibrium values decrease, but (symmetric) payoffs on the Pareto frontier increase. These changes can be clearly seen in Fig. 1b. Although incentives for leaving the insurance relationships are higher with high autarky values, high capital levels provide better insurance against aggregate shocks and higher consumption. However, in this example, for capital levels above 3.2, no risk-sharing is feasible. For \(k\in (3.2,4]\), the autarky values are too high, and thus the incentives to deviate are too strong to enable risk-sharing. There is, consequently, a limit to the amount of capital that can be accumulated in this risk-sharing arrangement. That limit depends on the parameters of the model, in particular on the amount of capital that can be grabbed by one agent.