Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Prerequisites. Undergraduate-level real analysis in \(\mathbb{R}^{N}\), basic notions of set theory. The notions of normed linear spaces and metric spaces pop up, but nothing will be lost by thinking of everything in the context of \(\mathbb{R}^{N}\).

1 Multifunctions

A multifunction Footnote 1 Φ from a set X to a set Y (notation: Φ: X ⇉ Y ) is a function Φ defined on X whose values are nonempty subsets of Y.

We’ll quite naturally refer to X as the domain of Φ, but perhaps somewhat less naturally will refer to Y as its range, and will further abuse terminology by defining Φ(E), the image of \(E \subset X\), to be the union of the sets Φ(x) as x ranges through E. Proceeding in this vein, the graph of Φ will be the subset of the product set X × Y defined by

$$\displaystyle{\text{graph}\,(\Phi ) =\bigcup _{x\in X}\{x\} \times \Phi (x) =\{ (x,y) \in X \times Y: y \in \Phi (x)\}.}$$

Example 6.1 (Inverse Maps).

Perhaps the most commonly encountered set-valued map is the inverse map induced by an ordinary function. For such a function f: X → Y, define \(f^{-1}: Y \rightrightarrows X\) by

$$\displaystyle{f^{-1}(y) =\{ x \in X: f(x) = y\}\qquad (y \in Y ).}$$

For example, if \(X = \mathbb{R}\) and f(x) = x 2 for each \(x \in \mathbb{R}\), then \(f^{-1}(y) =\{ \sqrt{y},-\sqrt{y}\}\), a set-valued map whose domain is here taken to be the non-negative real numbers, and whose range is \(\mathbb{R}\).

Exercise 6.1.

For the general inverse mapping \(f^{-1}: Y \rightrightarrows X\), what is \(\text{graph}\,(f^{-1})\), and how does it relate to the graph of f? How does this play out for the particular map \(f^{-1}\) where f(x) = x 2 (x real)?

Exercise 6.2 (A step map).

Consider the set-valued map \(\Phi: \mathbb{R} \rightrightarrows \mathbb{R}\) defined by

$$\displaystyle{\Phi (x) = \left \{\begin{array}{rrr} \{ - 1\}\quad & \text{if} & x <0 \\ {}[-1, 1]\quad & \text{if} & x = 0 \\ \{ + 1\}\quad & \text{if} & x> 0\end{array} \right.}$$

What is \(\Phi (\mathbb{R})\)? Sketch the graph of Φ; how does it differ from that of the “usual” step function?

If Φ: X ⇉ Y where X and Y are metric spaces, we’ll say Φ is closed-valued if Φ(x) is closed in Y for every x ∈ X, with similar definitions applying to all other topological or geometric properties (openness, compactness, convexity, …) that sets may have. We’ll say Φ has closed graph if \(\text{graph}\,(\Phi )\) is closed in the product space X × Y. For example, the step map defined in Exercise 6.2 above has closed graph.

Exercise 6.3.

Show that if X and Y are metric spaces and f: X → Y is continuous, then the graph of f −1: Y ⇉ X is closed in Y × X.

Exercise 6.4.

Suppose X and Y are metric spaces. Show that if Φ: X ⇉ Y has closed graph then Φ(x) is closed for each x ∈ X. Is the converse true?

Example 6.2 (The metric projection).

Suppose (X, d) is a metric space and K is a compact subset of X. Then for each x ∈ X there is an element \(\xi \in K\) that is nearest to x, i.e.,

$$\displaystyle{\inf _{k\in K}d(x,k) = d(x,\xi )\,.}$$

The metric projection of X onto K is the set-valued map P K that associates to a point of X the collection of all these elements of K that are nearest to x.

We’ve already encountered the metric projection in Sect. 4.4, where \(X = \mathbb{R}^{N}\) and K is both compact and convex. In that setting we called P K the “closest point map,” and showed that it’s an “ordinary” function in that each of its values is a singleton.

Exercise 6.5.

Prove that the metric projection of a metric space onto a compact subset has closed graph.

2 Best Response

The notion of “best response” arose during our work in Chap. 5 on Nash Equilibrium. In this section we’ll strip the concept down to its essentials and see how it really concerns a multifunction. We begin with an M × N real matrix A, and its associated quadratic function

$$\displaystyle{u(x,y) = xAy^{t}\qquad (x \in \mathbb{R}^{M},y \in \mathbb{R}^{N}).}$$

Exercise 6.6.

Show that u is continuous on \(\mathbb{R}^{M} \times \mathbb{R}^{N}\).

We’ll be concerned only with the values u takes on the cartesian product Π M × Π N of the standard simplices in \(\mathbb{R}^{M}\) and \(\mathbb{R}^{N}\), respectively, where—as in the last chapter—we think of each simplex as a collection of probability vectors representing mixed strategies for a two-player game, and u(x, y) as Player I’s “payoff” for playing strategy x ∈ Π M against Player II’s strategy y ∈ Π N . Recall that for each y ∈ Π N the function u(⋅ , y) is continuous on the compact subset Π M of \(\mathbb{R}^{M}\), hence there exists x  ∈ Π M such that \(u(x^{{\ast}},y) =\max _{x\in \Pi _{m}}u(x,y)\).

We’ve previously called the probability vector x a “best response” for the first player to the second player’s strategy y. In this section we’ll consider the set BR(y) of all best responses to y, thereby obtaining the “best-response multifunction” BR: Π N  ⇉ Π M for the payoff-matrix A.

Proposition 6.3.

Each value of BR is a nonvoid compact, convex subset of Π M . The graph of BR is closed in Π N × Π M.

Proof.

We’ve already noted that BR(y) is nonempty for each y ∈ Π N . Let’s fix y and write u y (⋅ ) = u(⋅ , y), so u y is an affine continuous function on Π M . Let μ be the first player’s payoff for some best response (hence for all best responses) to the second player’s strategy y, i.e., \(\mu = u(x^{{\ast}},y) = u_{y}(x^{{\ast}})\) for every x  ∈ BR(y). Thus \(\mathrm{BR}(y) = u_{y}^{-1}(\{\mu \})\). Now for each continuous function the inverse image of a closed set is closed, and for each affine map the inverse image of a convex set is convex (Exercise 5.2, page 59). Thus BR(y), being the image under \(u_{y}^{-1}\) of the closed, convex singleton {μ}, is itself a convex subset of Π M that is closed, and hence compact, in Π M .

As for the closed graph, suppose \(((y_{j},x_{j}))_{j=1}^{\infty }\) is a sequence of points in the graph of BR that converges in Π N × Π M to a point (y, x), necessarily also in Π N × Π M . We need to prove that (y, x) belongs to the graph of B, i.e., that x is a best response to y. For each j we know that x j is a best response to y j , i.e., if \(\xi \in \Pi _{M}\) is fixed, then \(u(x_{j},y_{j}) \geq u(\xi,y_{j}).\) By the continuity of u (Exercise 6.6) this inequality is preserved as \(j \rightarrow \infty\), so \(u(x,y) \geq u(\xi,y)\). Since this last inequality is true for each \(\xi \in \Pi _{M}\), the strategy x is a best response to y, i.e., (y, x) belongs to the graph of BR. □ 

3 The Kakutani Fixed-Point Theorem

To say x ∈ X is a fixed point of Φ: X ⇉ X means that x ∈ Φ(x).

Exercise 6.7 (Fixed points of some multifunctions).

  

  1. (a)

    For an “ordinary” function f: X → X, what does it mean to say that x ∈ X is a fixed point of \(f^{-1}: X \rightrightarrows X\)?

  2. (b)

    What are the fixed points of the step function of Exercise 6.2?

  3. (c)

    For the metric projection as in Example 6.2, what are the fixed points?

Here is the main theorem of this chapter.

Theorem 6.4 (Kakutani’s set-valued fixed-point theorem).

If C is a nonvoid compact, convex subset of \(\mathbb{R}^{N}\) and Φ: C ⇉ C is a multifunction with compact, convex values and closed graph, then Φ has a fixed point.

Were Φ to have a “continuous selection,” i.e., a continuous map φ: C → C with φ(x) ∈ Φ(x) for each x ∈ C, then Kakutani’s result would follow immediately from the Brouwer Fixed-Point Theorem. However under the hypotheses of Kakutani’s theorem such selections need not always exist. For example, the “step-map” Φ of Exercise 6.2, when restricted to (say) the closed interval \(C = [-1,1]\), satisfies those hypotheses, but has no continuous selection. Nevertheless there is an “approximate” version of this argument that will carry the day.

To simplify the statement of this “approximate-selection theorem:” for (X, d) a metric space, \(E \subset X\), and \(\varepsilon> 0\), define \([E]_{\varepsilon } =\bigcup _{x\in E}B_{\varepsilon }(x),\) where \(B_{\varepsilon }(x)\) is the open d-ball in X of radius \(\varepsilon\), centered at x. Thus \([E]_{\varepsilon }\) is the set of all points in X that lie a distance \(<\varepsilon\) from E. We’ll call \([E]_{\varepsilon }\) the “\(\varepsilon\)-expansion of E”; clearly, it’s an open subset of X.

Exercise 6.8.

Suppose C is a convex subset of \(\mathbb{R}^{N}\) (or for that matter, of a normed linear space). Show that the \(\varepsilon\)-expansion \([C]_{\varepsilon }\) is convex for every \(\varepsilon> 0\).

When the space in which such expansions take place is the product of two metric spaces, we’ll take the “product distance” to be the maximum of the distances in the individual factors. More precisely: if (X, ρ) and \((Y,\sigma )\) are metric spaces, then the metric in X × Y will be

$$\displaystyle{d((x,y),(x',y')): =\max \, [\rho (x,x'),\sigma (y,y')]\qquad ((x,y),(x',y') \in X \times Y ).}$$

Exercise 6.9.

For the product space \(\mathbb{R} \times \mathbb{R}\) (with the usual metric on each factor) describe and sketch the open \(\varepsilon\)-ball with center at the origin. What’s the relationship between this “product metric” and the Euclidean metric? Do they have the same convergent sequences?

Theorem 6.5 (The Approximate-Selection Theorem).

Suppose X is a compact metric space, Y a nonvoid compact, convex subset of a normed linear space, and \(\Phi: X \rightrightarrows Y\) a multifunction all of whose values are compact and convex. If graph (Φ) is closed, then for every \(\varepsilon> 0\) there exists a continuous function f: X → Y with \(\mathrm{graph}(f) \subset [\mathrm{graph}(\Phi )]_{\varepsilon }\).

Proof of Kakutani’s Theorem. Granting the truth of The Approximate-Selection Theorem, use that theorem to provide, for each positive integer n, a continuous function f n : C → C whose graph lies in \([\text{graph}\,(\Phi )]_{1/n}\). Brouwer’s theorem provides, for each n, a fixed point p n for f n , hence the pair (p n , p n ) lies in \(\text{graph}\,(f_{n})\), and so lies within 1∕n of \(\text{graph}\,(\Phi )\). Since C is compact the sequence of “approximate fixed points” (p n ) has a convergent subsequence—say convergent to p ∈ C. But then the corresponding subsequence of pairs (p n , p n ) converges in C × C to (p, p), which lies a distance zero from \(\text{graph}\,(\Phi )\). Since \(\text{graph}\,(\Phi )\) is closed, (p, p) belongs to it, i.e., p ∈ Φ(p). □ 

For the proof of the Approximate-Selection Theorem we need to explore the consequences for a multifunction of possessing a closed graph.

Lemma 6.6 (Semicontinuity Lemma).

Suppose X and Y are metric spaces with Y compact. If \(\Phi: X \rightrightarrows Y\) has closed graph, then for every \(\varepsilon> 0\) and x ∈ X there exists a ball B x centered at x such that \(\Phi (B_{x}) \subset [\Phi (x)]_{\varepsilon }.\)

Multifunctions satisfying the conclusion of this Proposition are said to be upper semicontinuous, reflecting the fact that for \(\xi\) close to x the set \(\Phi (\xi )\) “cannot be too large” compared to Φ(x).

Proof of Lemma.

Suppose Φ is not upper semicontinuous. Then there exists \(\varepsilon> 0\), x ∈ X, and a sequence (x n ) in X convergent to x such that Φ(x n ) is not contained in \([\Phi (x)]_{\varepsilon }\), i.e., for each index n there exists y n  ∈ Φ(x n ) that lies at least \(\varepsilon\)-distant from Φ(x). Since Y is compact, we may, upon passing to an appropriate subsequence, assume that (y n ) converges to some point y ∈ Y, which also lies at least \(\varepsilon\)-distant from Φ(x). Thus the sequence of pairs (x n , y n ), which belongs to \(\text{graph}\,(\Phi )\), converges in X × Y to the pair (x, y) which is not in \(\text{graph}\,(\Phi )\), so \(\text{graph}\,(\Phi )\) is not closed. □ 

Exercise 6.10 (Converse to the Semicontinuity Lemma).

Suppose X and Y are metric spaces and that Φ: X ⇉ Y is closed-valued and upper semicontinuous. Show that \(\text{graph}\,(\Phi )\) is closed.

Proof of Theorem 6.5. Fix \(\varepsilon> 0\). For each x ∈ X, Lemma 6.6 provides us with a ball B x centered at x for which \(\Phi (B_{x}) \subset [\Phi (x)]_{\varepsilon }\). We may take this ball to have radius \(<\varepsilon\). The collection of balls \(\frac{1} {2}B_{x}\) (center still at x, but radius half that of B x ) is an open cover of the compact space X, and so has a finite subcover which, to save on subscripts, we’ll write as \(\mathcal{B} =\{ \frac{1} {2}B_{1}, \frac{1} {2}B_{2},\,\ldots \, \frac{1} {2}B_{n}\}\), where B j is the original ball centered at x j  ∈ X. By Proposition B.6 of Appendix B there is a partition of unity {p 1, p 2, p n } subordinate to the covering \(\mathcal{B}\), i.e., for each index j the function p j : X → [0, 1] is continuous and vanishes off \(\frac{1} {2}B_{j}\), while the totality of these functions sums to 1 on X. For each index j choose y j  ∈ Φ(x j ); then define f: X → Y by

$$\displaystyle{ f(x) =\sum _{ j=1}^{n}p_{ j}(x)y_{j}\qquad (x \in X). }$$
(6.1)

Thus f is continuous on X, with values in the convex hull of Φ(X). We’ll be done if we can show that \(\text{graph}\,(f) \subset [\text{graph}\,(\Phi )]_{\varepsilon }\), i.e., that for x ∈ X the point (x, f(x)) of \(\text{graph}\,(f)\) lies within \(\varepsilon\) of \(\text{graph}\,(\Phi )\).

Fix x ∈ X and note that since the “partition function” p j vanishes identically off \(\frac{1} {2}B_{j}\), the sum on the right-hand side of Eq. (6.1) involves only those indices j for which \(x \in \frac{1} {2}B_{j}\). Let J = J(x) denote this collection of indices, so that \(f(x) =\sum _{j\in J}p_{j}(x)y_{j}\).

Let m be an index in J(x) corresponding to a ball B j of largest radius for j ∈ J(x). Note that for each j ∈ J(x) our point x lies within \(\text{radius}(\frac{1} {2}B_{j}) \leq \text{radius}(\frac{1} {2}B_{m})\) of the center x j , so all these points x j for j ∈ J(x) lie within radius(B m ) of each other. Thus for each j ∈ J(x) we have (from our semicontinuity-driven initial choice of balls): \(\Phi (x_{j}) \subset [\Phi (x_{m})]_{\varepsilon }\). In particular the point y j , chosen to lie in Φ(x j ), also lies in \([\Phi (x_{m})]_{\varepsilon }\). Conclusion: f(x) lies in the convex hull of \([\Phi (x_{m})]_{\varepsilon }\). But Φ(x m ) is convex, hence so is its \(\varepsilon\)-expansion \([\Phi (x_{m})]_{\varepsilon }\) (Exercise 6.8). Thus \(f(x) \in [\Phi (x_{m})]_{\varepsilon }\), i.e., there exists y ∈ Φ(x m ) with \(d(f(x),y) <\varepsilon\).

Finally, note that x, being a point of \(\frac{1} {2}B_{m}\), lies within \(\text{radius}(B_{m}) <\varepsilon\) of x m , so by our definition of the metric in X × Y the point (x, f(x)) lies within \(\varepsilon\) of \((x_{m},y) \in \text{graph}\,(\Phi )\), i.e., (x, f(x)) lies within \(\varepsilon\) of \(\text{graph}\,(\Phi )\), as promised. □ 

Remark. In this proof we used the compactness of Y only to deduce the upper semicontinuity of Φ from the closed-graph hypothesis. Thus we could eliminate this extra assumption on Y by requiring at the outset that Φ be upper semicontinuous.

4 Application to Nash Equilibrium

We’re now in a position to give Nash’s “one-page” proof [85] of Theorem 5.11 on the existence of Nash Equilibria for mixed-strategy extensions of non-cooperative finite games.

As usual, we’ll keep notation to a minimum by concentrating on the two-person situation. Here the original “pure strategy” game provides M strategies for Player I and N strategies for Player II. The payoff matrices are A and B for the respective players, and the mixed-strategy payoff functions are: u I(x, y) = xAy t for Player I, and u II(x, y) = xBy t, where x ∈ Π M and y ∈ Π N . Denote the best response multifunction for Player I by BRI, and for Player II by BRII. Recall that to say the strategy pair (x , y ) ∈ Π M × Π N is a Nash Equilibrium means that each is a best response to the other, i.e., that x  ∈ BRI(y ) and y  ∈ BRII(x ). So if we define BR: Π M × Π N  ⇉ Π M × Π N by

$$\displaystyle{\mathrm{BR}(x,y) =\mathrm{ BR}_{\mathrm{I}}(y) \times \mathrm{ BR}_{\mathrm{II}}(x)\qquad (x,y) \in \Pi _{M} \times \Pi _{N},}$$

then we’re saying that our strategy pair (x , y ) is a Nash Equilibrium if and only if it’s a fixed point of the multifunction BR.

Since the set Π M × Π N on which BR is acting is a compact, convex subset of \(\mathbb{R}^{MN}\), Kakutani’s Theorem 6.4 will produce the desired fixed point if we can verify that BR satisfies its hypotheses.

To this end, note that we’ve already shown (Proposition 6.3) the individual cartesian factors BRI and BRII of BR to have nonvoid compact convex values. Since these properties are preserved by cartesian products the values of BR are also nonvoid, compact, and convex. We’ve also shown that the factors of BR each have graphs that are closed in their ambient compact product spaces, hence they are compact. Now the graph of BR is homeomorphic via a permutation of coordinates to the cartesian product of the graphs of BRI and BRII, so it too is compact. Thus Kakutani’s theorem applies, and shows that BR has the desired fixed point. □ 

Notes

Kakutani’s Theorem. Theorem 6.5, the approximate selection result that did the heavy lifting in the proof of Kakutani’s Theorem (Theorem 6.4) is attributed variously to von Neumann [89] and more recently to Cellina [23]. The proof given here is taken from [16, pp. 59–62].

Kakutani’s original proof [58]. This takes place on an N-simplex where Kakutani constructs, for each triangulation, a piecewise-affine approximate selection.

Another famous Kakutani. Michiko Kakutani, the influential New York Times literary critic and 1998 Pulitzer Prize winner, is Shizuo Kakutani’s daughter.

von Neumann and Kakutani. As mentioned in the Notes to the previous chapter, von Neumann proved a special case of Theorem 6.4 on which he based his proof of the Minimax Theorem. The reference for this is [89].