Abstract
In this chapter, the concept of Berge equilibrium is introduced as a mathematical model of the Golden Rule. This concept was suggested by the Russian mathematician K. Vaisman in 1994. The Berge–Pareto equilibrium is formalized and sufficient conditions for the existence of such an equilibrium are established. As an application, the existence in the class of mixed strategies is proved.
Access provided by Autonomous University of Puebla. Download chapter PDF
Similar content being viewed by others
Celui qui croit pouvoir trouver en soi-même
de quoi se passer de tout le monde se trompe
fort; mais celui qui croit qu’on ne peut se
passer de lui se trompe encore davantage.
—La RochefoucauldFootnote 1
In this chapter, the concept of Berge equilibrium is introduced as a mathematical model of the Golden Rule. This concept was suggested by the Russian mathematician K. Vaisman in 1994. The Berge–Pareto equilibrium is formalized and sufficient conditions for the existence of such an equilibrium are established. As an application, the existence in the class of mixed strategies is proved.
2.1 What is the Content of the Golden Rule?
Virtue is its own reward.
—English proverb
In the religious-ethical foundations, most nations are guided by the same strategy of behavior, embodied in the demands of the so-called Golden Rule (see Chap. 1). It will hopefully become an established ethical rule for the behavior of the mankind. The well-known statement of the Golden Rule declares, “Behave to others as you would like them to behave to you” (from a lecture delivered by Academician A. Guseinov, Director of RAS Institute of Philosophy, during The IX Moscow Science Festival on October 10, 2014 at Moscow State University). It originates from the New Testament, see the Gospel of Luke, Chapter 6:31, precepting that “And as ye would that men should do to you, do ye also to them likewise.”
Prior to formalizing the concept of Berge equilibrium that matches the Golden Rule, we will present the concept of Nash equilibrium as a standard approach to resolve conflicts. In fact, a critical discussion of the latter has led to the Berge equilibrium, the new solution of noncooperative games that is cultivated in this book.
2.2 Main Notions
Suum cuique.Footnote 2
Nowadays, when the world shudders at the possibility of escalating military conflicts, the Golden Rule becomes more relevant than ever. Indeed, the Golden Rule is a possible way to avoid wars and blood-letting. The modern science of warfare relies mostly on the concept of Nash equilibrium. In this section, the definition of a Nash equilibrium is given, preceded by background material from mathematical theory of noncooperative games.
2.2.1 Preliminaries
Non multa sed multum.Footnote 3
Some general notions from the mathematical theory of noncooperative games that will be needed in the text are presented.
Which mythical means were used by Pygmalion to revivify Galatea? We do not know the true answer, but Pygmalion surely was an operations researcher by vocation: at some moment of time his creation became alive. This idea underlines creative activities in any field, including mathematical modeling. To build an integral entity from a set of odd parts means “to revivify” it in an appropriate sense:
This section is devoted to the revivification of a conflict.
What is a conflict? Looking beyond the common (somewhat criminal) meaning of this word, we will use the following notion from [189, p. 333]: “conceptually a conflict is any phenomenon that can be considered in terms of its participants, their actions, the outcomes yielded by these actions as well as in terms of parties interested in one way or another in these outcomes, including the nature of these interests.” As a matter of fact, game theory suggests mathematical models of optimal decision-making under conflict. The logical foundation of game theory is a formalization of three fundamental ingredients, namely, the features of a conflict, decision-making rules and the optimality of solutions. In this book, we study “rigid” conflicts only, in which each party is guided by his own reasons according to his perception and hence pursues individual goals, l’esprit les intérêst du clocher.Footnote 5
The branch of game theory dealing with such rigid conflicts is known as the theory of noncooperative games. The noncooperative games described in Chap. 2 possess a series of peculiarities. Let us illustrate them using two simple examples.
Example 2.1.1
Imagine several competing companies (firms) that supply the same product in the market. Product price (hence, the profit of each firm) depends on the total quantity of products supplied in the market. The goal of each firm is to maximize its profit by choosing an appropriate quantity of supply.
Example 2.1.2
The economic potential of an individual country can be assessed by a special indicator—a function that depends on controllable factors (taxation, financial and economic policy, industrial and agricultural development, foreign supplies, investments, credits, etc.) and also on uncontrollable factors (climate changes and environmental disasters, anthropogenic accidents, suddenly sparked wars, etc.). Each country seeks to achieve a maximal economic potential through a reasonable choice of the controllable factors with a proper consideration of the existing economic relations with other countries.
These examples elucidate well the character of noncooperative games.
The differentia specifica Footnote 6 of such games are the following.
First, the decision-making process involves several parties (decision makers, e.g., sellers or governments), which are often called players in game theory. Note that a priori they are competitors: quilibet (quisque) fortunae suae faber.Footnote 7
Second, each player has an individual goal (profit or economic potential maximization) and the goals are bound to each other: tout s’enchaine, tout se lie dans ce monde.Footnote 8 A dazzling success of one party may turn out to be a disaster for another party.
Third, each player uses his own tools for achieving his goal (for sellers, the quantity of products supplied; for a country, the controllable factors in Example 2.1.2); in game theory, the controllable factors of each player are called his strategies, while a specific strategy chosen by a player is his decision or action in a noncooperative game.
Let us list three important circumstances.
First, quantitative analysis in any field requires an appropriate mathematical model; this fully applies to noncooperative games. In the course of mathematical modeling, a researcher inevitably faces the risks of going too deep into details (“not see the wood for the trees”) and presenting the phenomenon under study in a rough outline (“throwing out the baby with the bathwater”). The mathematical model of a noncooperative game often includes the following elements:
-
the set of players;
-
for each player, the set of his strategies;
-
for each player, a scalar functional defined on the set of players’ strategies. The value of this functional is the degree to which a given player achieves his goal under given strategies. In game theory, the functional is called the payoff function (or utility function) of a given player.
Second, “many intricate phenomena become clear naturally if treated in terms of game theory.” [21, p. 97]. Following these ex cathedra Footnote 9 pronouncements by Russian game theory maître N. Vorobiev, we are employing the framework of noncooperative games in this book.
A series of conventional requirements have been established for a game-theoretic model (of course, including a sufficient adequacy to the conflict under consideration) as follows.
First, the model must incorporate all interested parties of the conflict (players).
Second, the model must specify possible actions of all parties (the strategies of players).
Third, the model must describe the interests of all parties (for each player and each admissible collection of actions chosen by all players, the model must assign a value called the payoff of that player).
The main challenges of game theory [24] are
-
(1∘)
the design of optimality principles;
-
(2∘)
the proof of existence of optimal actions for players;
-
(3∘)
the calculation of optimal actions.
Different game-theoretic concepts of optimality often reflect intuitive ideas of profitability, stability and equitability, rarely with an appropriate axiomatic characterization. Therefore, in most cases the notion of optimality in game theory (an optimal solution of a game) is not unique, prior or absolute.
We will make the normative Footnote 10 approach to noncooperative games the cornerstone of this book: it will be established which behavior of the players should be considered optimal (rational or reasonable) [47, 48].
Depending on the feasibility of joint actions among the players (coordination of their individual actions), the games are classified as noncooperative, cooperative, and coalitional [50].
In the noncooperative setup of a game (simply called a noncooperative game, see above), each player chooses his action (strategy) in order to achieve the best individual result for himself without any coordination with other players: chacun pour soi, chacun chez soi.Footnote 11
The cooperative setup of a game (cooperative game) is opposite to the noncooperative one. Here all players jointly choose their strategies in a coordinate way and, in some cases, even share the results (their payoffs). Alle für einen, einer für alle.Footnote 12
Finally, in the coalitional setup of a game (coalitional game), all players are partitioned into pairwise disjoint groups (coalitions) so that the members of each coalition act cooperatively while all coalitions play a noncooperative game with each other.
2.2.2 Elements of the Mathematical Model
Ad Disputandum
Consider several subsystems that are interconnected with each other. In economics, these can be industrial enterprises or sectors, countries, sellers in a market, producers of every sort and kind with the same type of products, and other economic systems (called firms in [124, p. 28]). In ecology, industrial enterprises with the same purification and treatment facilities, competing populations of different species (e.g., predators and preys), epidemics propagation and control. In the mechanics of controlled systems, a group of controlled objects (aircrafts, missiles) that attempt to approach each other or to capture an evader.
Each subsystem is controlled by a supervisor (henceforth called a player), who undertakes certain actions for achieving his goal based on available information. In social and economic systems, the role of players is assigned to the general managers of industrial enterprises and business companies, the heads of states, sellers (suppliers) and buyers (customers). In mechanical control systems, this role is played by the captains of ships or aircrafts and the chiefs of control centers.
Assume that, due to a priori conditions, the players have to follow the “Help yourself” slogan. This leads to the noncooperative setup of their interaction.
As an example, consider a simplified mathematical model of competition among N firms in a market.
Example 2.2.3
There are \(N \geqslant 2\) competing firms (players) that supply an infinitely divisible good of the same type (flour, sugar, etc.) in a market. The cost of one unit of good for firm i is \(c_i>0,\; i \in \mathbb {N}=\{1,\ldots ,N\}\). Suppose the number of market participants is sufficiently small so that the prices for goods depend directly on the quantity supplied by each firm. More specifically, denote by K the total supply of goods in the market; then the price p of one good can be calculated as \(p=\max \{a-Kb,0\}\), where a > 0 gives the constant price of one good without any supply in the market, while b > 0 is the elasticity coefficient that characterizes the price drop in response to the supply of one unit of good. Here a natural assumption is that \(c_i<a, \; i \in \mathbb {N}\), since otherwise the activity of firms makes no economic sense. In addition, the production capacities of the players are unlimited and they sell the goods at the price p.
Suppose the firms operate in stable (not extreme) conditions and hence their behavior is aimed at increasing profits. Denote by x i the quantity of goods supplied by firm i \((i\in \mathbb {N})\). Then the total supply of goods in the market is given by
while the profit of firm i is described by the function
where (as before) p is the unit price.
Another reasonable hypothesis is that a − Kb > 0, since otherwise p = 0 and production yields no benefit for all firms (the profits become negative, f i(x) = −c ix i < 0, i ∈N). In this case, the function
is the profit of firm i.
Therefore, in Example 2.2.3 the players are the competing firms and the action (strategy) of each player \( i \in \mathbb {N}\) consists in choosing the quantity \( x _i\in X_i= \left [ 0,+\infty \right )\) of its goods supplied in the market. Making its choice, each player i seeks to maximize its profit f i(x) (payoff) given the supplied quantities x = (x 1, …, x N) of all players.
Idée Générale
Now, let us clarify the framework of the noncooperative games studied in this chapter.
For this purpose, we will answer Quintilian’s questions “Quis? Quid? Ubi? Quibus auxiliis? Cur? Quomodo? Quando?”Footnote 13
Quis?
(Who?) In fact, a leading part in noncooperative games is assigned to the players. As mentioned earlier, players can be general managers of industrial enterprises and business companies, heads of states, sellers (suppliers) and buyers (customers), captains of ships or aircrafts and so on, i.e., those who have the right or authority to make decisions, give instructions and control their implementation (interestingly, some people considering themselves to be (fairly!) serious strongly object to such a game-theoretic interpretation of their activity). Each player has a corresponding serial number: 1, 2, …, i, …, N. Denote by \(\mathbb {N}=\{1,2,\ldots ,N\}\) the set of all players and let the set \(\mathbb {N}\) be finite. Note that games with an infinite number of players (called non-atomic games) are also studied in game theory [160, 171]. Players may form groups, i.e., coalitions \(\mathbb {K} \subseteq \mathbb {N}\). A coalition is any subset \(\mathbb {K}=\{i_1,\ldots ,i_k\}\)of the player set \(\mathbb {N}\). In particular, possible coalitions are singletons (the noncoalitional setup of the game) and the whole set \(\mathbb {N}\) (the cooperative setup of the game). A partition of the set \(\mathbb {N}\) into pairwise disjoint subsets forming N in union is a coalitional structure of the game:
For example, in the noncooperative three-player game \((\mathbb {N}=\{1,2,3\})\), there exist five possible coalitional structures, given by \(\mathcal {P}_1=\{\{1\},\{2\},\{3\}\}\), \(\mathcal {P}_2=\{\{1,2\},\{3\}\}\), \(\mathcal {P}_3=\{\{1\},\{2,3\}\}\), \(\mathcal {P}_4=\{\{1,3\},\{2\}\}\), \(\mathcal {P}_5=\{\{1,2,3\}\}\).
For a compact notation, we will sometimes consider only two-player games, letting \(\mathbb {N} =\{1,2\}\).
In Example 2.2.3, the players are the general managers of competing firms.
Quid?
(What?) Each player chooses and then uses his strategy. A strategy is understood as a rule that associates each state of the player’s awareness with a certain action (behavior) from a set of admissible actions (behaviors) given this awareness. For the head of a state, this is a direction of strategic development. In a sector composed of several industrial enterprises, a strategy of a general manager is the output of his enterprise, the price of products, the amount of raw materials and equipment purchased, supply contracts, investments, innovations and implementation of new technologies, payroll redistribution, penalties, bonuses and other incentive and punishment mechanisms. For a seller, a strategy is the price of one good; for the captain of a ship, own course (rudder angle, the direction and magnitude of reactive force).
Thus, the action of each player consists in choosing and using his individual strategy, which gives an answer to the question Quid? Speaking formally, the strategy of player i in the game Γ3 is x i while the strategy set of this player is denoted by Xi.
Ubi?
(Where?) Here the answer is short: in the conflict, more precisely put, in its mathematical model described by the noncooperative game. In Example 2.2.3, this is the market of goods.
Quibus Auxiliis? Quomodo?
(Who helped? How?) Actually the players affect the conflict using their strategies, which is the answer to both questions.
In Example 2.2.3, the firms choose the quantities of their goods supplied in the market as their strategies. The resulting situation in the market is the strategy profile in the corresponding noncooperative game.
Cur?
(Why?) The answer is: in order to assess the performance of each player. The noncooperative game (the mathematical model of a conflict adopted in our book) incorporates the payoff function of player i (\(i\in \mathbb {N}\)). The value of this function (called payoff or outcome in game theory) is a numerical assessment of the desired performance. In Example 2.2.3, the payoff function of player i has the form
It measures the profit of firm i in the single-stage game. The following circumstances should be taken into account while assessing the performance of each player in a noncooperative game.
First, the design of payoff functions (performance assessment criteria) is a rather difficult and at times subjective task: “Nous ne désirerions guére de choses avec ardeur, si nous connaissions parfaitement ce que nous dèsirons.”Footnote 14 [119, p. 55].
Sometimes, the goal consists in higher profit or lower cost; in other cases, in smaller environmental impact. Other goals are possible as well. As a rule, in a noncooperative game these criteria represent scalar functions defined on the set of all admissible strategy profiles. For the sake of definiteness, assume each player seeks to increase his payoff function as much as possible.
Second, in accordance with the noncoalitional setup of the game, the players act in an isolated way and do not form coalitions. Being guided by the Suum cuique slogan,Footnote 15 each player chooses his strategy by maximizing his own payoff.
As a result, each player endeavors to implement his cherished goal: “Chacun produit selon ses facultés et recoit selon ses besoins.”Footnote 16
Third, the decision-making process in the noncooperative game is organized as follows. Each player chooses and then uses his strategy, which yields a strategy profile of the game. The payoff function of each player is defined on the set of all admissible strategy profiles. The value of this function (payoff) is a numerical assessment of the player’s performance.
In game theory, both terms are equivalent and widespread! Person = player.
At a conceptual level, during the decision-making process in the noncoalitional game player i chooses his strategy x i ∈Xi so that
-
first, this choice occurs simultaneously for all N players;
-
second, no agreements or coalitions among the players and no information exchange are allowed during the game [178, p. 1].
Quando?
(When?) The answer to the last question of Quantilian’s system is the shortest: at the time of decision-making in the conflict (within its mathematical model—the noncoalitional game) through an appropriate choice of strategies by the players.
In principle, a conflict can be treated as a certain controlled system, a “black box” in which the players input their strategies and receive their payoffs at the output. This is a standard approach to “instantaneous, single-period, static” noncooperative games in general game theory [23]. However, in most applications (particularly, in economics and the mechanics of controlled systems), the controlled system itself undergoes some changes with time, and the players are able to vary their strategies during the whole conflict. The games whose state evolves in time are called dynamic. Hopefully, our next book will be focused on the analysis of dynamic games.
2.2.3 Nash Equilibrium
Politica del campanile.Footnote 17
A generally accepted solution concept for noncooperative games is the so-called Nash equilibrium.Footnote 18 Nash equilibrium is widely used in economics, military science, policy and sociology. Almost each issue of modern journals on operations research, systems analysis or game theory contains papers involving the concept of Nash equilibrium.
Thus, let us consider a noncooperative three-player game described by
where each player i = {1, 2, 3} chooses an individual strategy \(x_i \!\in \!\mathrm {X}_i \!\subseteq \mathbb {R}^{n_i}\) in order to increase his performance \(f_i(x\!=(x_1,\!x_2,\!x_3)\!)\), i.e., his payoff f i(x) in a current strategy profile x = (x 1, x 2, x 3) ∈ ∈X1 ×X2 ×X3 = X.
A Nash equilibrium is a pair \((x^{\mathrm {e}},f^{\mathrm {e}} = (f_1(x^{\mathrm {e}}),f_2(x^{\mathrm {e}}),f_3(x^{\mathrm {e}})) \in \mathrm {X} \times \mathbb {R}^{3}\) defined by the three equalities
Each player therefore acts selfishly, seeking to satisfy his individual ambitions regardless of the interests of the other players. As repeatedly mentioned earlier, this concept of equilibrium was suggested in 1949 by J. Nash, a Princeton University graduate at that time and a famous American mathematician and economist as we know him today. Moreover, 45 years later J. Nash, J. Harsanyi and R. Selten were awarded the Nobel Prize in Economic Sciences “for the pioneering analysis of equilibria in the theory of non-cooperative games.” Let us note two important aspects. First, owing to his research in the field of game theory, by the end of the twentieth century J. Nash became a leading American apologist of the Cold War. Second, the Nash equilibrium had been so widely used in economics, sociology, and military science that during the period 1994–2012 the Nobel Committee awarded seven Nobel Prizes for different investigations that to a large degree stemmed from the concept of Nash equilibrium. However, the selfish character of NE prevents it from “paving the way” towards a peaceful resolution of conflicts.
2.2.4 Berge Equilibrium
Que jamais le mérite avec lui ne
perd rien, Et que, mieux que du
mal, il se souvient du bien.Footnote 19
Almost all notions in the modern theory of measure
and integral go back to Lebesgue’s works,
and introduction of these notions was in some sense
a turning point of transition from the mathematics of
the 19th century to the science of the 20th century.
—VilenkinFootnote 20
A peaceful resolution of conflicts can be achieved using Berge equilibrium (BE). This concept appeared in 1994 in Russia, following a critical analysis of C. Berge’s book [202]. Interestingly, Berge wrote his book as a visiting professor at Princeton University, simultaneously with Nash, who also worked there under support of the Alfred P. Sloan Foundation.
A Berge equilibrium is a pair (x B, f B = (f 1(x B), f 2(x B), f 3(x B))) defined by the equalities
Equilibria (2.2.2) and (2.2.1) exhibit the following fundamental difference. In (2.2.1), each player directs all efforts to increase his individual payoff (the value of his payoff function) as much as possible. The antipode of (2.2.1) is (2.2.2), where each player strives to maximize the payoffs of the other players, ignoring his individual interests. Such an altruistic approach is intrinsic to kindred relations and occurs in religious communities. The elements of such altruism can be found in charity, sponsorship, and so on. The concept of Berge equilibrium also provides a solution to the Tucker problem in the well-known Prisoner’s Dilemma (see Example 2.6.1 below). Due to (2.2.2), an application of this equilibrium concept eliminates armed clashes and murderous wars. This is an absolute advantage of Berge equilibrium.
As a matter of fact, the Berge equilibrium had an unenviable fate. The publication of the book [202] in 1957 initiated a sharp response of Shubik [269, p. 821] (“…no attention has been paid to applications to economics…the book will be of a little direct interest to economists…”). Most likely, such a negative review in combination with Shubik’s authority in scientific community pushed away the Western experts in game theory and economics from the book [202]. In Russia, after its translation in 1961, the book was analyzed in depth (Russian researchers were not acquainted with Shubik’s review!) and the concept of Berge equilibrium was suggested on the basis of an appropriate modification of the notion of Nash equilibrium. The difference between Berge equilibrium and Nash equilibrium is that the former postulates stable payoffs against the deviations of all players and also reassigns the “ownership” of the payoff function (in the definition of a Nash equilibrium, the strategies of a separate player and all other players are interchanged). Note that the book [202] did not actually introduce the definition of Berge equilibrium, but it inevitably comes to mind while studying the results of Chaps. 1 and 5 of [202].
Subsequently, the Berge equilibrium was rigorously defined in 1994–1995 by K. Vaisman in his papers and dissertation [11, 13, 302], under the scientific supervision of V. Zhukovskiy. This concept was immediately applied in [280, 281] for noncooperative linear-quadratic positional games under uncertainty. Unfortunately, Vaisman’s sudden death at the age of 35 suspended further research on Berge equilibrium in Russia. At that time, however, the concept of Berge equilibrium was “exported from Russia” by Algerian postgraduates of V. Zhukovskiy Radjef [266] and Larbani [248]. Later on, it was actively used by Western researchers (e.g., see the survey [255] with over 50 references and also the recent review [131, pp. 53–56] published in Ukraine). As shown by these and more than 100 subsequent publications, most of research works are dedicated to the properties of Berge equilibrium, the specific features and modifications of this concept, and relations to Nash equilibrium. It seems that an incipient theory of Berge equilibrium will soon emerge as a rigorous mathematical theory. Hopefully, an intensive accumulation of facts will be replaced by the stage of evolutionary internal development.
This chapter reveals the internal instability of the set of Berge equilibria. To eliminate this negative feature, we suggest a method to construct a Berge equilibrium that is Pareto-maximal with respect to all other Berge equilibria. The method reduces to a saddle point calculation for an auxiliary zero-sum two-player game that is effectively designed using the original noncooperative game. As a supplement, we prove the existence of such a (Pareto refined) Berge equilibrium in mixed strategies under standard assumptions of mathematical game theory, i.e., compact strategy sets and continuous payoff functions of the players.
2.3 Compactness of the Set XB
The notion of infinity is our greatest friend;
it is also the greatest enemy of our peace in mind.
—PierpontFootnote 21
It is shown that the set of Berge equilibria is closed and bounded.
Thus, we consider the mathematical model of a conflict in the form of a noncooperative N-player game, \(N \geqslant 2\), described by an ordered triplet
Here \(\mathbb {N} = \{ 1,2,\ldots ,N \}\) denotes the set of players; each of the N players, forming no coalitions with other players, chooses his strategy (action) \(x_i \in \mathrm {X}_i \subseteq \mathbb {R}^{n_i}\) (throughout the book, the symbol \(\mathbb {R}^{k}, \; k \geqslant 1\), stands for the k-dimensional Euclidean space whose elements are ordered sets of k real numbers in the form of columns, with the standard scalar product and the Euclidean norm); such a choice yields a strategy profile \(x\!=(x_1,\ldots ,x_N) \!\in \! \mathrm {X}\! =\!\prod _{i \in \mathbb {N}}\!\mathrm {X}_i \!\subseteq \!\mathbb {R}^{n}\) (\(n\! = \!\sum _{i \!\in \mathbb {N}}\! n_i\)); a payoff function f i(x) defined on the set X numerically assesses the performance of player i \((i \in \mathbb {N})\); let (x∥z i) = (x 1, …, x i−1, z i, x i+1, …, x N) and f = (f 1, …, f N).
A pair \((x^{\mathrm {B}},f^{\mathrm {B}}) = ((x^{\mathrm {B}}_1,\ldots ,x^{\mathrm {B}}_N),(f_1(x^{\mathrm {B}}),\ldots ,f_N(x^{\mathrm {B}}))) \in \mathrm {X} \times \mathbb {R}^{N}\) is called a Berge equilibrium in game (2.3.1) if
In the sequel, we will consider mostly the strategy profiles x B from such pairs, also calling them Berge equilibria in game (2.3.1).
Property 2.3.1
If in the game Γ the sets Xi are closed and bounded, i.e., \(\mathrm {X}_i \in \mathrm {comp} \:\mathbb {R}^{n_i}\), and the payoff functions f i(⋅) are continuous, f i(⋅) ∈ ∈ C(X) \((i \in \mathbb {N})\), then the set XB of all Berge equilibria in the game Γ is compact in X (possibly, empty) and \(f(\mathrm {X}^B) \in \mathrm {comp} \:\mathbb {R}^{N}\).
Proof
Since XB ⊆X and \(\mathrm {X} \in \mathrm {comp} \:\mathbb {R}^{n}\), then XB is bounded. Thus, if we can show that XB is closed, then \(\mathrm {X}^B \in \mathrm {comp} \:\mathbb {R}^{n}\). Let us prove the closedness of XB by contradiction. Assume that, for a infinite sequence \(\{ x^{(k)} \}_{k=0}^{\infty }\), x (k) ∈XB, there exist a subsequence \(\{ x^{(k_r)} \}_{r=0}^{\infty }\) and a strategy profile x ∗∈X such that, first, \(\lim _{r \rightarrow \infty } x^{(k_r)} = x^{*}\) and, second, x ∗∉XB.
Since x ∗∉XB, there exist a strategy profile \(\bar {x} \in \mathrm {X}\) and a number \(j \in \mathbb {N}\) such that \(f_j( \bar {x}\|x^{*}_j)\! >\! f_j(x^{*})\), where \(x^{*}\!=(\!x^{*}_1,\ldots ,x^{*}_j,\ldots ,x^{*}_N\!)\) and, as before, \((\bar {x}\|x^{*}_j ) = (\bar {x}_1,\ldots ,\bar {x}_{j-1},x^{*}_j,\bar {x}_{j+1},\ldots ,\bar {x}_N)\).
Owing to the continuity of \(f_j( \bar {x}\|x_j )\) and f j(x) in x ∈X and the convergence \(\lim _{r \rightarrow \infty } x^{(k_r)} = x^{*}\), there exists an integer M > 0 such that, for \(r \geqslant M\), \(f_j( \bar {x}\| x^{(k_r)}_j ) > f_i(x^{(k_r)})\). This strict inequality contradicts \(f_j( x\| x^{\mathrm {B}}_j ) \leqslant f_j( x^{\mathrm {B}})\) ∀x ∈X, and the conclusion follows. \( \blacksquare \)
Corollary 2.3.1
Let the hypotheses of Property 2.3.1 be valid and the set XB ≠ ∅ in the game Γ. Then there exists a Berge equilibrium that is Pareto-maximal with respect to all other equilibria x B ∈XB in this game.
Indeed, since the set XB is compact, f i(⋅) ∈ C(X) \((i \in \mathbb {N})\) and the hypotheses of Property 2.3.1 hold, the N-criteria choice problem
has a Pareto-maximal alternative x B ∈XB [152, p. 149]. In other words, for every x ∈XB, the system of N inequalities
with at least one strict inequality, is inconsistent.
2.4 Internal Instability of the Set XB
It is easier to stop the Sun and move the Earth
than to decrease the sum of angles in a triangle,
to make parallels converge, and to drop perpendiculars
to the same line from a far distance.
—KaganFootnote 22
It is found that there may exist two Berge equilibria, in one of which each player has a strictly greater payoff than in the other.
Property 2.4.1
The set XB of all Berge equilibria can be internally unstable, i.e., in the game Γ there may exist two Berge equilibria x (1) and x (2) such that, for all \(i \in \mathbb {N}\),
Example 2.4.1
Consider a noncooperative two-player game (N = 2) of the form
In this game, the strategy profiles are x = (x 1, x 2) ∈ [−1, +1]2, the strategy sets of both players coincide, Xi = [−1, +1] (i = 1, 2), while the Berge equilibrium \(x^{\mathrm {B}}=(x^{\mathrm {B}}_1,x^{\mathrm {B}}_2)\) is defined by the inequalities
or
(these inequalities follow from (2.3.2)). Hence, \(x^{\mathrm {B}}_1 =x^{\mathrm {B}}_2 = \alpha \) for all α = const ∈ [−1, +1] (see Fig. 2.1a), and then \(f^{\mathrm {B}}_i = f_i(x^{\mathrm {B}}) = \alpha ^2\) for all α = const ∈ [−1, +1] (see Fig. 2.1b).
Thus, we have established that, first, there may exist a continuum of Berge equilibria (in Example 2.4.1, the set XB = AB as illustrated by Fig. 2.1a) and, second, the set XB is internally unstable, since f i(0, 0) = 0 < f i(1, 1) = 1 (i = 1, 2) (see Fig. 2.1b).
Hence, in the game Γ the players should use the Berge equilibrium that is Pareto-maximal with respect to all other Berge equilibria. We introduce the following definition for further exposition.
Definition 2.4.1
A strategy profile x ∗∈X is called a Berge–Pareto equilibrium (BPE) in the game Γ if
-
(1)
x ∗ is a Berge equilibrium in Γ (x ∗ satisfies conditions (2.3.2));
-
(2)
x ∗ is a Pareto-maximal alternative in the N-criteria choice problem
i.e., for any alternatives x ∈XB, the system of inequalities
with at least one strict inequality, is inconsistent.
In Example 2.4.1, we have two BPE, x (1) = (−1; −1) and x (2) = (+1; +1), with the same payoffs f i(x (1)) = f i(x (2)) = 1 (i = 1, 2).
Remark 2.4.1
If XB ≠ ∅, \(\mathrm {X}_i \in \mathrm {comp} \:\mathbb {R}^{n_i}\), and f i(⋅) ∈ C(X) \((i \in \mathbb {N})\), then Definition 2.4.1 relies on Corollary 2.3.1, stating that the set of BPE is nonempty under the two requirements above.
Interestingly, the set of Nash equilibria in the game Γ is also internally unstable (this is demonstrated by Example 2.4.1 with the change x 1 ↔ x 2).
In the forthcoming sections, we will establish sufficient conditions for the existence of a BPE, which are reduced to a saddle point calculation for an auxiliary zero-sum two-player game that is effectively designed using the original noncooperative game.
2.5 No Guaranteed Individual Rationality of the Set XB
Among the splendid generalizations effected by modern mathematics,
there is none more brilliant or more inspiring or more fruitful,
and none more commensurate with the limitless immensity of being itself,
than that which produced the great concept designated …
hyperspace or multidimensional space.
—KeyserFootnote 23
A Nash equilibrium has the property of individual rationality, whereas a Berge equilibrium generally does not, as illustrated by an example in this section. It is also established that there may exist a Berge equilibrium in which at least one player obtains a smaller payoff than the maximin.
Another negative property of a Berge equilibrium is the following.
Property 2.5.1
A Berge equilibrium may not satisfy the individual rationality conditions, as opposed to the Nash equilibrium x e in the game Γ2 (under the assumptions \(\mathrm {X}_i \in \mathrm {comp}\:\mathbb {R}^{n_i}\) and f i(⋅) ∈ C(X) \((i \in \mathbb {N})\), the game Γ2 (the game Γ with \(\mathbb {N} = \{1,2\}\)) satisfies the inequalities
known as the individual rationality conditions).
Example 2.5.1
Consider a noncooperative two-player game of the form
where x = (x 1, x 2). A Berge equilibrium \(x^{\mathrm {B}}=(x^{\mathrm {B}}_1,x^{\mathrm {B}}_2)\) in the game \(\Gamma _2^{\prime }\) is defined by the two equalities
The second equality holds only for the strategy \(x^{\mathrm {B}}_1=1\). Due to the strong convexity of f 1(x) in x 2 (which follows from the fact that \(\left . \frac {\partial ^2 f_1(x^{\mathrm {B}}_1 ,x_2)}{\partial x^2_2}\right |{ }_{x_2} = 2 > 0\)), the maximum of the function
is achieved on the boundary of X2, more specifically, at the point \(x^{\mathrm {B}}_2 = 1\). Thus, the game \(\Gamma _2^{\prime }\) has the Berge equilibrium x B = (1, 1), and the corresponding payoff is f 1(x B) = f 1(1, 1) = −1.
Now, find \(\max \limits _{x_1 \in \mathrm {X}_1} \min \limits _{x_2 \in \mathrm {X}_2} f_1(x_1,x_2)\) in two steps as follows. In the first step, construct a scalar function x 2(x 1) that implements the inner minimum:
By the strong convexity of f 1(x 1, x 2) in x 2,
yielding the unique solution x 2(x 1) = −x 1 and \(f_1[x_1]=f_1(x_1,x_2(x_1))=- 5 x^2_1\).
In the second step, construct the outer maximum, i.e., find
Consequently,
which shows that the individual rationality property may fail for a Berge equilibrium.
Remark 2.5.1
Individual rationality is a requirement for a “good” solution in both noncooperative and cooperative games: each player can guarantee the maximin individually, i.e., by his own maximin strategy, regardless of the behavior of the other players [173]. However, in a series of applications (especially for the linear-quadratic setups of the game), the maximin often does not exist. Such games were studied in the books [52, pp. 95–97, 110–116, 120] and [93, pp. 124–131].
In the case where game (2.3.1) has maximins, Vaisman suggested to incorporate the individual rationality property into the definition of a Berge equilibrium. Such equilibria are called Berge–Vaisman equilibria.
2.6 Two-Player Game
You don’t have to be a mathematician
to have a feel for numbers.
—NashFootnote 24
The specific features of Berge equilibria in two-player games are identified.
Non-antagonistic Case
Consider a special case of game (2.3.1) with two players, i.e., the game Γ in which \(\mathbb {N} = \{ 1,2 \}\). Then a Berge equilibrium \(x^{\mathrm {B}}=(x^{\mathrm {B}}_1,x^{\mathrm {B}}_2)\) is defined by the equalities
Recall that a Nash equilibrium x e in this two-player game is given by the conditions
A direct comparison of these independent formulas leads to the following result.
Property 2.6.1
A Berge equilibrium in game (2.3.1) with \(\mathbb {N} = \{ 1,2 \}\) coincides with a Nash equilibrium if both players interchange their payoff functions and then apply the concept of Nash equilibrium to solve the modified game.
Remark 2.6.1
In view of Property 2.6.1, a special theoretical study of Berge equilibrium in game (2.3.1) with \(\mathbb {N} = \{ 1,2 \}\) seems unreasonable, despite careful attempts by a number of researchers. In fact, all results concerning Nash equilibrium in a two-player game are automatically transferred to the Berge equilibrium setting (of course, with an appropriate “interchange” of the payoff functions, as described by Property 2.6.1).
Let us proceed with an example of a two-player matrix game in which the players have higher payoffs in a Berge equilibrium than in a Nash equilibrium (in the setting of game, this is an analog of the Prisoner’s Dilemma).
Also note the following interesting fact for game (2.3.1) with \(\mathbb {N} = \{ 1,2 \}\), \(f_i(x)=x^T_1 A_i x_1 +x^T_2 B_i x_2\), the strategies \(x_1 \in \mathbb {R}^{n_1}\) and \(x_2 \in \mathbb {R}^{n_2}\), where the matrices A i and B i of compatible dimensions are square, constant and symmetric, A 1 > 0, B 1 < 0, A 2 < 0, and B 2 > 0 (the notation A > 0 (<) stands for the positive (negative) definiteness of the quadratic form x TAx): in this game, there exist no Nash equilibria, while the strategy profile \((0_{n_1},0_{n_2})\) forms a Berge equilibrium (as before, 0k denotes the zero vector of dimension k).
Example 2.6.1
Consider the bimatrix game in which player 1 has two strategies, i.e., chooses between rows 1 and 2. Accordingly, the strategies of player 2 are represented by columns 1 and 2. For example, the choice of the strategy profile (1, 2) means that the payoffs of players 1 and 2 are 4 and 7, respectively.
According to the above definitions, in this bimatrix game the strategy profiles (2, 2) and (1, 1) are a Nash and Berge equilibrium, respectively. As 6 > 5, the payoffs of both players in the Berge equilibrium are strictly greater than their counterparts in the Nash equilibrium. The same result occurs in the Prisoner’s Dilemma, a well-known bimatrix game. Note that the paper [255] gave some examples of 2 × 2 bimatrix games in which the payoffs in a Nash equilibrium are greater than or equal to those in a Berge equilibrium.
Antagonistic Case
To conclude this section, consider the antagonistic case of game (2.3.1), which arises for Γ with \(\mathbb {N} = \{ 1,2 \}\) and f 2(x) = −f 1(x) = f(x). In other words, consider an ordered triplet
A conventional solution of the game Γa is the saddle point \(x^0 = (x^0_1,x^0_2) \in \mathrm {X}_1 \times \mathrm {X}_2\), which is formalized here by the chain of inequalities
Property 2.6.2
For the antagonistic case Γa of the game Γ, the Berge equilibrium \((x^{\mathrm {B}}_1,x^{\mathrm {B}}_2)\) matches the saddle point \((x^0_1,x^0_2)\) defined by (2.6.1).
The proof of this property follows immediately from the inequalities
and the identity f(x) = f 2(x) = −f 1(x) ∀ x ∈X. \(\blacksquare \)
2.7 Comparison of Nash and Berge Equilibria
There are only two kinds of certain knowledge:
Awareness of our own existence and the truths of mathematics.
—d’AlembertFootnote 25
A detailed comparison of Berge and Nash equilibria is made.
NE | BE |
Stability | |
against a unilateral deviation of a single player, since the inequality \(f_i (\!x^{\mathrm {e}}_1,\ldots ,x^{\mathrm {e}}_{i\!-\!1},x_i,x^{\mathrm {e}}_{i\!+\!1},\ldots ,x^{\mathrm {e}}_N \!) \!\! \leqslant \!\! f_i (\!x^{\mathrm {e}}) \) holding ∀x i ∈ Xi \((i\!\in \! \mathbb {N})\) implies that the payoff of the deviating player i is not greater than in the NE. | against the deviations of the coalition of all players except player i, since the inequality \(f_i (x_1,\ldots ,x_{i\!-\!1},\! x^{\mathrm {B}}_i \!\! ,x_{i\!+\!1},\ldots ,\! x_N \!) \!\! \leqslant \!\! f_i (x^{\mathrm {B}} \!) \) holding ∀x j ∈ Xj \((j \!\in \! \mathbb {N} \!\setminus \! \{i\}\), \(i\!\in \! \mathbb {N})\) implies that the payoff of each player i under such a deviation of the coalition of the other N − 1 players from the BE is not greater than in the BE. |
Individual rationality (IR) | |
Here and in the sequel, \(x_{\mathbb {N} \setminus \! \{i\}} \!\!=\!\! (x_1,\ldots ,x_{i\!-\!1},\) \(x_{i+1},\ldots ,x_N) \in {\mathbf {X}}_{\mathbb {N} \setminus \{i\}} = \prod \limits _{j \in \mathbb {N} \setminus \{i\}}\!\!\! {\mathbf {X}}_j.\) If x e exists and | Generally speaking, fails (see Property 2.5.1, Example 2.5.1, and Remark 2.5.1). |
\(f^{\mathrm {g}}_i = \max \limits _{x_i \!\in \! {\mathbf {X}}_i} \min \limits _{x_{\mathbb {N} \setminus \! \{i\}} \!\in \! {\mathbf {X}}_{\mathbb {N} \setminus \{i\}}} f(x\|x_i) =\) \(= \min \limits _{x_{\mathbb {N} \setminus \! \{i\}} \!\in \! {\mathbf {X}}_{\mathbb {N} \setminus \{i\}}} f_i(x\|x^{\mathrm {g}}_i)\) \((i\in \mathbb {N})\), then \(f_i(x^{\mathrm {e}}) \!\!\geqslant \!\! f^{\mathrm {g}}_i\) \((i \!\!\in \!\! \mathbb {N})\), i.e., NE satisfies the IR condition. | |
Internal instability | |
The set of NE is internally unstable (see proof in [54]). | The set of BE is internally unstable (see Property 2.4.1 and Example 2.4.1). |
To eliminate this drawback both for the NE and BE , | |
Pareto maximality with respect to | |
the other equilibria of a given type is required. | |
Saddle point (SP) in the game | |
\(\Gamma _{a} = \left \langle \{1,2\}, \{ {\mathbf {X}}_i \}_{i=1,2}, \{ f_1(x_1,x_2),f_2(x_1,x_2)=-f_1(x_1,x_2) \} \right \rangle \) | |
is a special case of NE and BE. | |
NE coincides with the SP \((x^{\mathrm {e}}_1,x^{\mathrm {e}}_2)\) of the form \(\max \limits _{x_1 \in {\mathbf {X}}_1} f_1(x_1,x^{\mathrm {e}}_2)\!\! =\) \(=f_1(x^{\mathrm {e}}_1,x^{\mathrm {e}}_2) = \min \limits _{x_2 \in {\mathbf {X}}_2} f_1(x^{\mathrm {e}}_1,x_2)\). | The BE coincides with the SP \((x^{\mathrm {B}}_1,x^{\mathrm {B}}_2)\) of the form \(\max \limits _{x_2 \in {\mathbf {X}}_2} f_1(x^{\mathrm {B}}_1,x_2) =\) \(=f_1(x^{\mathrm {B}}_1,x^{\mathrm {B}}_2) = \min \limits _{x_1 \in {\mathbf {X}}_1} f_1(x_1,x^{\mathrm {B}}_2)\). |
Conclusions
The main difficulty in many modern developments of mathematics
is not to learn new ideas but to forget old ones.
—SawyerFootnote 26
Nash equilibria have three undisputable advantages, namely, they are stable, coincide with the saddle point (containing this generally accepted concept as a special case), and satisfy the individual rationality condition. The first and second advantages are shared also by the Berge equilibria.
At the same time, Nash equilibria suffer from several drawbacks, namely, the internal instability of the set of NE and selfishness (each player seeks to increase his individual payoff, as by definition \(\forall i \in \mathbb {N}\): \(f_i(x^{\mathrm {e}})=\max \limits _{x_i \in \mathrm {X}_i} f_i(x^{\mathrm {e}}\|x_i)\)).
Internal instability is intrinsic to the set of BE too. This negative feature can be eliminated by requiring Pareto maximality for the NE and BE. The selfish nature of NE is eliminated using the altruistic orientation of BE (“help the others if you seek for their help”). This constitutes a clear merit of BE as a way of benevolent conflict resolution.
2.8 Sufficient Conditions
Mathematics as an expression of the human mind reflects the active will,
the contemplative reason, and the desire for aesthetic perfection.
Its basic elements are logic and intuition, analysis and construction,
generality and individuality. Though different traditions may emphasize
different aspects, it is only the interplay of these antithetic forces
and the struggle for their synthesis that constitute the life, usefulness,
and supreme value of mathematical science.
—CourantFootnote 27
2.8.1 Continuity of the Maximum Function of a Finite Number of Continuous Functions
I turn with terror and horror from this lamentable scourge
of continuous functions with no derivatives.Footnote 28
An auxiliary result from operations research is described, which will prove fruitful for the ensuing theoretical developments.
Consider N + 1 scalar functions φ i(x, z) = f i(x∥z i) − f i(z) \((i \in \mathbb {N})\) and \(\varphi _{N+1} (x,z) = \sum \limits _{i \in \mathbb {N}} \left [ f_i(x) - f_i(z) \right ]\) that are defined on the Cartesian product X ×Z; from this point on, all strategy profiles \(x=(x_1,\ldots ,x_N) \in \mathrm {X} = \prod _{i \in \mathbb {N}} \mathrm {X}_i \subset \mathbb {R}^{n}\) \((n = \sum _{i \in \mathbb {N}} n_i)\), and also x i, z i ∈Xi \((i \in \mathbb {N})\), \(z=(z_1,\ldots ,z_N) \in \mathrm {Z} = \mathrm {X} \subset \mathbb {R}^{n}\) (recall that (x∥z i) = (x 1, …, x i−1, z i, x i+1, …, x N)).
Lemma 2.8.1
If the N + 1 scalar functions φ j(x, z) \((j\!=\!1,\ldots ,N,N\!+\!1\!)\) are continuous on X ×Z while the sets X and Z are compact \((\mathrm {X},\mathrm {Z} \in \mathrm {comp} \: \mathbb {R}^{n})\), then the function
is also continuous on X ×Z.
The proof of a more general result can be found in many textbooks on operations research, e.g., [136, p. 54]; it was included even in textbooks on convex analysis [46, p. 146]. Note that function (2.8.1) is called the Germeier convolution of the functions φ j(x, z) \((j\!=\!1,\ldots , N\!+\!1\!).\)
Finally, note that our choice of Hermite’s quote on “terror and horror” refers to the fact that although each of the functions φ j(x, z) can be differentiable, in general this is not necessarily true for the function φ(x, z) defined by (2.8.1).
2.8.2 Reduction to Saddle Point Design
The result presented in this section is the pinnacle of our book.
Thus, using the payoff functions f i(x) of game (2.3.1), construct the Germeier convolution
with the domain of definition X × (Z = X).
A saddle point (x 0, z B) ∈X ×Z of the scalar function φ(x, z) in the zero-sum two-player game
is defined by the chain of inequalities
Theorem 2.8.1
If in the zero-sum two-player game Γ a, there exists a saddle point (x 0, z B), then the minimax strategy z B is a Berge–Pareto equilibrium in noncooperative game (2.3.1).
Proof
In view of (2.8.2), the first inequality in (2.8.3) with z = x 0 shows that φ(x 0, x 0) = 0. By (2.8.3) and transitivity, for all x ∈X,
Hence, for each \(i \in \mathbb {N}\) and all x ∈X,
Hence, for all x ∈X we have
Since the first N inequalities in (2.8.4) hold for all x ∈X, the strategy profile x B = z B satisfies the Berge equilibrium requirements (2.3.2) in the game Γ. The last equality in (2.8.4) where x ∈XB (the set of Berge equilibria) is a sufficient condition [152, p. 71] for x B = z B to be a Pareto-maximal alternative in the N-criteria choice problem \(\langle \mathrm {X}^{\mathrm {B}}, \{f_i(x)\}_{i \in \mathbb {N}} \rangle \). Thus, by Definition 2.4.1, the resulting strategy profile z B ∈X is a Berge–Pareto equilibrium in game (2.3.1). \(\blacksquare \)
Remark 2.8.1
Theorem 2.8.1 suggests the following design method of a Berge–Pareto equilibrium in the noncooperative game (2.3.1):
first, construct the function φ(x, z) using formula (2.8.2);
second, find the saddle point (x 0, z B) of the function φ(x, z) from the chain of inequalities (2.8.3).
Then the resulting strategy profile z B ∈X is a Berge–Pareto equilibrium in game (2.3.1).
2.8.3 Germeier Convolution
The world of curves has a richer texture than the world of points.
It has been left for the twentieth century to penetrate into this full richness.
—WienerFootnote 29
Let us associate with game (2.3.1) the N-criteria choice problem
where \(\mathrm {X} =\prod _{i\in \mathbb {N}}\mathrm {X}_i\) is the set of all admissible alternatives and f(x) = (f 1(x), …, f N(x)) is the N-dimensional vector criterion. In the problem Γν, a decision maker (DM) seeks to choose an alternative x ∈X in order to maximize the values of all N criteria (objective functions) f 1(x), …, f N(x).
2.8.3.1 Necessary and Sufficient Conditions
In Definition 2.4.1 we have used the following notion of a vector optimum for the problem Γν.
Definition 2.8.1
An alternative x P ∈X is called Pareto-maximal in the problem Γν if, for any x ∈X, the combined inequalities
with at least one strict inequality, are inconsistent. An alternative x S ∈X is called Slater-maximal in the problem Γν if, for any x ∈X, the combined strict inequalities
are inconsistent.
In this section, we are dealing with the Germeier convolution
where the constants μ i belong to the set M of positive vectors from \(\mathbb {R}^N\) (sometimes, with the unit sum of their components).
Note that if f i(x) = −ψ i(x), then formula (2.8.5) yields the standard Germeier convolution (with φ(x) = −ψ(x)) given by
since
Most applications employ Germeier convolutions of two types:
where a i = const > 0 are convolution parameters, i = 1, …, N, and
where μ i = const > 0 are convolution parameters, i = 1, …, N. Clearly, the transition from the first form to the second can be performed by the change of variables \(\mu _i = \frac {1}{a_i}\).
The following results were obtained in multicriteria choice theory.
Germeier’s theorem ([152, p. 66])
. Consider the N-criteria choice problem
and assume that the objective functions f i(x) are positive for all x ∈X and \(i\in \mathbb {N}\).
An alternative x S ∈X is Slater-maximal in Γ ν if and only if there exists a vector μ = (μ 1, …, μ N) ∈M such that
For the Slater-maximal alternatives x S ∈X, let \(\mu = \mu ^{\mathrm {S}} = (\mu _1^{\mathrm {S}},\ldots ,\mu _N^{\mathrm {S}})\), where
which leads to
Recall that M is the set of positive vectors \(\mu = (\mu _1,\ldots ,\mu _N) \in \mathbb {R}^N\) (possibly with the unit sum of components). The next result is a useful generalization of Germeier’s theorem.
Corollary 2.8.1 ([152, p. 67])
. Suppose x S ∈X and ζ i(y) \((i\in \mathbb {N})\) are increasing functions of the variable \(y \in \mathbb {R}\) that satisfy
An alternative x S is Slater-maximal in the multicriteria choice problem Γ ν if and only if
Corollary 2.8.2 ([152, p. 68])
. An alternative x S is Slater-maximal in the multicriteria choice problem Γ ν if and only if
Finally, a Pareto-maximal alternative x P in the problem Γ ν has the following property.
Proposition 2.8.1 ([152, p. 72])
. Let x P ∈ X and f i(x P) > 0 \((i\in \mathbb {N})\). An alternative x P is Pareto-maximal in the multicriteria choice problem Γ ν if and only if there exists a vector
such that f(x P) yields the maximum point of the function \(\sum _{i\in \mathbb {N}} f_i(x)\) on the set
2.8.3.2 Geometrical Interpretation
Consider the Germeier convolution in the case of two criteria in the choice problem Γν, i.e., f(x) = (f 1(x), f 2(x)). Assume that at some point A = (f 1(x A), f 2(x A)) one has
for some parameter values, i.e., \(f_1(x^A) = \frac {1}{\mu _1}\) and \(f_2(x^A) = \frac {1}{\mu _2}\) (see Fig. 2.2). Then the Germeier convolution takes the form
Then the following relations hold on the rays originating from the point A parallel to the coordinate axes:
-
1.
\(\mu _1 f_1(x) \geqslant 1\) and μ 1f 2(x) = 1 on the horizontal ray, or
-
2.
μ 1f 1(x) = 1 and \(\mu _1 f_2(x) \geqslant 1\) on the vertical ray.
Hence, \(\min _{i\in \mathbb {N}} \mu _i f_i(x) = 1\) on these rays. Consequently, the contour lines of the Germeier convolution coincide with the boundaries of the cone \(\{ f(x^A) + \mathbb {R}^2_{+} \}\), where \(\mathbb {R}^2_{+} = \{ f=(f_1,f_2) \mid f_i \geqslant 0 \: (i=1,2) \}\). In the same way, the contour lines of mini ∈{1;2}μ if i(x) = γ are defined by the vertical and horizontal rays originating from the point \(f(\widetilde {x}) =(f_1(\widetilde {x}),f_2(\widetilde {x}))\), where \(\mu _1 f_1(\widetilde {x}) = \mu _2 f_2(\widetilde {x}) = \gamma \). In other words, the contour lines of the Germeier convolution mini ∈{1;2}μ if i(x) = γ form the boundaries of the cone \(\{ f(\widetilde {x}) + \mathbb {R}^2_{+} \}\), where \(f(\widetilde {x}) = \gamma f(x^A)\).
In the general case of N criteria, the level surfaces form the boundaries of the cone \(\{ f(\widetilde {x}) + \mathbb {R}^N_{+} \}\), where \(f(\widetilde {x})\) is any point satisfying the relation \(\min _{i \in \mathbb {N}} \mu _i f_i(x) = \gamma = \mathrm {const} > 0\) \((i\in \mathbb {N})\). Therefore, the level surfaces of the Germeier convolution are the boundaries of the cone of points that dominate its vertex.
It is geometrically obvious that x S is a Slater-maximal alternative in the bi-criteria choice problem Γν (N = 2) if and only if the interior of the orthant \(\mathbb {R}^N_{+}\) shifted to the point f(x S) does not intersect f(X) (see Fig. 2.3).
2.9 Mixed Extension of a Noncooperative Game
The value of pure existence proofs consists precisely in that
the individual construction is eliminated by them and that
many different constructions are subsumed under one fundamental idea,
so that only what is essential to the proof stands out clearly;
brevity and economy of thought are the raison d’être
of existence proofs…To prohibit existence proofs…
is tantamount to relinquishing the science of mathematics altogether.
—HilbertFootnote 30
2.9.1 Mixed Strategies and Mixed Extension of a Game
The theory of probabilities is at bottom nothing but common sense
reduced to calculus; it enables us to appreciate with exactness that
which accurate minds feel with a sort of instinct for which ofttimes
they are unable to account…It teaches us to avoid the illusions
which often mislead us; …there is no science more worthy of
our contemplations nor a more useful one for admission to our system of
public education.
—LaplaceFootnote 31
The mixed extension of a game that includes mixed strategies and profiles as well as expected payoffs is formalized.
Let us, consider the noncooperative N-player game (2.3.1). For each compact set \(\mathrm {X}_i \subset \mathbb {R}^{n_i}\) \((i \in \mathbb {N})\), consider the Borel σ-algebra \(\mathcal {B}(\mathrm {X}_i)\), i.e., the minimal σ-algebra that contains all closed subsets of the compact set Xi (recall that a σ-algebra is closed under taking complements and unions of countable collections of sets).
Assuming that there exist no Berge–Pareto equilibria x B (see Definition 2.4.1) in the class of pure strategies x i ∈Xi \((i \in \mathbb {N})\), we will extend the set Xi of pure strategies x i to the mixed ones, using the approach of Borel [204], von Neumann [261], and Nash [257] and their followers [192, 194, 195, 197, 198]. Next, the idea is to establish the existence of (properly formalized) mixed strategy profiles in game (2.3.1) that satisfy the requirements of a Berge–Pareto equilibrium (an analog of Definition 2.4.1).
Thus, we use the Borel σ-algebras \(\mathcal {B}(\mathrm {X}_i)\) for the compact sets Xi \((i \in \mathbb {N})\) and the Borel σ-algebra \(\mathcal {B}(\mathrm {X})\) for the set of strategy profiles \(\mathrm {X}= \prod _{i \in \mathbb {N}}\mathrm {X}_i\), so that \(\mathcal {B}(\mathrm {X})\) contains all Cartesian products of arbitrary elements of the Borel σ-algebras \(\mathcal {B}(\mathrm {X}_i)\) \((i \in \mathbb {N})\).
In accordance with mathematical game theory, a mixed strategy ν i(⋅) of player i will be identified with a probability measure on the compact set Xi. By the definition in [122, p. 271] and notations in [108, p. 284], a probability measure is a nonnegative scalar function ν i(⋅) defined on the Borel σ-algebra \(\mathcal {B}(\mathrm {X}_i)\) of all subsets of the compact set \(\mathrm {X}_i \subset \mathbb {R}^{n_i}\) that satisfies the following conditions:
-
1.
\(\nu _i(\bigcup \nolimits _k Q^{(i)}_k) = \bigcup \nolimits _k \nu _i(Q^{(i)}_k)\) for any sequence \(\{Q^{(i)}_k\}_{k=1}^{\infty }\) of pairwise disjoint elements from \(\mathcal {B}(\mathrm {X}_i)\) (countable additivity);
-
2.
ν i(Xi) = 1 (normalization), which yields \(\nu _i(Q^{(i)}) \leqslant 1\) for all \(Q^{(i)} \in \mathcal {B}(\mathrm {X}_i)\).
Let {ν i} denote the set of all mixed strategies of player i \((i \in \mathbb {N})\).
Also note that the product measures ν(dx) = ν 1(dx 1) ⋯ ν N(dx N), see the definitions in [122, p. 370] (and the notations in [108, p. 123]), are probability measures on the strategy profile set X. Let {ν} be the set of such probability measures (strategy profiles). Once again, we emphasize that in the design of the product measure ν(dx) the role of a σ-algebra of subsets of the set X1 ×⋯ ×XN = X is played by the smallest σ-algebra \(\mathcal {B}(\mathrm {X})\) that contains all Cartesian products Q (1) ×⋯ × Q (N), where \(Q^{(i)} \in \mathcal {B}(\mathrm {X_i})\) \((i \in \mathbb {N})\). The well-known properties of probability measures [41, p. 288], [122, p. 254] imply that the sets of all possible measures ν i(dx i) \((i \in \mathbb {N})\) and ν(dx) are weakly closed and weakly compact ([122, pp. 212, 254], [180, pp. 48, 49]). As applied, e.g., to {ν}, this means that from any infinite sequence {ν (k)} (k = 1, 2, …) one can extract a subsequence \(\{\nu ^{(k_j)}\}\) (j = 1, 2, …) that weakly converges to a measure ν (0)(⋅) ∈{ν}. In other words, for any continuous scalar function φ(x) on X, we have
and ν (0)(⋅) ∈{ν}. Owing to the continuity of φ(x), the integrals \(\int \limits _{\mathrm {X}} \varphi (x)\nu (dx)\) (the expectations) are well-defined; by Fubini’s theorem,
and the order of integration can be interchanged.
Let us associate with game (2.3.1) in pure strategies its mixed extension
where, like in (2.3.1), \(\mathbb {N}\) is the set of players while {ν i} is the set of mixed strategies ν i(⋅) of player i. In game (2.9.1), each conflicting party \(i \in \mathbb {N}\) chooses its mixed strategy ν i(⋅) ∈{ν i}, thereby forming a mixed strategy profile ν(⋅) ∈{ν}; the payoff function of each player i, i.e., the expectation
is defined on the set {ν}.
For game (2.9.1), the notion of a Berge–Pareto equilibrium x ∗ (see Definition 2.4.1) has the following analog.
Definition 2.9.1
A mixed strategy profile ν ∗(⋅) ∈{ν} is called a Berge–Pareto equilibrium in the mixed extension (2.9.1) (equivalently, a Berge–Pareto equilibrium in mixed strategies in game (2.3.1)) if
first, the profile ν ∗(⋅) is a Berge equilibrium in game (2.9.1), i.e.,
and second, ν ∗(⋅) is a Pareto-maximal alternative in the N-criteria choice problem
i.e., for all ν(⋅) ∈{ν B}, the system of inequalities
with at least one strict inequality, is inconsistent.
Here and in the sequel,
\(\nu _{\mathbb {N} \setminus \{i\}}(dx_{\mathbb {N} \setminus \{i\}}) = \nu _1(dx_1) \cdots \nu _{i-1}(dx_{i-1}) \nu _{i+1}(dx_{i+1}) \cdots \nu _{N}(dx_{N})\), \((\nu \|\nu ^{*}_i)=(\nu _1(dx_1) \cdots \nu _{i-1}(dx_{i-1}) \nu ^{*}_{i}(dx_{i}) \nu _{i+1}(dx_{i+1}) \cdots \nu _{N}(dx_{N}))\), \(\nu ^{*}(dx)=\nu ^{*}_1(dx_1)\cdots \nu ^{*}_N(dx_N)\), \(\{ \nu _{\mathbb {N} \setminus \{i\}} \} = \{ \nu _{\mathbb {N} \setminus \{i\}} (\cdot ) \}\); in addition, {ν B(⋅)} denotes the set of Berge equilibria ν B(⋅), i.e., the strategy profiles that satisfy (2.9.2) with ν ∗ replaced by ν B. Let {ν ∗} be the set of mixed strategy profiles in game (2.9.1) that are given by the two requirements of Definition 2.9.1.
The following sufficient condition for Pareto maximality is obvious, see the statement below.
Remark 2.9.1
A mixed strategy profile ν ∗(⋅) ∈{ν} is Pareto-maximal in the game \(\tilde {\Gamma }_\nu = \langle \{\nu ^{\mathrm {B}}\}, \{f_i(\nu )\}_{i \in \mathbb {N}}\rangle \) if
2.9.2 Préambule
Proposition 2.9.1
In game (2.3.1), suppose the sets Xi are compact, the payoff functions f i(x) are continuous on X = X1 ×⋯ ×XN, and the set of mixed strategy Berge equilibria {ν B} that satisfy (2.9.2) with ν ∗ replaced by ν B is nonempty.
Then {ν B} is a weakly compact subset of the set of mixed strategy profiles {ν} in game (2.9.1).
Proof
To establish the weak compactness of the set {ν B}, take an arbitrary scalar function ψ(x) that is continuous on the compact set X and an infinite sequence of mixed strategy profiles
in game (2.9.1). Inclusion (2.9.3) (hence, {ν B}⊂{ν}) implies {ν (k)(⋅)}⊂{ν}. As mentioned earlier, the set {ν} is weakly compact; hence, there exist a subsequence \(\{\nu ^{(k_j)} (\cdot )\}\) and a measure ν (0)(⋅) ∈{ν} such that
We will show that ν (0)(⋅) ∈{ν B(⋅)}by contradiction. Assume that ν (0)(⋅) does not belong to {ν B}. Then for sufficiently large j, one can find a number \(i \in \mathbb {N}\) and a strategy profile \(\bar {\nu } (\cdot ) \in \{\nu \}\) such that
which clearly contradicts the inclusion \(\{\nu ^{(k_j)} (\cdot )\} \in \{\nu ^{\mathrm {B}}\}\).
Thus, we have proved the requisite weak compactness. \(\blacksquare \)
Corollary 2.9.1
In a similar fashion, one can prove the compactness (closedness and boundedness) of the set
in the criteria space \(\mathbb {R}^{N}\).
Proposition 2.9.2
If in game (2.9.1) the sets \(\mathrm {X}_i \in \mathrm {comp}\:\mathbb {R}^{n_i}\) and f i(⋅) ∈ C(X) \((i \in \mathbb {N})\), then the function
satisfies the inequality
for any μ(⋅) ∈{ν}, ν(⋅) ∈{ν}; recall that
Proof
Indeed, from (2.9.4) we have \(N\!+\!1\) inequalities of the form
for each x, z ∈ X. Integrating both sides of these inequalities with respect to an arbitrary product measure μ(dx)ν(dz) yields
for all μ(⋅) ∈ {ν}, ν(⋅) ∈ {ν} and each \(r\!=1,\ldots ,N\!+\!1\). Consequently,
which proves (2.9.5). \(\blacksquare \)
Remark 2.9.2
In fact, formula (2.9.5) generalizes the well-known property of maximization: the maximum of a sum does not exceed the sum of the maxima.
2.9.3 Existence Theorem
Good mathematicians see analogies.
Great mathematicians see analogies between analogies.
—BanachFootnote 32
The central result of Chap. 2—the existence of a Berge–Pareto equilibrium in mixed strategies—is established.
Theorem 2.9.1
If in game (2.3.1) the sets \(\mathrm {X}_i \in \mathrm {cocomp}\:\mathbb {R}^{n_i}\) and f i(⋅) ∈ C(X) \((i \in \mathbb {N})\), then there exists a Berge–Pareto equilibrium in mixed strategies.
Proof
Consider the auxiliary zero-sum two-player game
In the game Γa, the set X of strategies x chosen by player 1 (which seeks to maximize φ(x, z)) coincides with the set of strategy profiles of game (2.3.1); the set Z of strategies z chosen by player 2 (which seeks to minimize φ(x, z)) coincides with the same set X. A solution of the game Γa is a saddle point (x 0, z B) ∈X ×X; for all x ∈X and each z ∈X, it satisfies the chain of inequalities
Now, associate with the game Γa its mixed extension
where {ν} and {μ} = {ν} denote the sets of mixed strategies ν(⋅) and μ(⋅) of players 1 and 2, respectively. The payoff function of player 1 is the expectation
The solution of the game \(\tilde {\Gamma }^{\mathrm {a}}\) (the mixed extension of the game Γa) is also a saddle point (μ 0, ν ∗) defined by the two inequalities
for any ν(⋅) ∈{ν} and μ(⋅) ∈{ν}.
Sometimes, this pair (μ 0, ν ∗) is called the solution of the game Γa in mixed strategies.
In 1952, Gliksberg [30] established the existence of a mixed strategy Nash equilibrium for a noncooperative game of \(N \geqslant 2\) players. Applying this existence result to the zero-sum two-player game Γa as a special case, we obtain the following statement. In the game Γa, let the set \(\mathrm {X} \!\subset \! \mathbb {R}^{n}\) be nonempty and compact and let the payoff function φ(x, z) of player 1 be continuous on X ×X (note that the continuity of φ(x, z) is assumed in Lemma 2.8.1). Then the game Γa has a solution (μ 0, ν ∗) defined by (2.9.7), i.e., there exists a saddle point in mixed strategies.
In view of (2.9.4), inequalities (2.9.7) can be written as
for all ν(⋅) ∈{ν} and μ(⋅) ∈{ν}. Using the measure \(\nu _i(dz_i) = \mu ^0_i(dx_i)\) \((i \!\in \! \mathbb {N})\) (so that ν(dz) = μ 0(dx)) in the expression
we obtain φ(μ 0, μ 0) = 0 due to (2.9.6). Similarly, φ(ν ∗, ν ∗) = 0, and it follows from (2.9.7) that
The condition φ(μ 0, μ 0) = 0 and the chain of inequalities (2.9.7) give, by transitivity,
In accordance with Proposition 2.9.2, then we have
Therefore, for all j = 1, …, N + 1,
Consider two cases as follows.
Case I (j = 1, …, N). Here, by (2.9.10), (2.9.6) and the normalization of ν(⋅), we arrive at
By Definition 2.9.1, ν ∗(⋅) is a Berge equilibrium in mixed strategies in game (2.3.1).
Case II (j = N + 1). Again, using (2.9.10), (2.9.6) and the normalization of ν(⋅) and μ(⋅), we have
In accordance with Remark 2.9.1, the mixed strategy profile ν ∗(⋅) ∈ {ν} of game (2.3.1) is a Pareto-maximal alternative in the multicriteria choice problem
Thus, we have proved that the mixed strategy profile ν ∗(⋅) in game (2.3.1) is a Berge equilibrium that satisfies Pareto maximality. Hence, by Definition 2.9.1, the mixed strategy profile ν ∗(⋅) is a Berge–Pareto equilibrium in game (2.3.1). \(\blacksquare \)
2.10 Linear-Quadratic Two-Player Game
Verba docent, exempla trahunt.Footnote 33
Readers who studied Lyapunov’s stability theory surely remember algebraic coefficient criteria. The whole idea of such criteria is to establish the stability of unperturbed motion without solving a system of differential equations, by using the signs of coefficients and/or their relationships. In this section of the book, we are endeavoring to propose a similar approach to equilibrium choice in noncooperative linear-quadratic two-player games. More specifically, our approach allows one to decide about the existence of a Nash equilibrium and/or a Berge equilibrium in these games based on the sign of quadratic forms in the payoff functions of players.
2.10.1 Preliminaries
Consider a noncooperative linear-quadratic two-player game described by
A distinctive feature of Γ2 is the absence of constraints on the strategy sets Xi: the strategies of player i can be any column vectors of dimension n i, i.e., elements of the n i-dimensional Euclidean space \(\mathbb {R}^{n_i}\) with the standard Euclidean norm ∥⋅∥ and the scalar product. Let the payoff function of player i (i = 1, 2) have the form
where A i and C i are constant symmetric matrices, B i is a constant rectangular matrix, and a i and c i are constant vectors, all of compatible dimensions; ′ denotes transposition; det A denotes, the determinant of a matrix A. Henceforth, A < 0 \((>, \: \geqslant )\) means that the quadratic form z′Az is negative definite (positive definite, positive semidefinite, respectively). We will adopt the following rules of differentiation of bilinear quadratic forms with respect to the vector argument [19, 27]:
For a scalar function Ψ(x) of a k-dimensional vector argument x, sufficient conditions for
are
where 0k denotes a zero column vector of dimension k.
2.10.2 Berge Equilibrium
For the payoff functions (2.10.1), relations (2.10.3) yield the following sufficient condition for the existence of a Berge equilibrium in the game Γ2.
Proposition 2.10.1
Assume that in the game Γ 2
and
Then the Berge equilibrium \(x^{\mathrm {B}} =(x^{\mathrm {B}}_1,x^{\mathrm {B}}_2)\) is given by
Proof
By definition, a strategy profile \((x^{\mathrm {B}}_1,x^{\mathrm {B}}_2)=x^{\mathrm {B}}\) is a Berge equilibrium in the game Γ2 if and only if
In view of (2.10.3) and(2.10.1), sufficient conditions for the first equality in (2.10.7) to hold can be written as
Similarly, for the second equality in (2.10.7) we obtain
In accordance with (2.10.4), the matrices C 1 and A 2 are negative definite and hence the Berge equilibrium \((x^{\mathrm {B}}_1,x^{\mathrm {B}}_2)\) in the game Γ2 satisfies the linear nonhomogeneous system of matrix equations
Using the chain of implications \(\left [ A_2 < 0 \right ] \; \Rightarrow \; \left [ \mathrm {det} A_2 \neq 0 \right ] \; \Rightarrow \; \left [ \exists A^{-1}_2\right ]\), we multiply the first equation in (2.10.8) on the left by \(A^{-1}_2\) to get
Substituting this expression into the second equation of system (2.10.8), one obtains
or
Here we used the fact that
formula (2.10.11) is easily derived upon multiplying both sides of (2.10.10) on the left by the matrix \(\left [ C_1 - B^{\prime }_1 A^{-1}_2 B_2 \right ]^{-1}.\) Finally, using the resulting expression for \(x^B_2\) in (2.10.9), we arrive at the first equality of system (2.10.6).
In the same fashion, it is possible to solve system (2.10.8) by multiplying the second equation by \(C^{-1}_1\!\!\). This leads to
Proposition 2.10.2
Assume inequalities (2.10.4) and
hold in the game Γ 2 . Then the Berge equilibrium \(x^{\mathrm {B}} =(x^{\mathrm {B}}_1,x^{\mathrm {B}}_2)\) has the form
Remark 2.10.1
System (2.10.8) has a unique solution for A 2 < 0 and C 1 < 0. The two explicit forms of this solution presented above are equivalent and can be reduced to each other.
2.10.3 Nash Equilibrium
In this section, we derive similar results for the Nash equilibrium in the game Γ2. Instead of (2.10.7), we consider a Nash equilibrium \(x^{\mathrm {e}}=(x^{\mathrm {e}}_1,x^{\mathrm {e}}_2)\) defined by the two equalities
The sufficient conditions for implementing (2.10.13) take the form
The first two conditions give the linear nonhomogeneous system of matrix equations
As in Propositions 2.10.1 and 2.10.2, the conditions A 1 < 0 and C 2 < 0 allow us to establish the following results. The onus probandi Footnote 34 is left to the reader.
Proposition 2.10.3
Assume the inequalities
and
hold in the game Γ 2 . Then the Nash equilibrium \(x^{\mathrm {e}}=(x^{\mathrm {e}}_1,x^{\mathrm {e}}_2)\) has the form
Proposition 2.10.4
Assume inequalities (2.10.14) and
are satisfied in the game Γ 2 . Then the Nash equilibrium \(x^{\mathrm {e}}=(x^{\mathrm {e}}_1,x^{\mathrm {e}}_2)\) has the form
2.10.4 Auxiliary Lemma
Despite the negativa non probantur Footnote 35 principle of Roman law, we will rigorously obtain a result useful for discarding the games without any Berge and/or Nash equilibria.
Lemma 2.10.1
In the game Γ 2 with A 1 > 0, there exists no \(\bar {x}_1\) such that, for each fixed x 2,
In other words, the payoff function f 1 of the first player is not maximized in this game.
Proof
Let us “freeze” a certain strategy \(x_2 \in \mathbb {R}^{n_2}\) of the second player. Then the payoff function of the first player can be written in the form
where the column vector φ(x 2) of dimension n 1 and the scalar function ψ(x 2) depend on the frozen value x 2 only.
By the hypothesis of Lemma 2.10.1, the symmetric matrix A 1 is positive definite. In this case, the characteristic equation \(\mathrm {det} [A_1 - E_{n_1}\lambda ] = 0\) (where \(E_{n_1}\) denotes an identity matrix of dimensions n 1 × n 1) has n 1 positive real roots owing to symmetry and, in addition,
where λ ∗ > 0 is the smallest root among them. Thus, maximum (2.10.17) is not achieved if, for any large value m > 0, there exists a strategy \(x_1(m,x_2 )\in \mathbb {R}^{n_1}\) such that
Under (2.10.18), this inequality holds if
Let us construct a solution x 1(m, x 2) of inequality (2.10.19) in the form
where the constant β > 0 will be specified below and e 1’s the n 1-dimensional vector with all components equal to 1.
Substituting (2.10.20) into (2.10.19) yields the following inequality for β:
Hence, for any constant
and corresponding strategy \(x_1(m,x_2 ) = \beta e_{n_1}\) of the first player, we have
Thus, maximum (2.10.17) does not exist. \(\blacksquare \)
Remark 2.10.2
In this case, the game Γ2 with a matrix A 1 > 0 admits no Nash equilibria. In combination with Proposition 2.10.1, this shows that
-
1.
the game Γ2 with matrices \(A_1\!>\!0\) or (and) \(C_2\! >\!0\) and also \(A_2\!<\!0\), \(C_1\!<\!0\) that satisfies condition (2.10.5) has no Nash equilibria, but does have a Berge equilibrium defined by (2.10.6).
In a similar way, it is easy to obtain the following.
-
2.
If \(A_2\!<\!0\), \(C_1\! <\!0\), condition (2.10.5) or (2.10.12) holds and also \(A_1\!>\!0\) or (and) \(C_2\! >\!0\), then the game Γ2 has a Berge equilibrium only.
-
3.
If \(A_1\!<\!0\), \(C_2\! <\!0\), condition (2.10.15) or (2.10.16) holds and also \(A_2\!>\!0\) or (and) \(C_1\! >\!0\), then the game Γ2 has a Nash equilibrium only.
-
4.
If \(A_1\!>\!0\) or (and) \(C_2\! >\!0\) and also \(A_2\!>\!0\) or (and) \(C_1\! >\!0\), then the game Γ2 has none of these equilibria.
-
5.
If A 2 < 0, C 1 < 0, A 1 < 0, C 2 < 0 and also conditions (2.10.5) or (2.10.12) and (2.10.15) or (2.10.16) hold, then the game Γ2 has both types of equilibrium.
2.10.5 Concluding Remarks
Thus, we have considered the noncooperative linear-quadratic two-player game without constraints \((\mathrm {X}_i = \mathbb {R}^{n_i}, i=1,2)\) and with the payoff functions
Here ′ denotes transposition; A i and C i are constant symmetric matrices of dimensions n 1 × n 1 and n 2 × n 2, respectively; B i is a constant rectangular matrix of dimensions n 1 × n 2; finally, a i and c i are constant vectors of dimensions n 1 and n 2, respectively (i = 1, 2).
Based on Propositions 2.10.1–2.10.4, we introduce the following coefficient criteria for the existence of Nash and Berge equilibria in the game Γ2. Par acquit de conscience, Footnote 36 they are presented in form of Table 2.1.
How should one use Table 2.1? Just follow the three simple steps indicated below.
-
Step1.
First, check the signs of the quadratic forms with the matrices A 1, A 2, C 1, and C 2. For example, suppose A 1 < 0, C 2 < 0 (both matrices are negative definite), while A 2 > 0 (i.e., A 2 is positive definite).
-
Step2.
Find the corresponding row in Table 2.1 (in our case, the conditions A 1 < 0, C 2 < 0 and A 2 > 0 are in row 3); then verify the nondegeneracy of the matrix (2.10.15) in column 5, i.e., the condition \(\mathrm {det} \left [ C_2 - B^{\prime }_2 A^{-1}_1 B_1 \right ] \neq 0.\)
-
Step3.
As shown in columns 6 and 7 of Table 2.1, the game Γ2 with these matrices has no Berge equilibria, but has a Nash equilibrium for any matrices C 1, B i and any vectors a i, b i of compatible dimensions. The explicit form of this Nash equilibrium is given in Proposition 2.10.3.
Notes
- 1.
French “He who thinks he has the power to content the world greatly deceives himself, but he who thinks that the world cannot be content with him deceives himself yet more.” François de La Rochefoucauld (1613–1680) was a French classical writer; a quote from Réflexions ou Sentences et Maximes morales (1665).
- 2.
Latin “To each according to its own merits; to each his/her own.” This phrase appeared in philosophical dialogs and treatises On Duties 1, 5, 14, and Tusculan Disputations, Vol. 22, by Marcus Tullius Cicero (102–43 BC), a Roman statesman, lawyer, scholar, and writer.
- 3.
Latin “Not many, but much,” meaning not quantity but quality. This phrase belongs to Plinius the Younger (62–114 A.D.); see Letters, VII, 9.
- 4.
Osip E. Mandelshtam, (1891–1938), was a major Russian poet, prose writer, and literary essayist.
- 5.
French, meaning narrow-mindedness and a lack of understanding or even interest in the world beyond one’s own town’s boundaries.
- 6.
Latin, meaning a feature by which two subclasses of the same class of named objects can be distinguished.
- 7.
Latin “Every man is the artisan of his own fortune.” This phrase goes back to Appius Claudius Caecus (4–3 centuries BC), an outstanding statesman, legal expert and author of early Rome who was one of the first notable personalities in Roman history.
- 8.
French, meaning that all things in the worlds are interconnected.
- 9.
Latin “From the chair,” used with regard to statements made by people in positions of authority.
- 10.
There also exist other approaches to conflict analysis: descriptive, which is to find the resulting collections of players’ actions (the so-called strategy profiles) in a given conflict; constructive, which is to implement the desired (e.g., optimal) strategies in a given conflict; predictive, which is to forecast the actual result (outcome) of a given conflict.
- 11.
French “Every man for himself, every man to himself.”
- 12.
German “One for all and all for one.”
- 13.
Latin “For discussion.”
- 14.
Latin “Who? What? Where? Who helped? Why? How? When?”; a well-known system of seven questions for crime investigation suggested by Roman rhetorician Quintilian, Latin in full Marcus Fabius Quintilianus, (appr. 35–100 A.D.).
- 15.
French “We would yearn for very few things if we clearly understood what we wanted.” A quote from Réflexions ou Sentences et Maximes Morales by F. de La Rochefoucauld.
- 16.
Latin “To each his own,” or “May all get their due”; also, see the epigraph to Sect. 2.2.1.
- 17.
French “From each according to his ability, to each according to his needs.”
- 18.
Italian, “The policy of his/her own bell tower.” Used to describe narrow-mindness and commitment to local interests.
- 19.
John Forbes Nash, Jr. Born June 13, 1928, in Bluefield, West Virginia. Successfully graduated from the Carnegie Institute of Technology (now, Carnegie Mellon University) with bachelor’s and master’s degrees in mathematics. Richard Duffin, Nash’s undergraduate advisor at the Carnegie Institute of Technology, gave him a brief characterization, “He is a mathematical genius.” In 1948 Nash started his postgraduate study at Princeton University, where he was particularly influenced by International Economy, the faculty course of J. von Neumann, and by the famous book Theory of Games and Economic Behavior (1944), written by von Neumann together with O. Morgenstern. In 1949 Nash presented his thesis on equilibrium solutions of noncooperative games; after 45 years—in 1994—he was awarded the Noble Prize in Economic Sciences for that research. From 1951 to 1959 worked at the Cambridge at Massachusetts Institute of Technology (MIT). In 1958 Fortune called Nash “America’s brilliant young star of the ‘new mathematics.”’ In 1959 moved to California to work for the RAND Corporation and became a leading expert in the Cold War. Since 1959 suffered from a mental disorder (completely overcame the disease by 1980, to the great astonishment of doctors). Since 1980 again worked at Princeton University as a consulting professor. Was killed in a car crash on May, 24, 2015, at the age of 86. Throughout the world, Nash is well-known through R. Howard’s movie A Beautiful Mind (2001, featuring R. Crowe) based on S. Nasar’s book Beautiful Mind: The Life of Mathematical Genius and Nobel Laureate John Nash. The movie received four Oscars and the Golden Globe.
- 20.
“And will not let true merit miss its due, Remembering always rather good than evil.” A quote from Tartuffe, Scene VII, a famous theatrical comedy by Molière (1622–1873).
- 21.
Naum Ya. Vilenkin, (1920–1991), was a Soviet mathematician and student of A. G. Kurosh, who contributed to general algebra, topology, real-variable theory and functional analysis. A quote from Kvant, 1975, no. 8, p. 2.
- 22.
James P. Pierpont, (1866–1938), was an American mathematician. Known for research in the field of real and complex variable functions.
- 23.
Veniamin F. Kagan, (1860–1953), was a Russian and Soviet mathematician. A quote translated into English from Kvant, 1975, no. 6, p. 16.
- 24.
Cassius Jackson Keyser, (1862–1947), was an American mathematician of pronounced philosophical inclinations. A quote from On Mathematics and Mathematicians, R.E. Moritz, Ed., New York: Dover, 1958, pp. 360–361.
- 25.
From a PBS interview with John F. Nash.
- 26.
Jean Le Rond d’Alembert, (1717–1783), was a French mathematician, philosopher, and writer.
- 27.
Walter Warwick Sawyer, (1911–2008), was a British mathematician, mathematics educator and author, who popularized mathematics on several continents.
- 28.
Richard Courant, (1888–1972), was a German-born American mathematician, educator and scientific organizer who made significant advances in the calculus of variations. A quote from The Australian Mathematics Teacher, vols. 39–40, Australian Association of Mathematics Teachers, 1983, p. 3.
- 29.
From a letter of French mathematician Charles Hermite, (1822–1901), to Dutch mathematician Thomas Joannes Stieltjes, (1856–1894), written in 1893.
- 30.
Norbert Wiener, (1894–1964), was an outstanding American mathematician and philosopher, the father of cybernetics. A quote from his book I Am a Mathematician: the Later Life of a Prodigy, MIT Press, 1964.
- 31.
David Hilbert, (1862–1943), was a German mathematician who axiomatized geometry and contributed substantially to the establishment of the formalistic foundations of mathematics. Recognized as one of the most influential and universal mathematicians of the 19th and early 20th centuries. A quote from the book Hilbert by C.B. Reid, Springer, 1996.
- 32.
Pierre-Simon, marquis de Laplace, (1749–1827), was a French scholar who made important contributions to the development of mathematics, statistics, physics and astronomy. An English translation of a quote from his book Théorie Analytique des Probabilités, 1795.
- 33.
Stefan Banach, (1892–1945), was a Polish mathematician who founded modern functional analysis and helped contributed to the development of the theory of topological vector spaces. Generally considered one of the most important and influential mathematicians of the twentieth century.
- 34.
Latin “Words instruct, illustrations lead.”
- 35.
Latin “The burden of proof.”
- 36.
Latin “Negative statements are not proved.”
- 37.
French “For our peace of mind.”
References
Vaisman, K.S., Berge Equilibrium, Cand. Sci. (Phys.-Math.) Dissertation, St. Petersburg: St. Petersburg. Gos. Univ., 1995.
Vaisman, K.S., Berge Equilibrium, in Lineino-kvadratichnye differentsial’nye igry (Linear-Quadratic Differential Games), Zhukovskiy, V.I. and Chikrii, A.A., Eds., Kiev: Naukova Dumka, 1994, pp. 119–143.
Voevodin, V.V. and Kuznetsov, Yu.A., Matritsy i vychisleniya (Matrices and Calculations), Moscow: Nauka, 1984.
Vorobiev, N.N., Issledovanie operatsii, (Operations Research), Moscow: Nauka, 1972.
Vorobiev, N.N., Modern State of Game Theory, Uspekhi Mat. Nauk, 1970, vol. 25, no. 2, pp. 81–140.
Vorobiev, N.N., Teoriya igr dlya ekonomistov-kibernetikov (Game Theory for Economists-Cyberneticians), Moscow: Nauka, 1985.
Gantmacher, F.R., The Theory of Matrices. Vols. 1 and 2, New York: Chelsea Publishing, 1959.
Glicksberg, I.L., A Further Generalization of the Kakutani Fixed Point Theorem, with Application to Nash Equilibrium Points, Proc. Am. Math. Soc., 1952, vol. 3, no. 1, pp. 170–174.
Dunford, N. and Schwartz, J.T., Linear Operators, New York: Interscience, 1958. Translated under the title Lineinye operatory. Obshchaya teoriya, Moscow: Inostrannaya Literatura, 1962, vol. 1.
Dmitruk, A.V., Vypuklyi analiz. Elementarnyi vvodnyi kurs (Convex Analysis. An Elementary Introductory Course), Moscow: Makspress, 2012.
Dubina, I.N., Osnovy teorii ekonomicheskikh igr (Theoretical Foundations of Economic Games), Moscow: Knorus, 2010.
Egorov, A.I., Osnovy teorii upravleniya (Foundations of Control Theory), Moscow: Fizmatlit, 2004.
Zhautykov, O.A., Zhukovskiy, V.I., and Zharkynbaev, S.Zh, Differentsial’nye igry neskol’kikh lits (s zapazdyvaniem vremeni) (Differential Multiperson Games with Time Delay), Almaty: Nauka, 1988.
Zhitomirskii, G.I., Dynamic Problems with Conflict, Extended Abstract of Cand. Sci. Dissertation (Phys.-Math.), Leningrad State Univ., Leningrad, 1989.
Zhukovskiy, V.I., Kooperativnye igry pri neopredelennosti i ikh prilozheniya (Cooperative Games under Uncertainty and Their Applications), Moscow: URSS, 2010, 2nd ed.
Zhukovskiy, V.I., Vvedenie v differentsial’nye igry pri neopredelennosti. Ravnovesie po Neshu (Introduction to Differential Games under Uncertainty. Nash Equilibrium), Moscow: URSS, 2010.
Zhukovskiy, V.I. and Chikrii, A.A., Lineino-kvadratichnye differentsial’nye igry (Linear-Quadratic Differential Games), Kiev: Naukova Dumka, 1994.
Krasovskii, N.N. and Subbotin, A.I., Pozitsionnye differentsial’nye igry (Positional Differential Games), Moscow: Nauka, 1985.
Larets ostroslovov (Casket of Wisecrackers), Moscow: Izdatel’stvo Politicheskoi Literatury, 1991.
Lezhnev, A.V., Dinamicheskoe programmirovanie v ekonomicheskikh zadachakh (Dynamic Programming in Economic Problems), Moscow: BINOM, Laboratoriya Znanii, 2006.
Lyusternik, L.A. and Sobolev, V.I., Elementy funktsional’nogo analiza (Elements of Functional Analysis), Moscow: Nauka, 1969.
Makarov, V.A. and Rubinov, A.M., Matematicheskaya teoriya ekonomicheskoi dinamiki i ravnovesie (Mathematical Theory of Economic Dynamics and Equilibrium), Moscow: Nauka, 1973.
Mashchenko, S.O., Concept of Nash Equilibrium and Its Development, Zh. Obchisloval. Prikladn. Mat., 2012, no. 1(107), pp. 40–65.
Morozov, V.V., Sukharev, A.G., and Fedorov, V.V., Issledovanie operatsii v zadachakh i uprazhneniyakh (Operations Research in Problems and Exercises), Moscow: Vysshaya Shkola, 1986.
Podinovskii, V.V. and Noghin, V.D., Pareto-optimal’nye resheniya mnogokriterial’nykh zadach (Pareto Optimal Solutions of Multicriteria Problems), Moscow: Fizmatlit, 2007.
Rosenmüller, J., The Theory of Games and Markets, Elsevier, 1981.
Ulanov, V.A., On a Differentia Game with Infinite Number of Players, in Problemy teorii igr v obshchikh sistemakh (Problems of Game Theory in General Systems), Yakutsk. Gos. Univ., 1988, pp. 32–39.
Fedorov, V.V., Chislennye metody maksimina (Numerical Methods of Maximin), Moscow: Nauka, 1979.
Harsanyi J.C. and Selten R., A General Theory of Equilibrium Selection in Games, Cambridge: MIT Press, 1988.
Hille, E. and Phillips, R.S., Functional Analysis and Semi-Groups, American Mathematical Society Colloquium Publications, vol. 31, Providence: Am. Math. Soc., 1957. Translated under the title Funktsional’nyi analiz i polugruppy, Moscow: Inostrannaya Literatura, 1962.
Entsiklopediya kibernetiki. Tom 2. (Encyclopeadia of Cybernetics. Vol. 2), Kiev: Glavn. Redakts. Ukrainsk. Sovetsk. Entsiklop., 1974.
Yanovskaya, E.B., Equilibria in General Noncooperative Games and Their Mixed Extensions, in Teoretiko-igrovye voprosy prinyatiya resheniy (Game-Theoretic Issues of Decision-Making), Leningrad: Nauka, 1978, pp. 43–65.
Abalo, K.Y. and Kostreva, M.M., Some Existence Theorems of Nash and Berge Equilibria, Appl. Math. Lett., 2004, vol. 17, pp. 569–573.
Abalo, K.Y. and Kostreva, M.M., Berge Equilibrium: Some Recent Results from Fixed-Point Theorems, Appl. Math. and Comp., 2005, vol. 169, pp. 624–638.
Aumann, R.J. and Peleg, B., Von Neumann–Morgenstern Solutions to Cooperative Games without Side Payments, Bull. Amer. Math. Soc., 1960, vol. 66, pp. 173–179.
Basar, T., A Contrexample in Linear-Quadratic Game: Existence of Nonlinear Nash Solutions, J. Optimiz. Theory and Appl., 1974, vol. 14, no. 4, pp. 425–430.
Berge, C., Théorie générale des jeux ánpersonnes games, Paris: Gauthier Villars, 1957. (Russian translation: Berge, C., Obshchaya teoriya igr neskol’kikh lits, Moscow: Fizmatgiz, 1961).
Bertrand, J., Caleul des probabilities, Paris, 1888.
Larbani, M. and Nessah, R., A Note on the Existence of Berge and Berge–Nash Equilibria, Mathematical Social Sciences, 2008, vol. 55, pp. 258–271.
Milnor, J., Games Against Nature, in Decision Processes, New York: Wiley, 1954, pp. 49–60.
Musy, O., Pottier, A., and Tazdait, T., A New Theorem to Find Berge Equilibria, International Game Theory Review, 2012, vol. 14, no. 1, p. 1250005. DOI:10.1142/S0219198912500053
Nash, J.F., Non-Cooperative Games, Ann. Math., 1951, vol. 54, pp. 286–295.
Von Neumann, J., Zur Theorie der Gesellschaftspiele, Math. Ann., 1928, vol. 100, pp. 295–320.
Radjef, M., Sur l‘existence d‘un équilibre de Berge pour un jeu différential n-personnes (On the Existence of a Berge Equilibrium for an n-Person Differential Game), Cahiers Mathématiques de l‘Université d‘Oran, 1998, no. 1, pp. 89–93.
Shubik, M., Review of C. Berge “General theory of n-person games,” Econometrica, 1961, vol. 29, no. 4, p. 821.
Vaisman, K.S., The Berge Equilibrium for Linear–Quadratic Differential Game, The 3-rd Intern. Workshop on Multiple Criteria Problems under Uncertainty, Orekhovo-Zuevo, Russia, 1994, p. 96.
Vaisman, K.S. and Zhukovskiy, V.I., The Berge Equilibrium under Uncertainty, The 3-rd Intern. Workshop on Multiple Criteria Problems under Uncertainty, Orekhovo-Zuevo, Russia, 1994, pp. 97–98.
Zhukovskiy, V.I., Salukvadze, M.E., and Vaisman, K.S., Berge Equilibrium, Preprint. Tbilisi: Institute of Control Systems, 1994.
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this chapter
Cite this chapter
E. Salukvadze, M., I. Zhukovskiy, V. (2020). Static Case of the Golden Rule. In: The Berge Equilibrium: A Game-Theoretic Framework for the Golden Rule of Ethics. Static & Dynamic Game Theory: Foundations & Applications. Birkhäuser, Cham. https://doi.org/10.1007/978-3-030-25546-6_2
Download citation
DOI: https://doi.org/10.1007/978-3-030-25546-6_2
Published:
Publisher Name: Birkhäuser, Cham
Print ISBN: 978-3-030-25545-9
Online ISBN: 978-3-030-25546-6
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)