Abstract
The paper deals with construction of solutions in dynamical bimatrix games. It is assumed that integral payoffs are discounted on the infinite time interval. The dynamics of the game is subject to the system of differential equations describing the behavior of players. The problem of construction of equilibrium trajectories is analyzed in the framework of the minimax approach proposed by N. N. Krasovskii and A. I. Subbotin in the differential games theory. The concept of dynamical Nash equilibrium developed by A. F. Kleimenov is applied to design the structure of the game solution. For obtaining constructive control strategies of players, the maximum principle of L. S. Pontryagin is used in conjunction with the generalized method of characteristics for Hamilton–Jacobi equations. The impact of the discount index is indicated for equilibrium strategies of the game.
Access provided by CONRICYT-eBooks. Download chapter PDF
Similar content being viewed by others
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
6.1 Introduction
The dynamical bimatrix game with discounted integral payoff functionals is considered on the infinite horizon. Usually the discount parameter appears to be very uncertain value which reflects subjective components in economic and financial models. In this case models with discounted indices require an implementation of sensitivity analysis for solutions with respect to changing of the discount parameter. In the paper we build optimal control strategies based on Krasovskii minimax approach [10, 11], using constructions of Pontryagin maximum principle [21] and Subbotin technique of method of characteristics for generalized (minimax) solutions of Hamilton-Jacobi equations [22, 23]. Basing on constructed optimal control strategies we simulate equilibrium trajectories for dynamical bimatrix game in the framework of Kleimenov approach [8]. It is important to note that in considered statement we can obtain analytical solutions for control strategies depending explicitly on uncertain discount parameter. That allows to implement the sensitivity analysis of equilibrium trajectories with respect to changing of discount parameter and determine the asymptotical behavior of solutions when the discount parameter converges to zero. It is shown that control strategies and equilibrium solutions asymptotically converge to the solution of dynamical bimatrix game with average integral payoff functional considered in papers by Arnold [1].
It is worth to note that we use dynamical constructions and methods of evolutionary games analysis proposed in the paper [18]. To explain the dynamics of players’ interaction we use elements of evolutionary games models [2, 5, 6, 25, 27]. For the analysis of shifting equilibrium trajectories from competitive static Nash equilibrium to the points of cooperative Pareto maximum we consider ideas and constructions of cooperative dynamical games [20]. The dynamics of bimatrix game can be interpreted as a generalization of Kolmogorov’s equations for probabilities of states [9], which are widely used in Markov processes, stochastic models of mathematical economics and queuing theory. The generalization is understood in the sense that parameters of the dynamics are not fixed a priori and appear to be control parameters and are constructed by the feedback principle in the framework of control theory and differential games theory.
The solution of dynamical bimatrix games is based on construction of positional strategies that maximize own payoffs at any behavior of competing players, which means “guaranteeing” strategies [10, 11, 19]. The construction of solutions on the infinite horizon is divided into fragments with a finite horizon for which Pontryagin maximum principle is used [21] in accordance with constructions of positional differential games theory [11]. More precisely, elements of maximum principle are considered in the aggregate with the method of characteristics for Hamilton-Jacobi equations [12, 22, 24, 26]. The optimal trajectory in each time interval is constructed from pieces of characteristics while switching moments from one characteristic to another are determined by maximum principle. In this method switching moments and points generate switching lines in the phase space which determine the synthesis of optimal positional strategies. Let us note that analogous methods for construction of positional strategies are used in papers [7, 13,14,15,16,17].
In the framework of proposed approach we consider the model of competition on financial markets which is described by dynamical bimatrix game. For this game we construct switching curves for optimal control strategies and synthesize equilibrium trajectories for various values of the discount parameter. We analyze the qualitative behavior of equilibrium trajectories and demonstrate that equilibrium trajectories of dynamical bimatrix game provide better results than static Nash equilibrium. Results of the sensitivity analysis for obtained solutions are demonstrated. This analysis shows that switching curves of optimal control strategies for the series of the discount parameter values have the convergence property by the parameter. We provide calculations confirming the fact that equilibrium trajectories in the problem with discounting converge to the equilibrium trajectory in the problem with average integral payoff functional.
6.2 Model Dynamics
The system of differential equations which defines the dynamics of behavior for two players is investigated
The parameter x = x(t), 0 ≤ x ≤ 1, means the probability that first player holds to the first strategy (respectively, (1 − x) is the probability that he holds to the second strategy). The parameter y = y(t), 0 ≤ y ≤ 1, is the probability of choosing the first strategy by the second player (respectively, (1 − y) is the probability that he holds to the second strategy). Control parameters u = u(t) and v = v(t) satisfy conditions 0 ≤ u ≤ 1, 0 ≤ v ≤ 1, and can be interpreted as signals, that recommend change of strategies by players. For example, value u = 0 (v = 0) corresponds to the signal: “change the first strategy to the second one”. The value u = 1 (v = 1) corresponds to the signal: “change the second strategy to the first one”. The value u = x (v = y) corresponds to the signal: “keep the previous strategy”.
It is worth to note, that the basis for the dynamics (6.1) and its properties were examined in papers [18, 25]. This dynamics generalizes Kolmogorov’s differential equations for probabilities of states [9]. Such generalization assumes that coefficients of incoming and outgoing streams inside coalitions of players are not fixed a priori and can be constructed as positional strategies in the controlled process.
6.3 Local Payoff Functions
Let us assume that the payoff of the first player is described my the matrix A = a ij, and the payoff of the second player is described by the matrix B = b ij
Local payoff functions of the players in the time period t, t ∈ [t 0, +∞) are determined by the mathematical expectation of payoffs, given by corresponding matrices A and B in the bimatrix game, and can be interpreted as “local” interests of the players
Here parameters C A, α 1, α 2 and C B, β 1, β 2 are determined according to the classical theory of bimatrix games (see [27])
6.4 Nash Equilibrium in the Differential Game with Discounted Functionals
In this section we consider the non-zero sum differential game for two players with discounted payoff functionals on the infinite horizon
defined on the trajectories (x(⋅), y(⋅)) of the system (6.1).
Payoff functionals of the second player \(JD_{B}^{\infty }\), \(JD_{B}^{-}\), \(JD_{B}^{+}\) are determined analogously by replacement of the function g A(x, y) by the function g B(x, y).
Discounted functionals (6.2) are traditional for the problems of evolutionary economics and economic growth [6, 12], and are related to the idea of depreciation of financial funds in time. In the problems of optimal guaranteed control such functionals were considered in the paper [25]. Unlike payoff functionals optimized in each period, discounted functionals admit the possibility of loss in some periods in order to win in other periods and obtain better integral result in all periods. This fact allows the system to stay longer in favorable domains where values of local payoffs for the players are strictly better than values of static Nash equilibrium.
Let us introduce the notion of dynamical Nash equilibrium for the evolutionary game with the dynamics (6.1) and discounted payoff functionals (6.2) in the context of constructions of non-antagonistic positional differential games [8, 11, 18]. Let us define the dynamical Nash equilibrium in the class of positional strategies (feedbacks) U = u(t, x, y, ε), V = v(t, x, y, ε).
Definition 6.1
The dynamical Nash equilibria (U 0, V 0), U 0 = u 0(t, x, y, ε), V 0 = v 0(t, x, y, ε) from the class of controls by the feedback principle U = u(t, x, y, ε), V = v(t, x, y, ε) for the given problem is determined by inequalities
Here symbol X stands for the set of trajectories, starting from initial point and generated by corresponding postional strategies is the sense of the paper [11].
6.5 Auxiliary Zero-Sum Games
For the construction of desired equilibrium feedbacks U 0, V 0 we use the approach [8]. In accordance with this approach we construct the equilibrium using optimal feedbacks for differential games \(\varGamma _A = \varGamma _A^- \cup \varGamma _A^+\) and \(\varGamma _B = \varGamma _B^- \cup \varGamma _B^+\) with payoffs \(JD_A^{\infty }\) and \(JD_B^{\infty }\) (6.2). In the gamed Γ A the first player maximizes the functional \(JD_A^-(x(\cdot ),y(\cdot ))\) with the guarantee using the feedback U = u(t, x, y, ε), and the second player oppositely provides the minimization of the functional \(JD_A^+(x(\cdot ),y(\cdot ))\) using the feedback V = v(t, x, y, ε). Vice versa, in the game Γ B the second player maximizes the functional \(JD_B^-(x(\cdot ),y(\cdot ))\) with the guarantee, and the first player maximizes the functional \(JD_B^+(x(\cdot ),y(\cdot ))\).
Let us introduce following notations. By \(u_A^0=u_A^0(t,x,y,\varepsilon )\) and \(v_B^0=v_B^0(t,x,y,\varepsilon )\) we denote feedbacks that solve the problem of guaranteed maximization for payoff functionals \(JD_A^-\) and \(JD_B^-\) respectively. Let us note, that these feedbacks represent the guaranteed maximization of players’ payoffs in the long run and can be named “positive”. By \(u_B^0=u_B^0(t,x,y,\varepsilon )\) and \(v_A^0=v_A^0(t,x,y,\varepsilon )\) we denote feedbacks mostly favorable for opposite players, namely, those, that minimize payoff functionals \(JD_B^+\), \(JD_A^+\) of the opposite players. Let us call them “punishing”.
Let us note, that inflexible solutions of selected problems can be obtained in the framework of the classical bimatrix games theory. Let us propose for definiteness, (this proposition is given for illustration without loss of generality for the solution), that the following relations corresponding to the almost antagonistic structure of bimatrix game hold for the parameters of matrices A and B,
The following proposition is fair.
Lemma 6.1
Differential games \(\varGamma _A^-,\varGamma _A^+\) have equal values
and differential games \(\varGamma _B^-, \varGamma _B^+\) have equal values
for any initial position (x 0, y 0) ∈ [0, 1] × [1, 0]. These values, for example, can be guaranteed by “positive” feedbacks \(u_A^{cl}\) , \(v_B^{cl}\) corresponding to classical solutions x A , y B
“Punishing” feedbacks are determined by formulas
and correspond to classical solutions x B, y A (6.3), which generate the static Nash equilibrium NE = (x B, y A).
The proof of this proposition can me obtained by the direct substitution of shown strategies to corresponding payoff functionals (6.2).
Remark 6.1
Values of payoff functions g A(x, y), g B(x, y) coincide at points (x A, y B), (x B, y A)
The point NE = (x B, y A) is the “mutually punishing” Nash equilibrium, and the point (x A, y B) does not possess equilibrium properties in the corresponding static game.
6.6 Construction of the Dynamical Nash Equilibrium
Let us construct the pair of feedbacks, which consist the Nash equilibrium. For this let us combine “positive” feedbacks \(u_A^0,v_B^0\) and “punishing” feedbacks \(u_B^0,v_A^0\).
Let us choose the initial position (x 0, y 0) ∈ [0, 1] × [0, 1] and accuracy parameter ε > 0. Let us choose the trajectory \((x^0(\cdot ),y^0(\cdot )) \in X (x_0,y_0,U_A^0(\cdot ),v_B^0(\cdot ))\), generated by “positive” \(u_A^0=u_A^0(t,x,y,\varepsilon )\) and \(v_B^0=v_B^0(t,x,y,\varepsilon )\). Let us choose T ε > 0 such that
Let us denote by \(u_A^{\varepsilon }(t)\): [0, T ε) → [0, 1], \(v_B^{\varepsilon }(t)\): [0, T ε) → [0, 1] step-by-step implementation of strategies \(u_A^0,v_B^0\) such that the corresponding step-by-step trajectory (x ε(⋅), y ε(⋅)) satisfies the condition
From the results of the paper [8] the next proposition follows.
Lemma 6.2
The pair of feedbacks U 0 = u 0(t, x, y, ε), V 0 = v 0(t, x, y, ε), combines together “positive” feedbacks \(u_A^0\) , \(v_B^0\) and “punishing” feedbacks \(u_B^0\) , \(v_A^0\) according to relations
is the dynamical ε-Nash equilibrium.
Below we construct flexible “positive” feedbacks that generate trajectories (x fl(⋅), y fl(⋅)), which reduce to “better” positions than the inflexible dynamical equilibrium (x B, y A), (x A, y B) by both criteria \(JD_A^{\infty }(x^{fl}(\cdot ),y^{fl}(\cdot )) \geq v_A\), \(JD_B^{\infty }(x^{fl}(\cdot ),y^{fl}(\cdot )) \geq v_B\).
6.7 Two-Step Optimal Control Problems
For the construction of “positive” feedbacks \(u_{A}^{0}=u_{A}^{fl}(x,y)\), \(v_{B}^{0}=v_{B}^{fl}(x,y)\) we consider in this section the auxiliary two-step optimal control problem with discounted payoff functional for the first player in the situation, when actions of the second player are most unfavorable. Namely, let us analyze the optimal control problem for the dynamical system (6.1)
with the payoff functional
Here without loss of generality let us consider that t 0 = 0, T = T f, and terminal moment of time T f = T f(x 0, y 0) we determine later.
Without loss of generality, we assume that the value of the static game equals to zero
and next conditions hold
Let us consider the case when initial conditions (x 0, y 0) of the system (6.4) satisfy relations
Let us assume that actions of the second player are mostly unfavorable for the first player. For trajectories of the system (6.4), which start from initial positions (x 0, y 0) (6.8), these actions \(v_{A}^{0}=v_{A}^{cl}(x,y)\) are determined by the relation
Optimal actions \(u_{A}^{0}=u_{A}^{fl}(x,y)\) of the first player according to the payoff functional \(JD_{A}^{f}\) (6.5) in this situation can be presented as the two-step impulse control: it equals one from the initial time moment t 0 = 0 till the moment of optimal switch s and then equals to zero till the final time moment T f
Here the parameter s is the optimization parameter. The final time moment T f is determined by the following condition. The trajectory (x(⋅), y(⋅)) of the system (6.4), which starts from the line where x(t 0) = x A, returns to this line when x(T f) = x A.
Let us consider two aggregates of characteristics. The first one is described by the system of differential equations with the value on the control parameter u = 1
solutions of which are determined by the Cauchy formula
Here initial positions (x 0, y 0) satisfy conditions (6.8) and time parameter t satisfies the inequality 0 ≤ t < s.
The second aggregate of characteristics is given by the system of differential equations with the value of the control parameter u = 0
solutions of which are determined by the Cauchy formula
Here initial positions (x 1, y 1) = (x 1(s), y 1(s)) are determined by relations
and the time parameter t satisfies the inequality 0 ≤ t < p. Here the final time moment p = p(s) and the final position (x 2, y 2) = (x 2(s), y 2(s)) of the whole trajectory (x(⋅), y(⋅)) is given by formulas
The optimal control problem is to find such moment of time s and the corresponding switching point (x 1, y 1) = (x 1(s), y 1(s)) on the trajectory (x(⋅), y(⋅)), where the integral I = I(s) reaches the maximum value
On the Fig. 6.1 we depict the initial position IP, chosen on the line x = x A when y > y A, the characteristic CH, oriented on the vertex (1, 0) of the unit square, characteristics CH 1, CH 2, CH 3, oriented on the vertex (0, 0) of the unit square, switching points SP 1, SP 2, SP 3 of the motion along characteristics and final points of the motion FP 1, FP 2, FP 3, located of the line x = x A.
6.8 The Solution of the Two-Step Optimal Control Problem
We obtain the solution of the two-step optimal control problem (6.9)–(6.15), by calculating the derivative dI∕ds, presenting it as the function of optimal switching points (x, y) = (x 1, y 1), equating this derivative to zero dI∕ds = 0 and finding the equation F(x, y) = 0 for the curve, that consist from optimal switching points (x, y).
Sufficient maximum conditions in such construction are obtained from the fact that the integral I(s) has the property of monotonic increase by the variable s in the initial period, because the integrand g A(x, y) is positive, g A(x, y) > w A = 0, in the domain x > x A, y > y A. In the finite period the integral I(s) strictly monotonically decreases by the variable s, because the integrand g A(x, y) is negative, g A(x, y) < w A = 0, in the domain x > x A, y < y A.
Firstly let us calculate integrals I 1, I 2
Let us calculate derivatives dI 1∕ds, dI 2∕ds and present them as functions of optimal switching points (x, y) = (x 1, y 1)
While calculating the derivative dI 2∕ds let us take into the account next expressions for derivatives dx∕ds, dy∕ds, dp∕ds and the exponent e −p as functions of variables (x, y):
Let us introduce the new derivative q = e −p and obtain the expression for dI 2∕ds
Let us summarize derivatives dI 1∕ds and dI 2∕ds, equalize the expression to zero and express y by x in the following form:
Simplifying the expression we obtain the formula:
Taking into the account the fact that w A = 0 (6.6), we obtain a 22 = (α 1 α 2)∕C A. By substitution of this relation and the expression q = α 2∕(C A x) to previous formula we obtain:
Multiplying both parts on the expression by x (λ+2), we obtain:
Taking into the account relations x A = α 2∕C A and y A = α 1∕C A (6.7), we obtain the final expression for the switching curve \(M_A^1(\lambda )\):
To construct the final switching curve M A(λ) for the optimal strategy of the first player in the game with the discounted functional in the case C A > 0, we add to the curve \(M_{A}^{1}(\lambda )\) the similar curve \(M_{A}^{2}(\lambda )\) in the domain, where x ≤ y A and y ≤ y A
In the case when C A < 0, curves M A(λ), \(M_{A}^{1}(\lambda )\) and\(M_{A}^{2}(\lambda )\) are described by formulas
The curve M A(λ) divides the unit square [0, 1] × [0, 1] into two parts: the upper part
and the lower part
The “positive” feedback \(u_{A}^{fl}\) has the following structure
On the Fig. 6.2 we show switching curves \(M_{A}^{1}(\lambda )\), \(M_{A}^{2}(\lambda )\) for the first player. Directions of velocities \(\dot {x}\) are depicted by horizontal (left and right) arrows.
For the second player one can get similar switching curves M B(λ) for the optimal control problem with the discounted functional, corresponding to the matrix B. More precisely, in the case when C B > 0, the switching curve M B(λ) is given by relations
In the case when the parameter C B is negative C B < 0, curves M B(λ), \(M_{B}^{1}(\lambda )\) and \(M_{B}^{2}(\lambda )\) are determined by formulas
The curve M B(λ) divide the unit square [0, 1] × [0, 1] into two parts: the left part
and the right part
The “positive” feedback \(v_{B}^{fl}\) has the following structure
Remark 6.2
Let us note that in papers by Arnold [1] average integral payoff functionals were considered
In the paper [16] switching curves for optimal control strategies of players in the game with average integral functionals were obtained. For example, for the first player in the case when C A > 0 the switching curve in the domain x ≥ x A, y ≥ y A is described by relations
The asymptotical analysis of solutions (6.16) for the game with discounted payoff functionals shows, that according to L’Hospital’s rule, when the discount parameter λ tends to zero, the relation for switching curves (6.16) of the control strategy for the first player converges to switching curves (6.23) in the game with average integral payoff functionals (6.22).
On the Fig. 6.2. the solid line shows the switching curve of control strategies for the first player in the game with average integral payoff functionals, which is asymptotically approximated by solutions of the game with discounted functionals when λ ↓ 0. The dashed line and the dotted line show switching curves of control strategies for the first player in the game with discounted payoff functionals with values of the discount parameter λ = 0.1 and λ = 0.2, respectively.
On the Fig. 6.3 we show switching curves \(M_{B}^{1}(\lambda )\), \(M_{B}^{2}(\lambda )\) for the second player. Directions of velocities \(\dot {y}\) are depicted by vertical (up and down) arrows.
It is worth to clarify the asymptotical behavior of switching curves for optimal control when discount parameters can be infinitely large. In this case, one can check that switching curve M A(λ) for optimal control in the problem with discounted integral payoffs describing long-term interests of players converge to the switching line y = y A generated by the short-run payoff function g A(x, y) when the discount parameter λ tends to infinity. Such behavior of the switching curve M A(λ) is shown on the Fig. 6.4.
6.9 Guaranteed Values of Discounted Payoffs
Let us formulate the proposition, which confirms, that the “positive” optimal control by the feedback principle \(u_{A}^{fl}(x,y)\) (6.18) with the switching cure M A, defined by formulas (6.16), (6.17), guarantee that the value of discounted payoff functionals is more or equal than the value w A (6.6) of the static matrix game.
Theorem 6.1
For any initial position (x 0, y 0) ∈ [0, 1] × [0, 1] and for any trajectory
generated by the optimal control by the feedback principle \(u_{A}^{fl}=u_{A}^{fl}(x,y)\) there exists the final moment of time t ∗∈ [0, T A] such that in this moment of time the trajectory (x fl(⋅), y fl(⋅)) reaches the line where x = x A , namely x fl(t ∗) = x A . Then, according to the construction of the optimal control, that maximizes the integral (6.15) by the feedback principle \(u_{A}^{fl}\) , the following estimate holds
In particular, this inequality remains valid when time T tends to infinity
Inequalities (6.24), (6.25) mean, that the value of the discounted functional is not worse, than the value w A (6.6) of the static matrix game.
The analogous result is fair for trajectories, which generated by the optimal control \(v_{B}^{fl}\) (6.21), that corresponds to the switching curve M B (6.19), (6.20).
Proof
The result of the theorem follows from the fact that the value of the payoff functional (6.5) is maximum on the constructed broken line. In particular, it is more or equal, than the value of this functional on the trajectory which stays on the segment x = x A (see Fig. 6.1) with the control u(t) = x A. The value of the payoff functional on such trajectory is following
These arguments imply the required relation (6.24), which in the limit transition provides the relation (6.25). □
Remark 6.3
Let us consider the acceptable trajectory \((x_{AB}^{fl}(\cdot ),y_{AB}^{fl}(\cdot ))\), generated by “positive” feedbacks \(u_{A}^{fl}\) (6.18), \(v_{B}^{fl}\) (6.21). Then in accordance with the Theorem 6.1, next inequalities take place
and, hence, the acceptable trajectory \((x_{AB}^{fl}(\cdot ),y_{AB}^{fl}(\cdot ))\) provides the better result for both players, than trajectories, convergent to points of the static Nash equilibrium, in which corresponding payoffs are equal to values w A and w B.
6.10 Equilibrium Trajectories in the Game with Discounted Payoffs
Let us consider payoff matrices of players on the financial market, which reflect the data of investigated markets of stocks [3] and bonds [4] in USA. The matrix A corresponds to the behavior of traders, which play on increase of the course and are called “bulls”. The matrix B corresponds to the behavior of traders, which play on the depreciation of the course and are called “bears”. Parameters of matrices represent rate of return for stocks and bonds, expressed in the form of interest rates,
Characteristic parameters of static games are given at the following levels [27]
On the Fig. 6.5 we present broken lines of players’ best replies, saddle points NA, NB in static antagonistic games, the point of the Nash equilibrium NE in the static bimatrix game.
Let us note, that players of the coalition of “bulls” gain in the case of upward trend of markets, when players of both coalitions invest in the same market. And players of the coalition of “bears” make profit from investments in the case of downward trend of markets when players of the coalition of “bulls” move their investments from one market to another.
For the game of coalitions of “bulls” and “bears” we construct switching curves M A(λ), M B(λ) and provide calculations of equilibrium trajectories of the market dynamics with the value of the discount parameter λ = 0.1.
This calculations are presented on the Fig. 6.6. Here we show saddle points NA, NB in static antagonistic games, the point of the Nash equilibrium NE in the static bimatrix game, switching lines for players’ controls \(M_A(\lambda ) = M_A^1(\lambda ) \bigcup M_A^2(\lambda )\) and \(M_B(\lambda ) = M_B^1(\lambda ) \bigcup M_B^2(\lambda )\) in the dynamical bimatrix game with discounted payoff functionals for matrices A, B (6.26). The field of velocities of players is depicted by arrows.
The field of directions generates equilibrium trajectories, one of which is presented on the Fig. 6.6. This trajectory \(TR(\lambda )=(x_{AB}^{fl}(\cdot ),y_{AB}^{fl}(\cdot ))\) starts from the initial position IP = (0.1, 0.9) and moves along the characteristic in the direction of the vertex (1, 1) of the unit square [0, 1] × [0, 1] with control signals u = 1, v = 1. Then it crosses the switching line M B(λ), and the second coalition switches the control v from 1 to 0. Then, the trajectory TR(λ) moves in the direction of the vertex (1, 0) until it reaches the switching line M A(λ). Here players of the first coalition change the control signal u from 1 to 0. After that the movement of the trajectory is directed along the characteristic to the vertex (0, 0). Then the trajectory crosses the line M B(λ), on which the sliding mode arises, during which the switch of controls of the second coalition occurs, and the trajectory TR(λ) converge to the point \(IM(\lambda )=M_{A}(\lambda ) \bigcap M_{B}(\lambda )\) of the intersection of switching lines M A(λ), M B(λ).
References
Arnold, V.I.: Optimization in mean and phase transitions in controlled dynamical systems. Funct. Anal. Appl. 36, 83–92 (2002). https://doi.org/10.1023/A:1015655005114
Basar, T., Olsder, G.J.: Dynamic Noncooperative Game Theory. Academic Press, London (1982)
CNN Money [Electronic resource]. http://money.cnn.com/
Forex Market [Electronic resource]. http://www.fxstreet.ru.com/
Friedman, D.: Evolutionary games in economics. Econometrica 59(3), 637–666 (1991)
Intriligator, M.: Mathematical Optimization and Economic Theory. Prentice-Hall, New York (1971)
Klaassen, G., Kryazhimskii, A.V., Tarasyev, A.M.: Multiequilibrium game of timing and competition of gas pipeline projects. J. Optim. Theory Appl. 120(1), 147–179 (2004)
Kleimenov, A.F.: Nonantagonistic Positional Differential Games. Nauka, Yekaterinburg (1993)
Kolmogorov, A.N.: On analytical methods in probability theory. Uspekhi Mat. Nauk 5, 5–41 (1938)
Krasovskii, A.N., Krasovskii, N.N.: Control Under Lack of Information. Birkhauser, Boston (1995)
Krasovskii, N.N., Subbotin, A.I.: Game-Theoretical Control Problems. Springer, New York (1988)
Krasovskii, A.A., Taras’ev, A.M.: Dynamic optimization of investments in the economic growth models. Autom. Remote Control 68(10), 1765–1777 (2007)
Krasovskii, N.A., Tarasyev, A.M.: Search for maximum points of a vector criterion based on decomposition properties. Proc. Steklov Inst. Math. 269, 174 (2010). https://doi.org/10.1134/S0081543810060155
Krasovskii, N.A., Tarasyev, A.M.: Decomposition algorithm of searching equilibria in a dynamic game. Autom. Remote Control 76, 185 (2015). https://doi.org/10.1134/S0005117915100136
Krasovskii, N.A., Tarasyev, A.M.: Equilibrium Solutions in Dynamical Games. UrGAU, Yekaterinburg (2015)
Krasovskii, N.A., Tarasyev, A.M.: Equilibrium trajectories in dynamical bimatrix games with average integral payoff functionals. Math. Game Theory Appl. 8(2), 58–90 (2016)
Krasovskii, N.A., Kryazhimskiy, A.V., Tarasyev, A.M.: Hamilton–Jacobi equations in evolutionary games. Proc. Inst. Math. Mech. UrB RAS 20(3), 114–131 (2014)
Kryazhimskii, A.V., Osipov, Yu.S.: On differential-evolutionary games. Proc. Steklov Inst. Math. 211, 234–261 (1995)
Kurzhanskii, A.B.: Control and Observation Under Uncertainty. Nauka, Moscow (1977)
Petrosjan, L.A., Zenkevich, N.A.: Conditions for sustainable cooperation. Autom. Remote Control 76, 84 (2015). https://doi.org/10.1134/S0005117915100148
Pontryagin, L.S., Boltyanskii, V.G., Gamkrelidze, R.V., Mischenko, E.F.: The Mathematical Theory of Optimal Processes. Interscience Publishers, New York (1962)
Subbotin, A.I.: Minimax Inequalities and Hamilton–Jacobi Equations. Nauka, Moscow (1991)
Subbotin, A.I., Tarasyev, A.M.: Conjugate derivatives of the value function of a differential game. Dokl. AN SSSR 283(3), 559–564 (1985)
Subbotina, N.N.: The Cauchy method of characteristics and generalized solutions of Hamilton–Jacobi–Bellman equations. Dokl. Acad. Nauk SSSR 320(3), 556–561 (1991)
Tarasyev, A.M.: A Differential model for a 2 × 2-evolutionary game dynamics. IIASA Working Paper, Laxenburg, Austria, WP-94-063 (1994). http://pure.iiasa.ac.at/4148/1/WP-94-063.pdf
Ushakov, V.N., Uspenskii, A.A., Lebedev, P.D.: Geometry of singular curves of a class of time-optimal problems. Vestn. Sankt-Peterburgsk. Univ. 10(3), 157–167 (2013)
Vorobyev, N.N.: Game Theory for Economists and System Scientists. Nauka, Moscow (1985)
Acknowledgements
The paper is supported by the Project 18-1-1-10 “Development of the concept of feedback control, minimax approach and singular perturbations in the theory of differential equations” of the Integrated Program of UrB RAS
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer International Publishing AG, part of Springer Nature
About this chapter
Cite this chapter
Krasovskii, N., Tarasyev, A. (2018). The Impact of Discounted Indices on Equilibrium Strategies of Players in Dynamical Bimatrix Games. In: Petrosyan, L., Mazalov, V., Zenkevich, N. (eds) Frontiers of Dynamic Games. Static & Dynamic Game Theory: Foundations & Applications. Birkhäuser, Cham. https://doi.org/10.1007/978-3-319-92988-0_6
Download citation
DOI: https://doi.org/10.1007/978-3-319-92988-0_6
Published:
Publisher Name: Birkhäuser, Cham
Print ISBN: 978-3-319-92987-3
Online ISBN: 978-3-319-92988-0
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)