Abstract
In this paper we propose an algorithm for solving the split feasibility problem \(x\in C, Ax\in Q\) with C being the solution set of an equilibrium problem and A can be nonlinear. The proposed algorithm is a combination between the projection method for the equilibrium problem and the gradient method for the inclusion \(Ax\in Q\). The convergence of the algorithm is investigated. A numerical example for a jointly constrained Nash equilibrium model in electricity production market is provided to demonstrate the behavior of the algorithm.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
In 1994, Censor and Elfving [12] first introduced the split feasibility problem in finite-dimensional Hilbert spaces for modeling inverse problems which arise from phase retrievals and in medical image reconstruction. In this setting the problem is stated as
where C, Q are convex subsets in \(\mathcal {R}^n\), \(\mathcal {R}^m\) respectively, and \(A: \mathcal {R}^n \rightarrow \mathcal {R}^m\) is a linear continuous mapping on \(\mathcal {R}^n\). Problem (SFP) can also be applied to study intensity-modulated radiation therapy and other practical problems see, for instance [8, 10, 11, 13, 15] and the references therein. An important special case of Problem (SFP) is obtained when C is the solution set of the equilibrium problem given as
where K is a convex subset in \(\mathcal {R}^n\) and f is a monotone bifunction with f(x, .) being convex on K and \(f(x,x) = 0 \) for every \(x \in K\). It is well known that some important problems such as optimization, variational inequality, Kakutani fixed point, saddle point ones and Nash equilibrium model can be formulated equivalently in the form of (EP), see e.g. [7, 35] and the survey paper [6].
Some solution methods for solving split feasibility problem (SFP) when C and/or Q are solution sets of some other problems such as fixed point, optimization, variational inequality, equilibrium ones have been developed see [2, 9, 14, 18, 23, 33, 34, 38,39,40] and the survey paper [31]. These methods either use the adjoint operator of A or are based upon a convex mathematical programming formulation, and therefore they fail to apply to the case when A is nonlinear. A nonlinear split feasibility problem with C and Q being the intersection of convex subsets has been considered in [30], where the problem was equivalently formulated as a differentiable mathematical program. Under the assumption that the objective function of the latter program is convex, the problem was further formulated as a cooercive variational inequality that was solved by a splitting projection algorithm.
The purpose of this paper is to propose an algorithm for solving Problem (SFP) in finite dimensional Euclidean spaces where C is the solution set of a paramonotone equilibrium problem and A is a quasilinear operator defined by quasilinear functions. Quasilinear functions play an important role in mathematics and many real-life problems see e.g. [4, 19, 27, 32]. The algorithm that we propose for Problem (SFP) is a combination between the projection method for inclusion \(Ax \in Q\) and the gradient projection method for Problem (EP). This algorithm can be considered as an extension to Problem (SFP) of the one by Santos and Scheimberg in [36] for (EP). The main difference between these algorithms is that our algorithm uses an additional projection in order to handle the requirement \(Ax \in Q\), where A may not be linear. In the case of equilibrium problem (EP) considered in [36], the set \(Q\equiv \mathcal {R}^n\), then our [36] algorithm collapses into the one in [36].
The paper is organized as follows. In the next section we recall some definitions and properties of quasiconvex and quasilinear functions. The algorithm and its convergence analysis are presented in Sect. 3. An application to jointly constraint Nash equilibrium models is discussed and some computational results for a coupled constrained Nash–Cournot equilibrium model in electricity production market are reported in the last section.
2 Preliminaries
The following lemmas will be used for validity and convergence of the algorithm.
Lemma 1
([5], p. 61) Let C be a nonempty closed convex subset in a Hilbert space \(\mathcal {H}\) and \(P_C(x)\) be the metric projection of x onto C. Then
-
(i)
\(\langle x-y, P_C(x)-P_C(y) \rangle \ge \Vert P_C(x) -P_C(y)\Vert ^2 \quad \forall x,y \in \mathcal {H}\);
-
(ii)
\(\langle x-P_C(x), P_C(x)-y \rangle \ge 0 \quad \forall x\in \mathcal {H}, y\in C. \)
Lemma 2
([1]) Let \(\{v_k\}\) and \(\{\delta _k\}\) be nonnegative sequences of real numbers satisfying \(v_{k+1}\le v_k+\delta _k\) with \(\sum _{k=1}^{\infty } \delta _k <+\, \infty \). Then the sequence \(\{v_k\}\) is convergent.
The quasiconvex functions were first introduced by De Finetti in [20]. This class of functions are widely used in optimization, game theory, economics and other fields.
Definition 1
([4]) Let \(X\subset \mathbb {R}^n\) be a convex set and \(\varphi : X\rightarrow R\).
-
(i)
\(\varphi \) is called quasiconvex on X if its sublevel set
$$\begin{aligned} S_{\varphi ,\alpha }=\left\{ x\in X: \quad \varphi (x)\le \alpha \right\} \end{aligned}$$is convex for every \(\alpha \in \mathbb {R}\).
-
(ii)
\(\varphi \) is called quasiconvave on X if \(-\varphi \) is quasiconvex on X.
-
(iii)
\(\varphi \) is called quasilinear on X if it is both quasiconvex and quasiconcave.
Two important quasilinear functions are \(\varphi (x):= \log x\) and \(\varphi (x) := \frac{a^Tx+b}{c^Tx+d}\). The first function is quasilinear on \(\mathcal {R}_{++}\), while the second one is quasilinear on the set \(\left\{ x: \quad c^Tx+d \not = 0\right\} \).
The following proposition provides a characterization of the quasiconvex functions that will be used in the sequel.
Proposition 1
([21]) The following statements are equivalent
-
(i)
\(\varphi \) is quasiconvex on X.
-
(ii)
For any \(x,y \in X\) and \(\lambda \in \left[ 0, 1\right] \) one has
$$\begin{aligned} \varphi (\lambda x_1+(1-\lambda ) x_2)\le \max \left\{ \varphi (x_1), \varphi (x_2)\right\} . \end{aligned}$$(1)
Consequently, \(\varphi \) is quasilinear on X if and only if for every \(x,y \in X\) and \(\lambda \in \left[ 0, 1\right] \) we have
Theorem 1
([4, 21]) Suppose that \(\varphi :\mathcal {R}^n \rightarrow \mathcal {R}\) is differentiable on an open convex set containing X. Then \(\varphi \) is quasiconvex on X if and only if
It is easy to see that if \(\varphi _i\) is quasiconvex on X for every \(i=1,2, \dots ,m\), then \(\varphi (x):=\max _{i=1,2,\dots ,m} \varphi _i (x)\) is quasiconvex on X.
3 Algorithm and its convergence analysis
In this section we describe a projection method for solving the following split feasibility problem
where \(\emptyset \not = K \subseteq \mathcal {R}^n\) is convex, \(f: K \times K \rightarrow \mathcal {R}\), \(\emptyset \not = Q \subseteq \mathcal {R}^m\) and F is a map from \(\mathcal {R}^n\) to \(\mathcal {R}^m\).
In what follows we suppose that Problem (NSEP) admits a solution and that
-
(A1)
\(Q=Q_1\times Q_2\times \dots \times Q_m\) where \(Q_i\) is a closed convex subset of \(\mathcal {R}\) for every \(i=1,2,\dots ,m\);
-
(A2)
\(F=(F_1,F_2, \dots , F_m)\) where \(F_i:\mathcal {R}^n\rightarrow \mathcal {R}\) is quasilinear, i.e., both quasiconvex and quasiconcave, and continuously differentiable on an open set containing K.
We recall some well known definitions on monotonicity of a bifunction see e.g. [24, 26]
Definition 2
Let K be a convex set and \(S \subseteq K\). A bifunction \(f :K\times K \rightarrow \mathcal {R}\) is said to be
-
(a)
monotone on K with respect to S if
$$\begin{aligned} f(x,y) + f(y,x) \le 0 \quad \forall x\ \in S, y\in K; \end{aligned}$$ -
(b)
pseudomonotone on K with respect to S if
$$\begin{aligned} f(x,y) \ge 0 \Rightarrow f(y,x) \le 0, \quad \forall x\in S, y\in K; \end{aligned}$$ -
(c)
paramonotone on K with respect to S if
$$\begin{aligned} x\in S, y\in K: f(x,y) = f(y,x) = 0 \Rightarrow y \in S. \end{aligned}$$
Paramonotone bifunctions have been introduced and studied in [24] and used in some papers [36, 40]. Clearly, \(f(x,y) := \varphi (y) - \varphi (x)\) is paramonotone for every function \(\varphi \). Another example for paramonotone bifunction is \(f(x,y) := \langle Ax+b, y-x \rangle \) where A is a \((n\times n)\)-nonsymmetric matrix such that \(\hat{A}=\frac{1}{2}(A+A^T)\) is positive semidefinite and \(\ker (\hat{A})\subset \ker (A)\). In [24] (Prop. 3.2), Iusem proved that \(f(x,y) := \langle Ax+b, y-x \rangle \) is paramonotone if and only if \(\hat{A}\) is positive semidefinite and \(\ker (\hat{A})\subset \ker (A)\). Note that since A is nonsymmetric, \(Ax+b\) can not be expressed as the gradient of any function.
Let \(Sol({ EP})\) denote the solution set of the equilibrium problem
Then it is not hard to see that under Assumptions (A1), (A2), Problem (NSEP) can take the form
with C being the solution set of (EP). Moreover, the function \(p_i(x):=|(I-P_{Q_i}) (F_i(x))|^2\) is quasiconvex and hence, \(p(x):=\max _{i=1,2,\dots ,m} p_i(x)\) is quasiconvex, too.
The use of the function p is motivated mainly by two facts. The first one is that the split convex feasibility problem
introduced by Censor et al where C and Q are closed convex sets, has been considered in a lot papers when F is a linear bounded mapping. In this case, there are two approaches to solve the problem. The first one uses the adjoint operator of the linear mapping, while the second one formulates the problem equivalently as a convex programming problem. However both these approaches fail to apply to the case F is not linear. The second fact is that the quasiconvex function and its properties were well studied in some references [19,20,21, 32], and some algorithms were developed to find a local or a global minimizer of a quasiconvex function over a convex set, see [19, 21, 28, 32]. Note that in Problem (OP), the feasible domain C is not given explicitly, but the solution set of an equilibrium problem.
For each \(x \in K\), let
then we have the following lemma:
Lemma 3
Under the assumptions (A1) and (A2), it holds that
-
(i)
The function \(p_i\) is quasiconvex and differentiable on K;
-
(ii)
The function p is quasiconvex on K.
Proof
-
(i)
Since \(F_i\) is differentiable on K, then it is easy to see that \(p_i(x)= | (I-P_{Q_i}) (F_i(x))|^2\) is also differentiable on K and
$$\begin{aligned} \nabla p_i(x)= 2 (I-P_{Q_i})(F_i(x))\nabla F_i(x). \end{aligned}$$Since \(F_i\) is quasilinear on K for every i, then for any \(x_1, x_2 \in K\) and \(\lambda \in \left[ 0,1\right] \) we have
$$\begin{aligned} \min \left\{ F_i(x_1), F_i(x_2)\right\} \le F_i(\lambda x_1+(1-\lambda ) x_2)\le \max \left\{ F_i(x_1), F_i(x_2)\right\} . \end{aligned}$$Hence, there exists \(\alpha \in \left[ 0,1\right] \) such that
$$\begin{aligned} F_i(\lambda x_1+(1-\lambda ) x_2) = \alpha F_i(x_1)+ (1-\alpha ) F_i(x_2). \end{aligned}$$It is well-known that the function \(|(I-P_{Q_i})(.)|^2\) is convex. Then one has
$$\begin{aligned}&p_i(\lambda x_1+ (1-\lambda ) x_2)\\&\qquad =|(I-P_{Q_i})(F_i(\lambda x_1+(1-\lambda ) x_2))|^2\\&\qquad =|(I-P_{Q_i}) \left[ \alpha F_i(x_1) +(1-\alpha ) F_i(x_2)\right] |^2\\&\qquad \le \alpha |(I-P_{Q_i})(F_i(x_1))|^2+ (1-\alpha ) |(I-P_{Q_i})(F_i(x_2))|^2\\&\qquad = \alpha p_i(x_1)+(1-\alpha )p_i(x_2)\\&\qquad \le \max \left\{ p_i(x_1), p_i(x_2)\right\} \quad \forall x_1, x_2 \in K, \end{aligned}$$which implies that \(p_i\) is quasiconvex on K.
-
(ii)
This assertion comes directly from (i) and the definition of p. \(\square \)
We need the following widely used assumptions for bifunction f.
(A3) For every \(x\in K\), the function f(x, .) is convex, subdifferentiable, f(., x) is upper semicontinuous on an open convex set containing K and \(f(x,x) = 0\) for every \(x\in K\).
(A4) The bifunction f is pseudomonotone on Kwith respect to the solution set S of Problem (EP), that is
For a fixed \(\epsilon \ge 0\), let \(\partial ^\epsilon _2f(x,x)\) denote the \(\epsilon \)-subdifferentiable of the convex function f(x, .) at x, that is
Note that since the function f(x, .) is convex and subdifferentiable on an open convex set containing K, the \(\epsilon \)-subdifferential mapping \(\partial ^\epsilon _2f( x,.)\) maps a bounded set in K to a bounded set.
The algorithm then can be described as follows:
Algorithm 1.
Take a positive number \(\delta \) and real sequences \(\{\delta _k\}\), \(\{\beta _k\}\), \(\{\epsilon _k\}\) satisfying the conditions
Step 0: Seek \(x_1\in K\) and let \(k:=1\).
Iteration k: Given \(x_k\in K\). Take \(g_k\in \partial _2^{\epsilon _k}f(x_k,x_k)\) and define
Compute \(y_k=P_K(x_k-\alpha _k g_k)\).
If \(\nabla p_i(y_k)=0 \quad \forall i\in I(y_k)\) then take \(\widehat{h}_k=0\);
Otherwise, take \(0\ne h_k \in co \left\{ \nabla p_i(y_k) , \quad i\in I(y_k)\right\} \), where \( co \) stands for the convex hull notion. Set
and compute
then increase k by one and go to iteration k.
Remark 1
-
(i)
If \(\epsilon _k=0\), \(x_k=y_k\) and \( p(x_k)=0\), then \(x_k\) is an exact solution. So, we can call \(x_k\) is an \(\epsilon \)-solution if \(\epsilon _k\le \epsilon \), \(||x_k-y_k||\le \epsilon \) and \( p(x_k)\le \epsilon \).
-
(ii)
If \(Q_i \equiv \mathcal {R}\) for every i, then \(p_i(x) = |(I - P_{Q_i}) F_i(x)| = 0\) for every i. Thus Problem (NSEP) becomes the equilibrium problem \(({ EP})\). In this case Algorithm 1 becomes the one in [36] with exact projection. In fact, since \(\widehat{h}_k=0\) for every k, the iteration k of Algorithm 1 reads
Iteration k: Given \(x_k\in K\). Take \(g_k\in \partial _2^{\epsilon _k}f(x_k,x_k)\) and define
Compute \(x_{k+1}=P_K(x_k-\alpha _k g_k)\).
The proof of the lemma below can be done by a similar way as the one in [36].
Lemma 4
([36]) For every k, the following inequalities hold
-
(i)
\(\alpha _k \Vert g_k\Vert \le \beta _k;\)
-
(ii)
\( \Vert y_k-x_k\Vert \le \beta _k.\)
The following lemmas will be used in the proof of the convergence for the proposed algorithm.
Lemma 5
It holds that
Proof
By nonexpansiveness of \(P_K\), one has
Lemma 6
-
(i)
Under Assumptions (A1), (A2), (A3), for \(z\in K\), it holds that
$$\begin{aligned} \Vert x_{k+1}-z\Vert ^2\le \Vert x_k-z||^2 +2\alpha _k f(x_k,z)-2\alpha _k \langle \widehat{h}_k, y_k -z\rangle +A_k, \end{aligned}$$(7)where \(A_k= 2(\alpha _k\epsilon _k +\beta _k^2)+ \alpha ^2_k.\)
-
(ii)
If there exist \(z\in K\), \(\epsilon >0\) and \(\delta > 0\) such that
$$\begin{aligned} p(y)< p(y_k)-\delta \quad \forall y\in B(z, \epsilon ), \end{aligned}$$then
$$\begin{aligned} \langle \widehat{h}_k, y_k -z \rangle \ge \epsilon \quad \forall k \end{aligned}$$whenever \(\widehat{h}_k \not = 0\).
Proof
-
(i)
It is clear that,
$$\begin{aligned} \Vert y_k-z\Vert ^2= & {} \Vert z-x_k+x_k-y_k\Vert ^2\\= & {} \Vert x_k-z\Vert ^2-\Vert x_k-y_k\Vert ^2+2\langle x_k-y_k, z-y_k\rangle \\\le & {} \Vert x_k-z\Vert ^2+2\langle x_k-y_k, z-y_k\rangle . \end{aligned}$$Since \(y_k=P_K(x_k-\alpha _k g_k)\), we have
$$\begin{aligned}&\langle y_k-x_k+\alpha _k g_k, z-y_k\rangle \ge 0 \\&\quad \Leftrightarrow \langle \alpha _k g_k, z-y_k\rangle \ge \langle x_k-y_k, z-y_k\rangle . \end{aligned}$$Hence,
$$\begin{aligned} \Vert y_k-z\Vert ^2\le & {} \Vert x_k-z\Vert ^2+2\langle \alpha _k g_k, z-y_k\rangle \nonumber \\= & {} \Vert x_k-z\Vert ^2+2\langle \alpha _k g_k, z-x_{k}\rangle +2\langle \alpha _k g_k, x_{k}-y_k\rangle . \end{aligned}$$(8)From \(g_k\in \partial _2^{\epsilon _k} f(x_k, x_k)\) it follows that
$$\begin{aligned}&f(x_k,z)-f(x_k, x_k)\ge \langle g_k, z-x_k \rangle -\epsilon _k\nonumber \\&\qquad \Leftrightarrow f(x_k, z)+\epsilon _k \ge \langle g_k, z-x_k\rangle . \end{aligned}$$(9)On the other hand, by Lemma 4, it holds that
$$\begin{aligned} \langle \alpha _k g_k, x_k -y_k\rangle\le & {} \alpha _k\Vert g_k\Vert \Vert x_k - y_k\Vert \le \beta _k^2. \end{aligned}$$(10)From (8), (9) and \(\alpha _k >0\), it follows that
$$\begin{aligned} \Vert y_k-z\Vert ^2 \le \Vert x_k-z\Vert ^2+2\alpha _k f(x_k,z)+2\alpha _k\epsilon _k+2\beta _k^2, \end{aligned}$$(11)which, by Lemma 5, implies
$$\begin{aligned} \Vert x_{k+1}-z\Vert ^2\le \Vert x_k-z||^2 +2\alpha _k f(x_k,z)-2\alpha _k \langle \widehat{h}_k, y_k -z\rangle +A_k, \end{aligned}$$(12)where \(A_k= 2(\alpha _k\epsilon _k +\beta _k^2)+ \alpha ^2_k.\)
-
(ii)
Since \(z+\epsilon \widehat{h}_k \in B(z, \epsilon )\), by the assumption, we have
$$\begin{aligned} p(z+\epsilon \widehat{h}_k) < p(y_k). \end{aligned}$$Then from the definitions of p and \(I(y_k)\), we obtain
$$\begin{aligned} p_i(z+\epsilon \widehat{h}_k) < p_i(y_k) \quad \forall i \in I(y_k). \end{aligned}$$Applying Theorem 1 with \(\varphi = p_i\), \(y = z+\epsilon \widehat{h}_k\), and \(x = y_k\) we have
$$\begin{aligned} \langle \nabla p_i(y_k), z+\epsilon \widehat{h}_k -y_k\rangle \le 0, \quad \forall i\in I(y_k). \end{aligned}$$Thus, since \(\widehat{h}_k \not =0\), \(\widehat{h}_k = \frac{h_k}{\Vert h_k\Vert }\) with \(h_k \in co \left\{ \nabla p_i(y_k) , \quad i\in I(y_k)\right\} \). So,
$$\begin{aligned} \langle \widehat{h}_k, z+\epsilon \widehat{h}_k -y_k\rangle \le 0. \end{aligned}$$Hence,
$$\begin{aligned} \langle \widehat{h}_k, y_k -z \rangle \ge \epsilon . \square \end{aligned}$$
Before stating the convergence result of the algorithm we recall [16] that the directional derivative of a local Lipschitz function \(\varphi : \mathcal {R}^n \rightarrow \mathcal R\) at x in direction d is defined by
The subdifferential (in Clarke’s sense) of f at x then is defined as
Note that, since \(p(x):= \max \{p_i(x): i \in I(x)\}\), and \(p_i\) is locally Lipschitz, differentiable, by Proposition 2.3.12 [16], we have \(\partial ^0 p(x) = co\{\nabla p_i(x): i\in I(x)\}\), where co stands for the convex hull notion. As usual, we say that \(x^*\in C\) is a stationary point of the problem \(\min _{x\in C}\varphi (x)\) (or \(x^*\) is a stationary point of \(\varphi \) over C), if \(0\in \partial ^0 \varphi (x^*)+ N_C(x^*)\), in particular, \(0\in \partial ^0 \varphi (x^*)\).
Let S denote the solution set of Problem (NSEP), then we have the following convergence result:
Theorem 2
Under the assumptions (A1)–(A4) we suppose further that f is paramonotone with respect to the solution set \(Sol({ EP})\) and that the sequence \(\{g_k\}\) is bounded. Then the sequence \(\{x_k\}\) converges to a solution of Problem (NSEP) or to a solution of equilibrium problem (EP) which is also a stationary point of problem \(\min \{p(x): x\in K\}\).
More precisely, let
then
-
(i)
If \(\sum _{k\in J} \alpha _k=+\, \infty \), the sequence \(\{x_k\}\) converges to a solution of (NSEP).
-
(ii)
If \(\sum _{k\in J} \alpha _k<+\, \infty \), the sequence \(\{x_k\}\) converges to a solution \(x^*\) of the equilibrium problem (EP), which is also a stationary point of problem \(\min \{p(x): x\in K\}\).
Remark 2
The condition for which the sequence \(\{g_k\}\) is bounded has been given in [36]. Note that in the variational inequality case where \(f(x,y) := \langle \phi (x), y-x \rangle \), the sequence \(\{g_k\}\) is bounded whenever \(\phi \) is continuous. The condition \(\sum _{k\in J} \alpha _k=+\, \infty \) is satisfied if there exists an integer number \(k_0\) such that \(\widehat{h}_k\not =0\) for \(k\ge k_0\).
Proof
(of the theorem) We divide the proof into several steps.
Step 1: First we prove that, in any case, the sequence \(\{\Vert x_k-z\Vert ^2\}\) is convergent for all \(z\in S\), and hence \(\left\{ x_k\right\} \) and \(\left\{ y_k\right\} \) are both bounded.
In fact, since f is pseudomonotone on K with respect to the solution set of \(({ EP})\), one has
On the other hand, since z is a minimizer of p, we have \(p(y_k)\ge p(z).\) Hence, from the definitions of p and \(p_i\), it follows that
Since \(h_k \in co \left\{ \nabla p_i(y_k) , \quad i\in I(y_k)\right\} \) and \(\widehat{h}_k = 0\) or \(\widehat{h}_k = \frac{h_k}{\Vert h_k\Vert },\) one has
Then, by virtue of Lemma 6 (i), we obtain
where \(A_k= 2 (\alpha _k\epsilon _k +\beta _k^2)+ \alpha ^2_k.\) Since \(\alpha _k=\frac{\beta _k}{\gamma _k}\) with \(\gamma _k=\max \{\delta _k, \Vert g_k\Vert \}\),
Since \(\delta _k>\delta >0\),
which together with \(\sum _{k=1}^{+\, \infty } \beta _k^2<+\, \infty \) implies
Using Lemma 2, we see that \(\{\Vert x_k-z\Vert ^2\}\) is convergent for every \(z\in S\). Hence, \(\{x_k\}\) is bounded. Then, since \(y_k = P_K(x_k -\alpha _k g_k)\), by Lemma 1, the sequence \(\{y_k\}\) is bounded too.
Now we consider two distinct cases.
-
(i)
Case 1. \(\sum _{k\in J} \alpha _k =+\, \infty \).
Step 2(i): We show that, for \(z\in S\),
Indeed, by Lemma 6(i), for every k, we have
Summing up we obtain
Since z is a solution of Problem (NSEP), by pseudomonotonicity of f, we have \(f(x_k,z)\le 0\) and \(\langle \widehat{h}_k, y_k-z\rangle \ge 0\). Hence,
But from \(\sum _{k\in J} \alpha _k=+\, \infty \), it holds that
Step 3(i): We show that if \(z\in S\) and \(\left\{ x_{k_j}\right\} \) is a subsequence of \(\left\{ x_k\right\} \) such that
and \(x^*\) is a cluster point of \(\left\{ x_{k_j}\right\} \), then \(x^*\) belongs to S.
Combining (18) with \(f(x_{k_j},z)\le 0\) and \(\langle \widehat{h}_{k_j}, y_{k_j}-z\rangle \ge 0\), we obtain
Without loss of generality, we can assume that \(x_{k_j}\) converges to \(x^*\) as \(j \rightarrow \infty \). Since f(., z) is upper semicontinuous,
By assumption (A4), f is pseudomonotone on K with respect to S, we have \(f(x^*,z) \le 0\) if \(z\in S\). Thus, \(f(x^*,z)=0\). Again, by the pseudomonotonicity of f, \(f(z,x^*)\le 0\). Hence, \(f(z,x^*)=f(x^*,z)=0\). Then using the paramonotonicity of f, we can conclude that \(x^*\) is a solution of (EP).
Now, it remains to prove that \(p(x^*)=0\). Indeed, otherwise, there exists \(\alpha >0\) such that
Since \(\lim _{j \rightarrow \infty } x_{k_j}=x^*\) and \(\lim _{k \rightarrow \infty } \Vert x_k-y_k\Vert \le \lim _{k \rightarrow \infty } \beta _k =0\), we have \(\lim _{j \rightarrow \infty } y_{k_j}=x^*\). Thus,
and therefore there exists \(j_0\in \mathbb {N}\) such that
Note that \(0=p(z)<\frac{\alpha }{2}\) for every \(z\in S\). Since p is continuous, there exists \(\epsilon >0\) such that
Then \(p(y)<p(y_{k_j})\) for all \(j\ge j_0\) and \(y\in B(z,\epsilon )\). Since \(\widehat{h}_{k_j}\not = 0\) for every j, by applying Lemma 6(ii), we obtain
which contradicts with (19). Thus \(p(x^*)=0\), which means that \(x^*\in S\).
Step 4(i): Now we prove that \(\left\{ x_k\right\} \) converges to a solution of (NSEP).
In Step 1(i), we have seen that the sequence \(\left\{ \Vert x_k-z\Vert \right\} \) converges for every \(z\in S\). Combining this fact with Step 3(i) to obtain
-
(ii)
Case 2. \(\sum _{k\in J} \alpha _k <+\, \infty \)
Step 2(ii): We show that, for \(z\in S\),
Indeed, by Lemma 6, for every k, we have
Summing up we obtain
By the assumption that \(\{\Vert g_k\Vert \}\) is bounded, there exists \(L>\delta \) such that \(\Vert g_k\Vert \le L\) for every k. Thus,
Hence,
Since z is a solution of Problem (NSEP), by pseudomonotonicity of f, we have \(f(x_k,z)\le 0\) and \(\langle \widehat{h}_k, y_k-z\rangle \ge 0\). The it follows from (21) that
But from \(\sum _{k=1}^{\infty } \frac{\beta _k}{\delta _k}=+\, \infty \), it holds that
Step 3(ii): We show that if \(z\in S\) and \(\left\{ x_{k_j}\right\} \) is a subsequence of \(\left\{ x_k\right\} \) such that
and \(x^*\) is a cluster point of \(\left\{ x_{k_j}\right\} \), then \(x^*\) belongs to Sol(EP).
Indeed, without loss of generality, we can assume that \(x_{k_j}\) converges to \(x^*\) as \(j \rightarrow \infty \). We have
Combining this with \(f(x_{k_j},z)\le 0\) and \(\langle \widehat{h}_{k_j}, y_{k_j}-z\rangle \ge 0\), we obtain
By the same argument as in Step 3(i), we can show that \(x^*\) is a solution of (EP).
Now, if there exists a subsequence of \(\left\{ k_j\right\} \) such that \(\widehat{h}_{k_j}\not =0\), then by the same argument as in Step 3(i), we can conclude that \(x^*\) belongs to S. Otherwise, there exist \(j_1\) such that for all \(j\ge j_1\),
which implies that \(\nabla p_i(y_{k_j})=0\) for all \(i \in I(y_{k_j})\). Note that \(\left\{ y_{k_j}\right\} \) also converges to \(x^*\).
Step 4(ii): We prove that the sequence \(\{\Vert x_k-x^*\Vert \}\) converges, and hence the whole sequence \(\left\{ x_k\right\} \) converges to \(x^*\).
Since \(x^*\in K\), by Lemma 6(i), one has
In Step 3(ii), we have shown that \(x^*\) belongs to Sol(EP), and, by pseudomonotonicity of f,
So,
Note that if \(k \not \in J\), then \(\widehat{h}_k =0\) and thus,
If \(k \in J\), then \(\widehat{h}_k \not =0\) and \(\Vert \widehat{h}_k\Vert =1\), hence
Note that \(\left\{ y_k\right\} \) is bounded, so there exists \(M>0\) large enough such that
Then,
The it follows from (24) that
where
Since \(\sum _{k=1}^{\infty } A_k<+\, \infty \) and \(\sum _{k\in J} \alpha _k<+\, \infty \), we have
Thanks to Lemma 2, we see that \(\{\Vert x_k-x^*\Vert ^2\}\) is convergent, and hence \(\left\{ x_k\right\} \) converges to \(x^*\), as, by Step 3(ii), it has a subsequence converging to \(x^*\). More precisely, if there exists \(N_0\) such that \(\hat{h}_k \not =0\) for every \(k\ge N_0\), then by the same argument as in Case 1, we can show that \(x^*\) is a solution of Problem (NSFP). If \(\hat{h}_k =0\) for infinitely many k, then there exists an infinite subsequence \(\{y_{k_j}\}\) such that \(\nabla p_i(y_{k_j})=0\) for every j and \(i\in I(y_{k_j})\). Let
Since \(\nabla p_i(y_{k_j})=0\) for infinitely many j, and \( I(y_{k_j}) \subseteq \{1, \ldots ,m\}\), we have \(I_0 \not =\emptyset \). Letting \(j\rightarrow \infty \) we obtain \(\nabla p_i(x^*)=0\) for every \(i \in I_0\). On the other hand, for \(i\in I_0\),
which implies \(I_0 \subseteq I(x^*)\). Thus \(0\in \text {co} (\{\nabla p_i(x^*)=0\ i\in I_0\}) \subseteq \text {co} (\{\nabla p_i(x^*)=0\ i\in I(x^*)\})\) that means that \(x^*\) is a stationary point of p over C. Since both sequences \(\{x_k\}\), \(\{y_k\}\) converging to \(x^*\), the proof is complete. \(\square \)
4 Numerical experiments
In this section, we test our algorithm, denoted by NSEP and compare it with the MACEP algorithm proposed in [37] for solving constrained equilibrium problems on 3 examples. The first one is a differentiated jointly constrained Nash equilibrium model. The second is an example taken from [37] for a Nash–Cournot model. In the last example, we consider a NSEP problem where each function \(F_i(x_i)=\log (a_i x_i+b_i)\) . The algorithms have been code in Matlab 7.8 on a 8Gb RAM core i7.
Example 1
(Application to Jointly Constrained Nash Equilibrium Models)
Suppose that there are \(i = 1, \ldots , n\) players participating in a game. Each player can take an individual action, which is represented by a vector \(x_i\) in \(\mathcal {R}\). All players together can take a collective action \(x \in \mathcal {R}^n\). Each player i uses a payoff function \(f_i\) which depends on actions of other players. Then the Nikaido–Isoda function of the game is defined as
where the vector \(x[y_i]\) is obtained from x by replacing component \(x_i\) by \(y_i\). Let \(K_i \subset \mathcal {R}\) be the strategy of player i. Then the strategy set of the game is \(K:= K_1 \times \cdots \times K_n\). We recall that a point \(x^* \in K\) is a Nash equilibrium point of the game if
It is well known that \(x^*\) is an equilibrium point if and only if \(f(x^*, y) \ge 0 \quad \forall y\in K\). In some practical models an equilibrium point is required to satisfy certain additional constraints defined as \(F(x) \in Q\), where F is a mapping from \(\mathcal {R}^n\) to Q with Q being a convex subset in another space \(\mathcal {R}^m\). Clearly, the problem of finding such an equilibrium point can take the form of Problem (NSEP).
To illustrate let us consider a model in electricity production where it is assumed that a company possesses n-plants, each of them producing a type of electricity, for instance, nuclear, solar, wind, hydro and thermoelectricity.
We suppose that the price for producing at plant i is an affine function given by \(p_i(x_1, \ldots ,x_n) := \alpha - \sum _{k=1}^n\tau _{ik}x_k\) for every i, where \(\alpha > 0 \) (in general large), \(\tau _{ik} > 0\). This price function arises in differentiated good models [22], where the users prefer the commodity produced by one plant to the other ones, for example, many users prefer solar and wind electricity to thermoelectricity or nuclear ones. Note that when \(t_{ik} = \tau \) for every i and k, the price function becomes the usual one. The profit made by plant i takes the form
where \(c_i(x_i)\) is the cost (including the fee for environmental pollution) for producing \(x_i\) by plant i. In general \(c_i\) is an increasingly convex function. The convexity means that the cost for producing a unit increases as the amount of the production gets larger.
Actually, the company seeks to maximize its profit by choosing a corresponding production level at each plant under the presumption that the production at other plants are parametric input. Let \(K_i\) be the strategy set for plant i, that is the the production level \(x_i\) must be chosen in \(K_i\). A commonly used approach to this model is based upon the famous Nash equilibrium concept by using the Nikaido–Isoda function defined in (27). This function has been used to models in electricity markets and others [17, 25].
In practice the level of the production at each plant should satisfy a certain ratio, for example, the ratio of hydroelectricity \(x_1\) and the total production \(\sum _{j\not =1} x_j\) of all other electricity types should be restricted in a given percent, that can be written as \(l_1 \le \frac{ x_1 }{\sum _{j\not =1} x_j} \le u_1\), where \(l_1\) and \(u_1\) are given constants. In this case, the problem of finding an equilibrium point that satisfies joint constraints leads to a nonlinear split feasibility problem of the form (NSEP). Coupled constraint models in electricity market and in the River Basin Pollution game was considered in some papers, see e.g. [3, 29, 40] and the references therein.
We tested the proposed algorithm with the cost function given by
and the price functions
where \(r_j\), \(q_j\) and each \(\tau _{ij}\) are randomly generated in the interval [0, 20], [0, 3] and in [0, 1 / n] respectively. For these cost and price functions, by using Proposition 3.2 in [24], one can check that the bifunction f defined by (27) with \(f_i\) being given by (28) is paramonotone.
We took the strategy set \(K_i:= [0, 6]\) for every i and we require that the ratio of each type of electricity and the sum of all electricity production is less than fifty percent, which is represented by the constraints \(0\le F_i(x)\le 0.5\) for every \(i=1, \ldots ,m\).
We choose the sequences of the parameters as
and we computed the model with \(m=5\) and different number n from 10 to 50 and
where \(I_i\) is the set of plants that produce type i of electricity.
The computational results are shown in Table 1 with different sizes, one hundred problems have been tested for each size. We stopped the computation at iteration k if \(\Vert x_k - y_k\Vert \le 10^{-4}\) and \(p(x_k)\le 10^{-4}\) or the number of iterations exceeds 20,000.
Figure 1 shows that the errors \(p(x_k)\) and \(\Vert x_k-y_k\Vert \) both go to 0 as the number of iterations k goes to \(+\, \infty \). We can see that in this example, both of the two errors decrease quickly as the number of iteration k gets larger. However, in some cases, the value of function \(p(x_k)\) may not be monotone decreasingly. Nevertheless, for all tested problems, the value of the function p at the stoping iterate is very small.
We also tested the algorithm with different values of \(\beta _k\): \(\frac{7}{2(k+1)}\), \(\frac{10}{(k+1)}\), \(\frac{20}{(k+1)}\)\(\frac{50}{k+1}\), \(\frac{100}{k+1}\). We observe that the choice of the parameter \(\beta _k\) plays a crucial role for efficiency of the algorithm, in general, the difference \(\Vert x_k-y_k\Vert \) goes to 0 quite slow, when this parameter is too small or too big, see Fig. 2.
Recently, the authors in [37] considered the equilibrium problem
where C, D are convex subsets in \(\mathcal {R}^n\) and f is a finite bifunction defined on an open set containing C and D.
In order to compare our algorithm, denoted by NSEP, with the one, denoted by MACEP, in [37] on the above example, we formulate Problem (NSEP) in the form of Problem EP(C, D, f) by taking \(C:\equiv K\), \(D:= \{ x: F(x) \in Q\}\). and apply the algorithm MACEP in [37] to solve the model. We try to use the same stopping criteria \(\max \{\Vert x_k-y_k\Vert ,p(x_k)\} \le 10^{-4} \), and see that the algorithm MACEP takes a lot of time even in the small dimension \(n=10,k=5\).
Table 2 reports the computed results in average on the first 150-iterations for these two algorithms.
From the table, one can see that, for this example, the computational CPU time of MACEP algorithm is less than that of NSEP algorithm in the first 150-iterations, but the error obtained by NSFP goes to zero more quickly than MACEP.
The plot about error of two algorithms can be seen in Fig 3:
Figure 3 shows that the errors of MACEP at the first 30 iterations decrease more quickly, but from iteration 30, the NSFP is convergent to zero more quickly.
Example 2
([37])
We test our algorithm on Example 4.3 in [37], where the bifunction is given by
with
As in Example 4.3 ([37]),
By setting
and \(F(x)=Mx\) where
we can easily see that \(D=\left\{ x|F(x)\in Q\right\} \).
We choose \(x^0=(1;2;1;1;1)\) and \(\lambda _k=k/(k+1),\rho _k=1,\beta _k=40/k\), \(x^0=(1;2;1;1;1)\). The computed results are reported in table 3:
Example 3
We consider the (NSEP) problem where
with \(F_i(x_i)= \log (a_ix_i+b_i)\), \(a=(1,2,3,4,5)\), \(b_i=14\) for all \(i=1,\dots ,5\), and
The bifuction f is defined by
where \(q=( -28, -29,-27, -29, -29, -29, -28, -29, -28, -29)\),
Note that this bifunction f satisfies all assumptions for ensuring convergence of both algorithms. We take \(\rho _k=3,\beta _k=7/2(k+1)\), \(x^0=(3;3;3;3;3; 0; 0; 0; 0; 0)\) and test MACEP and our algorithm. Computed results are reported in Tables 4 and 5.
5 Conclusion
We have proposed an algorithm for solving the split feasibility problem of finding a solution x of a paramonotone equilibrium problem such that F(x) belongs to a convex set in another space, where F can be a quasilinear mapping. The proposed algorithm is a combination between the projection one for both the equilibrium and inclusion problems. Applications to jointly constrained Nash equilibrium problem have been discussed and some preliminary computational results for coupled constraint Nash–Cournot models in electricity production have been reported.
References
Alber, Y.I., Iusem, A.N., Solodov, M.V.: On the projected subgradient method for nonsmooth convex optimization in a Hilbert space. Math. Program. 81, 23–35 (1998)
Anh, T.V., Muu, L.D.: A projection-fixed point method for a class of bilevel variational inequalities with split fixed point constraints. Optimization 65, 1129–1143 (2016)
Aussel, D., Bendottib, P., Pitek, M.: Nash equilibriumin a pay-as-bid electricity market Part 2—best response of a producer. Optimization 66, 1027–1053 (2017)
Avriel, M.: Nonlinear Programming: Analysis and Methods. Prentice-Hall, Englewood Cliffs (1976)
Bauschke, H.H., Combettes, P.L.: Convex Analysis and Monotone Operator Theory in Hilbert Spaces. Springer, Berlin (2011)
Bigi, G., Castellani, M., Pappalardo, M., Panssacantando, M.: Existence and solution methods for equilibria. Eur. Oper. Res. 227, 1–11 (2013)
Blum, E., Oettli, W.: From optimization and variational inequalities to equilibrium problems. Math. Stud. 63, 127–149 (1994)
Bnouhachem, A., Noor, M.A., Khalfaoui, M., Zhaohan, S.: On descent-projection method for solving the split feasibility problems. J. Glob. Optim. 54, 627–639 (2012)
Byrne, C.: Iterative oblique projection onto convex sets and the split feasibility problems. Inverse Prob. 18, 441–453 (2002)
Byrne, C.: A unified treatment of some iterative algorithms in signal processing and image reconstruction. Inverse Prob. 20, 103–120 (2004)
Censor, Y., Bortfeld, T., Martin, B., Trofimov, A.: A unified approach for inversion problems in intensity-modulated radiation therapy. Phys. Med. Biol. 51, 2353–2365 (2006)
Censor, Y., Elfving, T.: A multiprojections algorithm using Bregman projections in a product spaces. Numer. Algorithms 8, 221–239 (1994)
Censor, Y., Elfving, T., Kopf, N., Bortfeld, T.: The multiple-sets split feasibility problem and its applications for inverse problems. Inverse Prob. 21, 2071–2084 (2005)
Censor, Y., Gibali, A., Reich, S.: Algorithms for the split variational inequality problem. Numer. Algorithms 59, 301–323 (2012)
Censor, Y., Segal, A.: The split common fixed point problems for directed operators. J. Convex Anal. 16, 587–600 (2009)
Clarke, F.H.: Optimization and Nonsmooth Analysis. SIAM, Philadelphia (1983)
Contreras, J., Klusch, M., Krawczyk, J.B.: Numerical solution to Nash–Cournot equilibria in coupled constraint electricity markets. EEE Trans. Power Syst. 19, 195–206 (2004)
Deepho, J., Kumam, W., Kumam, P.: A new hybrid projection algorithm for solving the split generalized equilibrium problems and the system of variational inequality problems. J. Math. Model. Algorithms 13, 405423 (2014)
Dragomir, S.S., Pearce, C.E.M.: Jensen’s inequality for quasiconvex functions. Numer. Algebra Control Optim. 2(2), 279–291 (2012)
Finetti, D., Sulle, B.: Stratificazioni conversse. Ann. Mat. Pura Appl. 30, 173–183 (1949)
Giorgi, G.: A simple way to prove the characterization of differentiable quasiconvex functions. Appl. Math. 5, 1226–1228 (2014)
Gorbachuk, V.M.: Cournot–Nash and Bertrand–Nash equilibria for heterogeneous duopoly of differentiated products. Cyber. Syst. Anal. 46, 25–26 (2010)
Hieu, D.V.: Two hybrid algorithms for solving split equilibrium problems. Int. J. Comput. Math. 95(3), 561–583 (2018)
Iusem, A.N.: On some properties of paramonotone operator. Convex Anal. 5, 269–278 (1998)
Jing-Yuan, W., Smeers, Y.: Spatial oligopolistic electricity models with Cournot generators and regulated transmission prices. Oper. Res. 47, 102112 (1999)
Konnov, I.V.: Combined Relaxation Methods for Variational Inequalities. Springer, Berlin (2000)
Kiwiel, K.C.: Convergence and efficiency of subgradient methods for quasiconvex minimization. Math. Program. Ser. A 90, 1–25 (2001)
Kraikaew, R., Saejung, S.: On split common fixed point problems. J. Math. Anal. Appl. 415, 513–524 (2014)
Krawczyk, J.B., Uryasev, S.: Relaxation algorithms to find Nash equilibria with economic applications. Environ. Model. Assess. 5, 63–73 (2000)
Li, Z., Hana, D., Zhang, W.: Self-adaptive projection-type method for nonlinear multiple-sets split feasibility problem. Inverse Probl. Sci. Eng. 21, 155–170 (2013)
Lopez, G., Martin-Maquez, V., Wang, F., Xu, H.K.: Solving the split feasibility problem without prior knowledge of matrix norms. Inverse Prob. 28, 085–094 (2012)
Luc, D.L.: Characterisations of quasiconvex functions. Bull. Aust. Math. Soc. 48, 393–406 (1993)
Moudafi, A.: Split monotone variational inclusions. J. Optim. Theory Appl. 150, 275–283 (2011)
Moudafi, A., Thakur, B.S.: Solving proximal split feasibility problems without prior knowledge of operator norms. Optim. Lett. 8, 2099–2110 (2014)
Muu, L.D., Oettli, W.: Convergence of an adaptive penalty scheme for finding constrained equilibria. Nonlinear Anal. 18, 1159–1166 (1992)
Santos, P., Scheimberg, S.: An inexact subgradient algorithm for equilibrium problems. Comput. Appl. Math. 30, 91–107 (2011)
Santos, P., Scheimberg, S.: A modified projection algorithm for constrained equilibrium problems. Optimization 66(12), 2051–2062 (2017)
Shehu, Y., Ogbuisi, F.U.: Convergence analysis for proximal split feasibility problems and fixed point problems. J. Appl. Math. Comput. 48, 221–239 (2015)
Vuong, P.T., Strodiot, J.J., Nguyen, V.H.: A gradient projection method for solving split equality and split feasibility problems in Hilbert spaces. Optimization 64, 2321–2341 (2015)
Yen, L.H., Muu, L.D., Huyen, N.T.T.: An algorithm for a class of split feasibility problems: application to a model in electricity production. Math. Methods Oper. Res. 84(3), 549–565 (2016)
Acknowledgements
We would like to thank the editor and the referees very much for their useful comments, remarks and suggestions that helped us very much to improve quality of the paper. The first author would also like to thank the Vietnam Institute for Advanced Study in Mathematics (VIASM) for the support during her visit.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
This paper is supported by the NAFOSTED, Grant 101.01-2017.315.
Rights and permissions
About this article
Cite this article
Yen, L.H., Huyen, N.T.T. & Muu, L.D. A subgradient algorithm for a class of nonlinear split feasibility problems: application to jointly constrained Nash equilibrium models. J Glob Optim 73, 849–868 (2019). https://doi.org/10.1007/s10898-018-00735-0
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10898-018-00735-0