Abstract
The purpose of this paper is to study and analyze two new extragradient methods for solving non-Lipschitzian and pseudo-monotone variational inequalities in real Hilbert spaces. Under suitable conditions, weak and strong convergence theorems of the proposed methods are established. We present academic and numerical examples for illustrating the behavior of the proposed algorithms.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
We focus on the following classical variational inequality (VI) ([9, 10]) which consists in finding a point \(x^*\in C\) such that
where C is a nonempty closed convex subset in a real Hilbert space H, and is a single-valued mapping \(A: H\rightarrow H\). As commonly done, we denote by VI(C, A) the solution set of VI (1.1). Variational inequalities are fundamental in a broad range of mathematical and applied sciences; the theoretical and algorithmic foundations as well as the applications of variational inequalities have been extensively studied in the literature and continue to attract intensive research. For the current state of the art in a finite-dimensional setting, see for instance [8] and the extensive list of references therein.
Many authors have proposed and analyzed several iterative methods for solving the variational inequality (1.1). The simplest one is the following projection method, which can be seen as an extension of the projected gradient method introduced for solving optimization problems.
for each \(n\ge 1\), where \(P_C\) denotes the metric projection from H onto C. It is known that the assumptions which imply convergence of this method are quite restrictive and require A to be L-Lipschitz continuous and \(\alpha \)-strongly monotone (see Chapter 13 in [19]) and \( \lambda \in \left( 0, \dfrac{2\alpha }{L^2}\right) \).
One way to weaken the (1.2) convergence assumptions, Korpelevich [20] (also independently by [1]) proposed a double projection method known as the extragradient method in Euclidean space when A is monotone and L-Lipschitz continuous. The iterative step of the method is as follows:
where \(\lambda _n \in \left( 0,\dfrac{1}{L}\right) \).
In recent years, the extragradient method was further extended to infinite-dimensional spaces in various ways, see, e.g. [2,3,4,5, 16, 23, 24, 30,31,32, 34] and the references therein.
Obviously, when A is not Lipschitz continuous or its Lipschitz constant L is difficult to compute/approximate, Korpelevich’s method fails since the step-size \(\lambda _n\) depends on this. So, Iusem [14] proposed the following algorithm in a way to avoid this obstacle.
Algorithm 1.1
Similar extensions have been developed by many authors, for example Khobotov [18] and Marcotte [25]. These algorithms assume that A is monotone and continuous on C. Thus, in this spirit, we wish to construct an extragradient modification which converges under a weaker condition in Hilbert spaces. To be more specific, we assume that A is a uniformly continuous pseudo-monotone operator. Our scheme also make use of a different Armijo-type line-search and then A is only assumed to be pseudo-monotone on C in the sense of Karamardian [17]. Weak and strong convergence of these proposed algorithms is established in real Hilbert spaces.
The paper is organized as follows. We first recall some basic definitions and results in Sect. 2. Our algorithms are presented and analyzed in Sect. 3. In Sect. 4, we present some numerical experiments which demonstrate the algorithms performances as well as provide a preliminary computational overview by comparing it with some related algorithms.
2 Preliminaries
Let H be a real Hilbert space and C be a nonempty, closed and convex subset of H. The weak convergence of \(\{x_n\}_{n=1}^{\infty }\) to x is denoted by \(x_{n}\rightharpoonup x\) as \(n\rightarrow \infty \), while the strong convergence of \(\{x_n\}_{n=1}^{\infty }\) to x is written as \(x_n\rightarrow x\) as \(n\rightarrow \infty .\) For each \(x,y\in H\) and \(\alpha \in {\mathbb {R}}\), we have
Definition 2.1
Let \(T:H\rightarrow H\) be an operator.
-
1.
The operator T is called L-Lipschitz continuous with \(L>0\) if
$$\begin{aligned} \Vert Tx-Ty\Vert \le L \Vert x-y\Vert \ \ \ \forall x,y \in H. \end{aligned}$$(2.1)if \(L=1\) then the operator T is called nonexpansive and if \(L\in (0,1)\), T is called a contraction.
-
2.
The operator T is called monotone if
$$\begin{aligned} \langle Tx-Ty,x-y \rangle \ge 0 \ \ \ \forall x,y \in H. \end{aligned}$$(2.2) -
3.
The operator T is called pseudo-monotone if
$$\begin{aligned} \langle Tx,y-x \rangle \ge 0 \Longrightarrow \langle Ty,y-x \rangle \ge 0 \ \ \ \forall x,y \in H. \end{aligned}$$(2.3) -
4.
The operator T is called \(\alpha \)-strongly monotone if there exists a constant \(\alpha >0\) such that
$$\begin{aligned} \langle Tx-Ty,x-y\rangle \ge \alpha \Vert x-y\Vert ^2 \ \ \forall x,y\in H. \end{aligned}$$ -
5.
The operator T is called sequentially weakly continuous if for each sequence \(\{x_n\}\) we have: \(\{x_n\}\) converges weakly to x implies \({Tx_n}\) converges weakly to Tx.
It is easy to see that every monotone operator is pseudo-monotone, but the converse is not true.
For every point \(x\in H\), there exists a unique nearest point in C, denoted by \(P_Cx\) such that \(\Vert x-P_Cx\Vert \le \Vert x-y\Vert \ \forall y\in C\). \(P_C\) is called the metric projection of H onto C. It is known that \(P_C\) is nonexpansive.
Lemma 2.2
[12] Let C be a nonempty closed convex subset of a real Hilbert space H. Given \(x\in H\) and \(z\in C\). Then \(z=P_Cx\Longleftrightarrow \langle x-z,z-y\rangle \ge 0 \ \ \forall y\in C.\)
Lemma 2.3
[12] Let C be a closed and convex subset in a real Hilbert space H, \(x\in H\). Then
-
(i)
\(\Vert P_Cx-P_Cy\Vert ^2\le \langle P_C x-P_C y,x-y\rangle \ \forall y\in C\);
-
(ii)
\(\Vert P_C x-y\Vert ^2\le \Vert x-y\Vert ^2-\Vert x-P_Cx\Vert ^2 \ \forall y\in C;\)
-
(iii)
\(\langle (I-P_C)x-(I-P_C)y,x-y\rangle \ge \Vert (I-P_C)x-(I-P_C)y\Vert ^2 \ \forall y\in C.\)
For properties of the metric projection, the interested reader could be referred to [12, Section 3].
The following Lemmas are useful for the convergence of our proposed methods.
Lemma 2.4
[7] For \(x\in H\) and \(\alpha \ge \beta >0\) the following inequalities hold.
Lemma 2.5
[15] Let \(H_1\) and \(H_2\) be two real Hilbert spaces. Suppose \(A : H_1 \rightarrow H_2\) is uniformly continuous on bounded subsets of \(H_1\) and M is a bounded subset of \(H_1\). Then A(M) (the image of M under A) is bounded.
Lemma 2.6
([6], Lemma 2.1) Consider the problem VI(C, A) with C being a nonempty, closed, convex subset of a real Hilbert space H and \(A : C \rightarrow H\) being pseudo-monotone and continuous. Then, \(x^*\) is a solution of VI(C, A) if and only if
Lemma 2.7
[26] Let C be a nonempty set of H and \(\{x_n\}\) be a sequence in H such that the following two conditions hold:
-
(i)
for every \(x\in C\), \(\lim _{n\rightarrow \infty }\Vert x_n-x\Vert \) exists;
-
(ii)
every sequential weak cluster point of \(\{x_n\}\) is in C.
Then \(\{x_n\}\) converges weakly to a point in C.
Lemma 2.8
[22] Let \(\{a_n\}\) be a sequence of non-negative real numbers such that there exists a subsequence \(\{a_{n_i}\}\) of \(\{a_n\}\) such that \(a_{n_{i}}<a_{n_{i}+1}\) for all \(i\in {\mathbb {N}}\). Then there exists a nondecreasing sequence \(\{m_k\}\) of \({\mathbb {N}}\) such that \(\lim _{k\rightarrow \infty }m_k=\infty \) and the following properties are satisfied by all (sufficiently large) number \(k\in {\mathbb {N}}\):
In fact, \(m_k\) is the largest number n in the set \(\{1,2,\ldots ,k\}\) such that \(a_n<a_{n+1}\).
The next technical lemma is very useful and used by many authors, for example Liu [21] and Xu [35]. Furthermore, a variant of Lemma 2.9 has already been used by Reich in [29].
Lemma 2.9
Let \(\{a_n\}\) be sequence of non-negative real numbers such that:
where \(\{\alpha _n\}\subset (0,1)\) and \(\{b_n\}\) is a sequence such that
-
(a)
\(\sum _{n=0}^\infty \alpha _n=\infty \);
-
(b)
\(\limsup _{n\rightarrow \infty }b_n\le 0.\)
Then \(\lim _{n\rightarrow \infty }a_n=0.\)
3 Main results
In this section, we introduce the two new extragradient modifications for solving the VI (1.1). We present the weak and strong convergence of the schemes under the assumptions.
Condition 3.1
The feasible set C of the VI (1.1) is a nonempty, closed, and convex subset of the real Hilbert space H.
Condition 3.2
The VI (1.1) associated operator \(A:C\rightarrow H\) is a pseudo-monotone, sequentially weakly continuous on C and uniformly continuous on bounded subsets of C.
Condition 3.3
The solution set of the VI (1.1) is nonempty, that is VI\((C,A)\ne \emptyset \).
3.1 Weak convergence
Algorithm 3.1
We start the algorithm’s convergence analysis by proving that (3.1) terminates after finite steps.
Lemma 3.2
Assume that Conditions 3.1–3.3 hold. The Armijo line-search rule (3.1) is well defined. In addition, we have \(\lambda _n\le \gamma .\)
Proof
If \(x_n\in VI(C,A)\) then \(x_n=P_{C}(x_n-\gamma Ax_n)\) and \(m_n=0\). We consider the situation \(x_n\notin VI(C,A)\) and assume the contrary that for all m we have
By Cauchy–Schwartz inequality, we have
Combining (3.2) and (3.3) we find
This implies that
Since \(x_n\in C\) for all n and \(P_C\) is continuous, we have \(\lim _{m\rightarrow \infty }\Vert x_n-P_C(x_n-\gamma l^m Ax_n)\Vert =0.\) Since A is uniformly continuous on bounded subsets of C (Condition 3.2), we get that
Combining (3.4) and (3.5) we get
Assume that \(z_m=P_C(x_n-\gamma l^m Ax_n)\) we have
This implies that
Taking the limit \(m\rightarrow \infty \) in (3.7) and using (3.6), we obtain
which implies that \(x_n\in VI(C,A)\) this is a contraction. \(\square \)
Remark 3.3
-
1.
In the proof of Lemma 3.2 we do not use the pseudomonotonicity of A.
-
2.
Now we show that if \(x_n=y_n\) then stop and \(x_n\) is a solution of VI(C, A). Indeed, we have \(0<\lambda _n\le \gamma \), which together with Lemma 2.4 we get
$$\begin{aligned} 0=\dfrac{\Vert x_n-y_n\Vert }{\lambda _n}=\dfrac{\Vert x_n-P_C(x_n-\lambda _n Ax_n)\Vert }{\lambda _n}\ge \dfrac{\Vert x_n-P_C(x_n-\gamma Ax_n)\Vert }{\gamma }. \end{aligned}$$This implies that \(x_n\) is a solution of VI(C, A).
Lemma 3.4
Assume that Conditions 3.1–3.3 hold and let \(\{x_n\}\) be any sequence generated by Algorithm 3.1. Then we have
Proof
Recall one of the metric projection property
By denoting \(y=x_n-\lambda _n Ax_n\) and \(x=x_n\), we get
thus
which, together with \(\lambda _n\le \gamma \) we find
\(\square \)
Lemma 3.5
Assume that Conditions 3.1–3.3 hold and let \(\{x_n\}\) be any sequence generated by Algorithm 3.1. Then we have
Proof
Indeed, let \(p\in VI(C,A)\), since \(y_n\in C\), we have \(\langle Ap, y_n-p\rangle \ge 0\). Due to the pseudomonotonicity of A, we get
On the other hand, according to Lemma 3.4 we have
Now, using (3.1), (3.8) and (3.9), we get
Since \(\lambda _n\le \gamma \) we get
\(\square \)
Remark 3.6
From Lemma 3.5 we see that if \(Ay_n=0\) then \(x_n=y_n\), this implies that \(x_n\) is a solution of VI(C, A).
Lemma 3.7
Assume that Conditions 3.1–3.3 hold and let \(\{x_n\}\) be any sequence generated by Algorithm 3.1. If there exists a subsequence \(\{x_{n_k}\}\) of \(\{x_n\}\) such that \(\{x_{n_k}\}\) converges weakly to \(z\in C\) and \(\lim _{k\rightarrow \infty }\Vert x_{n_k}-y_{n_k}\Vert =0\), then \(z\in VI(C,A).\)
Proof
We have \(y_{n_k}=P_C(x_{n_k}-\lambda _{n_k}Ax_{n_k})\) thus,
or equivalently
This implies that
Now, we show that
For showing this, we consider two possible cases. Suppose first that \(\liminf _{k\rightarrow \infty }\lambda _{n_k}>0\). We have \(\{x_{n_k}\}\) is a bounded sequence, A is uniformly continuous on bounded subsets of C. By Lemma 2.6, we get that \(\{Ax_{n_k}\}\) is bounded. Taking \(k\rightarrow \infty \) in (3.10) since \(\Vert x_{n_k}-y_{n_k}\Vert \rightarrow 0\), we get
Now, we assume that \(\liminf _{k\rightarrow \infty }\lambda _{n_k}=0\). Assume \(z_{n_k}=P_{C}(x_{n_k}-\lambda _{n_k}.l^{-1}Ax_{n_k})\), we have \(\lambda _{n_k}l^{-1}>\lambda _{n_k}\). Applying Lemma 3.5, we obtain
Consequently, \(z_{n_k}\rightharpoonup z\in C\), this implies that \(\{z_{n_k}\}\) is bounded, and due to Condition 3.2, we get that
By the Armijo line-search rule (3.1), we have
That is,
Combining (3.12) and (3.13), we obtain
Furthermore, we have
This implies that
Taking the limit \(k\rightarrow \infty \) in (3.14), we get
Therefore, the inequality (3.11) is proved. Next, we show that \(z\in \mathrm{VI}(C,A)\).
Now we choose a sequence \(\{\epsilon _k\}\) of positive numbers decreasing and tending to 0. For each k, we denote by \(N_k\) the smallest positive integer such that
where the existence of \(N_k\) follows from (3.11). Since \(\{ \epsilon _k\}\) is decreasing, it is easy to see that the sequence \(\{N_k\}\) is increasing. Furthermore, for each k, since \(\{x_{N_k}\}\subset C\) we have \(Ax_{N_k}\ne 0\) and, setting
we have \(\langle Ax_{N_k}, x_{N_k}\rangle = 1\) for each k. Now, we can deduce from (3.15) that for each k
Since the fact that A is pseudo-monotone, we get
This implies that
Now, we show that \(\lim _{k\rightarrow \infty }\epsilon _k v_{N_k}=0\). Indeed, we have \(x_{n_k}\rightharpoonup z \text { as } k \rightarrow \infty \). Since A is sequentially weakly continuous on C, \(\{ Ax_{n_k}\}\) converges weakly to Az. We have that \(Az \ne 0\) (otherwise, z is a solution). Since the norm mapping is sequentially weakly lower semicontinuous, we have
Since \(\{x_{N_k}\}\subset \{x_{n_k}\}\) and \(\epsilon _k \rightarrow 0\) as \(k \rightarrow \infty \), we obtain
which implies that \(\lim _{k\rightarrow \infty } \epsilon _k v_{N_k} = 0.\)
Now, letting \(k\rightarrow \infty \), then the right-hand side of (3.16) tends to zero by A is uniformly continuous, \(\{x_{N_k}\}, \{v_{N_k}\}\) are bounded and \(\lim _{k\rightarrow \infty }\epsilon _k v_{N_k}=0\). Thus, we get
Hence, for all \(x\in C\) we have
By Lemma 2.6, we obtain \(z\in VI(C,A)\) and the proof is complete. \(\square \)
Remark 3.8
When the mapping A is monotone, it is not necessary to impose the sequential weak continuity on A.
Theorem 3.9
Assume that Conditions 3.1–3.3 hold. Then any sequence \(\{x_n\}\) generated by Algorithm 3.1 converges weakly to an element of VI(C, A).
Proof
We divide the proof into two claims.
Claim 1
We show that
Indeed, since \(P_C\) is nonexpansive we get
which, together with Lemma 3.5 we obtain
Substituting \(\beta _n=\dfrac{1-\mu }{\gamma }\dfrac{\Vert x_n-y_n\Vert ^2}{\Vert Ay_n\Vert ^2}\) into (3.17), we get
Claim 2
Now, we show that \(\{x_n\}\) converges weakly to an element of VI(C, A). Thanks to Claim 1, we have
This implies that for all \(p\in VI(C,A)\) then \(\lim _{n\rightarrow \infty }\Vert x_{n}-p\Vert \) exists, thus the sequence \(\{x_n\}\) is bounded. Consequently, \(\{y_n\}\) is bounded.
On the other hand, according to Claim 1, we get
which implies that
Since \(\{y_n\}\subset C\) is bounded, A is uniformly continuous on bounded subsets of C, according to Lemma 2.5 we get \(\{Ay_n\}\) is bounded, which together with (3.18) we get
Since \(\{x_n\}\) is a bounded sequence, there exists the subsequence \(\{x_{n_k}\}\) of \(\{x_{n}\}\) such that \(\{x_{n_k}\}\) converges weakly to \(z\in C\). It implies from Lemma 3.7 and (3.19) that \(z\in \mathrm{VI}(C,A)\).
Therefore, we showed that:
-
(i)
For every \(p \in \mathrm{VI}(C,A)\), then \(\lim _{n\rightarrow \infty }\Vert x_n-p\Vert \) exists;
-
(ii)
Every sequential weak cluster point of the sequence \(\{x_n\}\) is in VI(C, A).
By Lemma 2.7 the sequence \(\{x_n\}\) converges weakly to an element of \(\mathrm{VI}(C,A).\)
\(\square \)
Remark 3.10
Our result is more general than related results in the literature and hence might be applied for a wider class of mappings. For example, we next present the advantage of our method compared with the recent result [33, Theorem 3.1].
As in Theorem 3.9, \(A:C\rightarrow H\) is assumed to be uniformly continuous on bounded subsets instead of Lipschitz continuous in [33].
3.2 Strong convergence
In this subsection, we introduce our second extragradient modification which is based on Halpern method [13] (see also [28]) and hence has a strong convergence property. An additional assumption for the analysis of our method is the following.
Condition 3.4
Let \(\{\alpha _n\}\) be a real sequences in (0, 1) such that
The proposed algorithm is of the form
Algorithm 3.11
Theorem 3.12
Assume that Conditions 3.1–3.4 hold. Then any sequence \(\{x_n\}\) generated by Algorithm 3.11 converges strongly to \(p\in \mathrm{VI}(C,A)\), where \(p=P_{\mathrm{VI}(C,A)}x_0.\)
Proof
Similar to the proof of Theorem 3.9, and in order to keep it simple, we divide the proof into four claims.
Claim 1
We prove that \(\{x_n\}\) is bounded. Indeed, thanks to Claim 1 in the proof of Theorem 3.9, we have
This implies that
Using (3.21), we have
Thus, the sequence \(\{x_n\}\subset C\) is bounded. Consequently, the sequences \(\{y_n\}, \{z_n\}, \{Ay_n\}\) are bounded.
Claim 2
We prove that
for some \(M>0.\) Indeed, we have
Substituting (3.20) into (3.23), we get
This implies that
where \(D:=\sup \{2\Vert x_0-p\Vert \Vert x_{n+1}-p\Vert : n\in {\mathbb {N}}\}\).
Claim 3
We prove that
Using (3.21) we have
Claim 4
Now, we will show that the sequence \(\{\Vert x_n-p\Vert ^2\}\) converges to zero by considering two possible cases on the sequence \(\{\Vert x_n-p\Vert ^2\}\).
Case 1: There exists an \(N\in {\mathbb N}\) such that \(\Vert x_{n+1}-p\Vert ^2\le \Vert x_n-p\Vert ^2\) for all \(n\ge N.\) This implies that \(\lim _{n\rightarrow \infty }\Vert x_n-p\Vert ^2\) exists. It implies from Claim 2 that
On the other hand,
Using (3.24), it implies that
Since \(\{x_n\}\subset C\) is a bounded sequence, we assume that there exists a subsequence \(\{x_{n_j}\}\) of \(\{x_n\}\) such that \(x_{n_j}\rightharpoonup z\in C\) and
Since \( \lim _{n\rightarrow \infty } \Vert x_n-y_n\Vert =0\) and \(x_{n_j}\rightharpoonup z\in C\), according Lemma 3.7, we get \(z\in VI(C,A)\). On the other hand,
Thus
Using (3.26), \(p=P_{VI(C,A)}x_0\) and \(x_{n_k} \rightharpoonup z\in VI(C,A)\), we get
which, together with Claim 3, it implies from Lemma 2.9 that
Case 2: There exists a subsequence \(\{\Vert x_{n_j}-p\Vert ^2\}\) of \(\{\Vert x_{n}-p\Vert ^2\}\) such that \(\Vert x_{n_j}-p\Vert ^2 < \Vert x_{n_j+1}-p\Vert ^2\) for all \(j\in {\mathbb {N}}\). In this case, it follows from Lemma 2.8 that there exists a nondecreasing sequence \(\{m_k\}\) of \({\mathbb {N}}\) such that \(\lim _{k\rightarrow \infty }m_k=\infty \) and the following inequalities hold for all \(k\in {\mathbb {N}}\):
According to Claim 2, we have
This implies that
As proved in the first case, we obtain
and
and
Combining (3.27) and Claim 3, we obtain
This implies that
which, together with (3.27) we get
It implies from (3.28) and (3.29) that \(\limsup _{k\rightarrow \infty }\Vert x_{k}-p\Vert ^2=0\), that is \(x_k\rightarrow p\) as \(k\rightarrow \infty .\) \(\square \)
To end this section, we next present an academic example of variational inequality problem in an infinite dimensional space, where the cost function A is pseudo-monotone, L-Lipschitz continuous and sequentially weakly continuous on C but A fails to be a monotone mapping on H.
Example
Consider the Hilbert space
equipped with the inner product and induced norm on H:
for any \(u=(u_1,u_2,\ldots ,u_n,\ldots ), v=(v_1,v_2,\ldots ,v_n,\ldots )\in H.\)
Consider the set and the mapping:
where \(\alpha >1\) is a positive real number.
With this C and A, it is easy to see that VI\((C,A)=\{0\}\) and moreover, A is pseudo-monotone, sequentially weakly continuous and uniformly continuous on C but A fails to be Lipschitz continuous on H.
First observe that since \(\alpha >1\), we get that
Now let \(u,v\in C\) be such that \(\langle Au,v-u\rangle \ge 0\). This implies that \(\langle u,v-u\rangle \ge 0\).
Consequently,
meaning that A is pseudo-monotone.
Now, since C is compact, the mapping A is uniformly continuous and sequentially weakly continuous on C.
Finally, we show that A is not Lipschitz continuous on H. Assume to the contrary that A is Lipschitz continuous on H, i.e., there exists \(L>0\) such that
Let \(u=(L,0,...,0,...)\) and \(v=(0,0,...,0,...)\), then
Thus, \(\Vert Au-Av\Vert \le L\Vert u-v\Vert \) is equivalent to
equivalently
which implies that \(\alpha <1\), and this leads to a contraction and thus A is not Lipschitz continuous on H.
Remark 3.13
It should be emphasized here that the example established in Section 4 in [33] is not sequentially weakly continuous.
Remark 3.14
Thank to the referee’s comment, we wish to point out that since the proximity operator of a proper, lower semicontinuous and convex function is a generalization of the metric projection and our convergence analysis mainly use the firmly nonexpansiveness of the metric projection, and it is known that the proximity operator is indeed firmly nonexpansive, our proposed method can be modified to solve a more general variational inequalities.
4 Numerical illustrations
In this section, we present an example illustrating the behavior and advantages of our proposed schemes. The numerical example which is the Kojima–Shindo Nonlinear Complementarity Problem (NCP), see e.g., [27].
Example
In this example, we test our algorithms behavior for solving The Kojima–Shindo Nonlinear Complementarity Problem (NCP) with \(n=4\), see e.g., [27]. The cpu time is measured in seconds using the intrinsic MATLAB function cpu time. The VI feasible set is \(C:=\{x\in {\mathbb {R}}_{+}^{4}\mid x_{1}+x_{2}+x_{3}+x_{4}=4\}\) and A is given as follows:
The solution if the problem is \((\sqrt{6}/2,0,0,0.5)\). In our experiments, we choose the stopping criteria as \(\left\| x_{n}-y_n\right\| \le 10^{-3}\). The projection onto the feasible set C is performed using CVX version 1.22. Other parameters are: \(\gamma =0.2, l=0.3, \mu =0.6\), we choose the stopping criterion \(||x_n-y_n||<10^{-5}\). In [11, Algorithm 3.1], we choose \(\varepsilon =0.2, \beta =0.5\) and \(\alpha _{-1}=0.7\).
The starting point for all experiments is \(x_0=(1,1,1,1)\). All computations were performed using MATLAB R2017a on an Intel Core i5-4200U 2.3 GHz running 64-bit Windows. In Fig. 1, the performances of Algorithms 3.1, 3.11 and [11, Algorithm 3.1] are presented.
5 Conclusions
In this paper, we proposed two extragradient extensions for solving non-Lipschitzian pseudo-monotone variational inequalities in real Hilbert spaces. Under suitable and standard conditions, we establish weak and strong convergence theorems of the proposed schemes. Our work extends and generalizes some existing results in the literature and academic and numerical experiments demonstrate the behavior and potential applicability of the methods.
References
Antipin, A.S.: On a method for convex programs using a symmetrical modification of the Lagrange function. Ekon. i Mat. Metody. 12, 1164–1173 (1976)
Censor, Y., Gibali, A., Reich, S.: The subgradient extragradient method for solving variational inequalities in Hilbert space. J. Optim. Theory Appl. 148, 318–335 (2011)
Censor, Y., Gibali, A., Reich, S.: Strong convergence of subgradient extragradient methods for the variational inequality problem in Hilbert space. Optim. Meth. Softw. 26, 827–845 (2011)
Censor, Y., Gibali, A., Reich, S.: Extensions of Korpelevich’s extragradient method for the variational inequality problem in Euclidean space. Optimization 61, 1119–1132 (2011)
Censor, Y., Gibali, A., Reich, S.: Algorithms for the split variational inequality problem. Numer. Algorithms. 56, 301–323 (2012)
Cottle, R.W., Yao, J.C.: Pseudo-monotone complementarity problems in Hilbert space. J. Optim. Theory Appl. 75, 281–295 (1992)
Denisov, S.V., Semenov, V.V., Chabak, L.M.: Convergence of the modified extragradient method for variational inequalities with non-Lipschitz operators. Cybern. Syst. Anal. 51, 757–765 (2015)
Facchinei, F., Pang, J.S.: Finite-Dimensional Variational Inequalities and Complementarity Problems. Springer Series in Operations Research, vols. I and II. Springer, New York (2003)
Fichera, G.: Sul problema elastostatico di Signorini con ambigue condizioni al contorno. Atti Accad. Naz. Lincei VIII. Ser. Rend. Cl. Sci. Fis. Mat. Nat 34, 138–142 (1963)
Fichera, G.: Problemi elastostatici con vincoli unilaterali: il problema di Signorini con ambigue condizioni al contorno. Atti Accad. Naz. Lincei, Mem., Cl. Sci. Fis. Mat. Nat. Sez. I, VIII. Ser 7, 91–140 (1964)
Gibali, A.: A new non-Lipschitzian projection method for solving variational inequalities in Euclidean spaces. J. Nonlinear Anal. Optim. Theory Appl. 6, 41–51 (2015)
Goebel, K., Reich, S.: Uniform Convexity, Hyperbolic Geometry, and Nonexpansive Mappings. Marcel Dekker, New York (1984)
Halpern, B.: Fixed points of nonexpanding maps. Bull. Am. Math. Soc. 73, 957–961 (1967)
Iusem, A.N.: An iterative algorithm for the variational inequality problem. Comput. Appl. Math. 13, 103–114 (1994)
Iusem, A.N., Gárciga Otero, R.: Inexact versions of proximal point and augmented Lagrangian algorithms in Banach spaces. Numer. Funct. Anal. Optim. 22, 609–640 (2001)
Kassay, G., Reich, S., Sabach, S.: Iterative methods for solving systems of variational inequalities in refelexive Banach spaces. SIAM J. Optim. 21, 1319–1344 (2011)
Karamardian, S.: Complementarity problems over cones with monotone and pseudo-monotone maps. J. Optim. Theory Appl. 18, 445–454 (1976)
Khobotov, E.N.: Modifications of the extragradient method for solving variational inequalities and certain optimization problems. USSR Comput. Math. Math. Phys. 27, 120–127 (1987)
Konnov, I.V.: Equilibrium Models and Variational Inequalities. Mathematics in Science and Engineering. Elsevier, Amsterdam (2007)
Korpelevich, G.M.: The extragradient method for finding saddle points and other problems. Ekon. i Mat. Metody. 12, 747–756 (1976)
Liu, L.S.: Ishikawa and Mann iteration process with errors for nonlinear strongly accretive mappings in Banach space. J. Math. Anal. Appl. 194, 114–125 (1995)
Maingé, P.E.: A hybrid extragradient-viscosity method for monotone operators and fixed point problems. SIAM J. Control Optim. 47, 1499–1515 (2008)
Malitsky, Y.V.: Projected reflected gradient methods for monotone variational inequalities. SIAM J. Optim. 25, 502–520 (2015)
Malitsky, Y.V., Semenov, V.V.: A hybrid method without extrapolation step for solving variational inequality problems. J. Glob. Optim. 61, 193–202 (2015)
Marcotte, P.: Application of Khobotov’s algorithm to variational inequalities and network equilibrium problems. Inf. Syst. Oper. Res. 29, 258–270 (1991)
Opial, Z.: Weak convergence of the sequence of successive approximations for nonexpansive mappings. Bull. Am. Math. Soc. 73, 591–597 (1967)
Pang, J.S., Gabriel, S.A.: NE/SQP: a robust algorithm for the nonlinear complementarity problem. Math. Program. 60, 295–337 (1993)
Reich, S.: Strong convergence theorems for resolvents of accretive operators in Banach spaces. J. Math. Anal. Appl. 75, 287–292 (1980)
Reich, S.: Constructive Techniques for Accretive and Monotone Operators. Applied Nonlinear Analysis, pp. 335–345. Academic Press, New York (1979)
Solodov, M.V., Svaiter, B.F.: A new projection method for variational inequality problems. SIAM J. Control Optim. 37, 765–776 (1999)
Thong, D.V., Hieu, D.V.: Modified subgradient extragradient method for variational inequality problems. Numer. Algorithms. 79, 597–610 (2018)
Thong, D.V., Hieu, D.V.: Inertial extragradient algorithms for strongly pseudomonotone variational inequalities. J. Comput. Appl. Math. 341, 80–98 (2018)
Vuong, P.T.: On the weak convergence of the extragradient method for solving pseudo-monotone variational inequalities. J. Optim. Theory Appl. 176, 399–409 (2018)
Vuong, P.T., Shehu, Y.: Convergence of an extragradient-type method for variational inequality with applications to optimal control problems. Numer. Algorithms. https://doi.org/10.1007/s11075-018-0547-6 (2018)
Xu, H.K.: Iterative algorithms for nonlinear operators. J. Lond. Math. Soc. 66, 240–256 (2002)
Acknowledgements
The authors would like to thank Professor Simeon Reich and the referees for their comments on the manuscript which helped in improving earlier version of this paper.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Thong, D.V., Gibali, A. Extragradient methods for solving non-Lipschitzian pseudo-monotone variational inequalities. J. Fixed Point Theory Appl. 21, 20 (2019). https://doi.org/10.1007/s11784-018-0656-9
Published:
DOI: https://doi.org/10.1007/s11784-018-0656-9