Abstract
In this paper, we introduce a modified projection method and give a strong convergence theorem for solving variational inequality problems in real Hilbert spaces. Under mild assumptions, there exists a novel line-search rule that makes the proposed algorithm suitable for non-Lipschitz continuous and pseudo-monotone operators. Compared with other known algorithms in numerical experiments, it is shown that our algorithm has better numerical performance.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
We research the classical variational inequality problem, which is defined as follows: find \(x^*\in \mathscr {C}\) such that
where \(\mathscr {C}\) is a nonempty closed convex subset of a real Hilbert space \(\mathcal {H}\), \(\mathscr {A}:\mathcal {H}\rightarrow \mathcal {H}\) is a nonlinear operator. For simplicity, we denote the solution set of variational inequality problem (1) by \(Sol(\mathscr {C}, \mathscr {A})\).
The variational inequality problem plays an important role in nonlinear analysis research, which not only promotes the development of optimization problems, mathematical statistics, fixed point theory and other mathematical disciplines, but also has a very wide range of applications in engineering, mechanics, physics and economics. So far, many methods for solving variational inequalities have been produced. In this paper, we mainly study a class of projection methods.
It is well known that one of the most classical projection methods is the gradient projection method involving the strongly monotone and Lipschitz continuous operator. Due to its mandatory conditions, it was later improved by Korpelevich [1] into the following algorithm, which is often referred to as the extragradient algorithm:
where \(\tau \in (0,\frac{1}{L})\) and the operator \(\mathscr {A}\) is monotone and L-Lipschitz continuous. The extragradient algorithm was first used to solve the saddle point problem, and was later applied to the variational inequality problem by scholars. Since then, many meaningful results have been produced. For example, Yao et al. [2, 3] constructed extragradient algorithms involving Lipschitz continuous and monotone operators for solving variational inequalities and fixed point problems. Vuong [4] proved the convergence of the extragradient algorithm under the pseudo-monotone condition, and the operator is Lipschitz continuous and sequentially weakly continuous.
However, it is worth noting that when the operator \(\mathscr {A}\) is non-Lipschitz continuous or the Lipschitz constant L is not easy to calculate, the above extragradient algorithm will fail due to the difficulty in determining the value of \(\tau \). To overcome this shortcoming, the authors apply different techniques. Thong [5] used the line-search process so that the involved operators only need to satisfy uniform continuity. Tan et al. [6, 7] and Duvocelle et al. [8] introduced different self adaptive rules so that the algorithms involving Lipschitz continuous and pseudo-monotone operators do not need to calculate the Lipschitz constant. In addition, Iusem [9] combined the line-search rule with the extragradient algorithm and obtained the following algorithm:
where \(\tau _{n}:=\gamma l^{m_n},\, m_n:=\min \{m\in \mathbb {N}:\gamma l^m\Vert \mathscr {A}x_n-\mathscr {A}y_n\Vert \le \mu \Vert x_n-y_n\Vert \},\,\mu , l\in (0,1)\) and \(\lambda _n:=\frac{\langle \mathscr {A}y_n,x_n-y_n\rangle }{\Vert \mathscr {A}y_n\Vert ^2}\). Under suitable assumptions, Iusem established a weak convergence theorem without the need for the Lipschitz continuity.
On the basis of the algorithm proposed by Iusem [9], Iusem et al. [10], Thong et al. [11, 12] and Xie et al. [13] have more or less improved the line-search rule and obtained the convergence of the corresponding algorithms. Specifically, the improved algorithm in [12] is as follows:
where \(\tau _{n}:=\gamma l^{m_n},\, m_n:=\min \{m\in \mathbb {N}:\gamma l^m\langle \mathscr {A}x_n-\mathscr {A}y_n,x_n-y_n\rangle \le \mu \Vert x_n-y_n\Vert ^2\},\,\mu , l\in (0,1)\) and \(\lambda _n:=\frac{1-\mu }{\gamma }\frac{\Vert x_n-y_n\Vert ^2}{\Vert \mathscr {A}y_n\Vert ^2}\). They effectively modified the line-search rule, thereby increasing the range of choices for the sequence \(\{\tau _n\}\). Under suitable assumptions, a strong convergence theorem is obtained by introducing Halpern-type iteration.
In addition, it is noted that during the operation of the extragradient algorithm, each iteration needs to calculate the projection on the feasible set \(\mathscr {C}\) twice, but the generality of \(\mathscr {C}\) will increase the computational complexity. To this end, the authors have proposed various methods to improve this situation. Vuong et al. [14] and Reich et al. [15] proposed different projection algorithms by using different line-search rules, and obtained the corresponding strong convergence theorems by combining Halpern iteration and viscosity iteration respectively. Censor et al. [16, 17] constructed a subgradient half-space \(\{x\in \mathcal {H}:\langle x_n-\tau \mathscr {A}x_n-y_n,x-y_n\rangle \le 0\}\) to replace the feasible set \(\mathscr {C}\) in the second projection process in the extragradient algorithm, which accelerated the iteration rate of the algorithm. Tseng [18] directly omitted the second projection and used \(y_n-\tau (\mathscr {A}y_n-\mathscr {A}x_n)\) to replace the computation of \(P_\mathscr {C}(x_n-\tau \mathscr {A}y_n)\). Based on the algorithm proposed by Tseng, Reich et al. [19] and Yao et al. [20] add a single inertial term and a double inertial term respectively, and the convergence results of the related algorithms are obtained by combining self adaptive rule. In this paper, we mainly research such a projection-type algorithm modified by Vuong and Shehu [14]. The detailed algorithm is as follows:
Vuong and Shehu [14] introduced a half-space \(C_n\) in Algorithm 1 to replace the second-step projection in the iterative process, which is an interesting improvement and reduces the difficulty of projection computation. At the same time, they obtained the strong convergence theorem involving uniformly continuous, pseudo-monotone and sequentially weakly continuous operators by combining the line-search rule and Halpern-type iteration method. Based on the result of Algorithm 1, Reich et al. [15] made some improvements while speeding up the convergence of the following algorithm:
Reich et al. [15] proposed a different line-search rule than in Algorithm 1, thereby adjusting the half-space \(C_n\). The operator \(\mathscr {A}\) in Algorithm 2 is pseudo-monotone, uniformly continuous and satisfies \(\Vert \mathscr {A}q\Vert \le \liminf _{n\rightarrow \infty }\Vert \mathscr {A}x_n\Vert \) whenever \(\{x_n\}\subset \mathscr {C}\) and \(x_n\rightharpoonup q\). On the other hand, the introduced viscosity iteration method further accelerates the convergence process of Algorithm 2. Under the imposition of appropriate assumptions on the parameters, they obtained a strong convergence theorem.
Motivated and inspired by the above results, we propose a new algorithm for solving variational inequalities with uniformly continuous pseudo-monotone operators. Precisely, we create a novel line-search rule to determine the value of the key sequence and make appropriate adjustments to \(C_n\). In addition, we introduced inertial technology to speed up iteration efficiency. Finally, by comparing with known results in numerical experiments, it is confirmed that our proposed algorithm indeed has better behavior.
2 Preliminaries
In this section, we recall some basic concepts and facts.
Let \(\mathcal {H}\) be a real Hilbert space. Then
and
for every \(x,y\in \mathcal {H}\) and \(\alpha \in \mathbb {R}\).
Let \(\mathscr {C}\) be a nonempty subset of \(\mathcal {H}\). Then an operator \(\mathscr {A}:\mathscr {C} \rightarrow \mathcal {H}\) is called
(a) L-Lipschitz continuous with \(L>0\) if
If \(L\in (0,1)\), \(\mathscr {A}\) is called a contraction.
(b) monotone if
(c) pseudo-monotone if
(d) sequentially weakly continuous if for each sequence \(\{x_n\}\subset \mathscr {C}\) we have: \(x_n\rightharpoonup x\) implies \(\mathscr {A}x_n\rightharpoonup \mathscr {A}x\) as \(n\rightarrow \infty \).
For any point \(x\in \mathcal {H}\), it is obvious that there exists a unique nearest point in \(\mathscr {C}\), denoted by \(P_\mathscr {C} x\) satisfying \(\Vert x-P_\mathscr {C}x\Vert \le \Vert x-y\Vert ,\ \forall y\in \mathscr {C}\). \(P_\mathscr {C}\) is called the metric projection of \(\mathcal {H}\) onto \(\mathscr {C}\). The projection formula will be applied to numerical experiments (Sect. 4). The projection of x onto a half-space \(\mathscr {C}_{u,v}=\{x:\langle u,x\rangle \le v\} \) is computed by
Throughout this paper, let \(\mathscr {C}\) be a nonempty closed convex subset of a real Hilbert space \(\mathcal {H}\). Next, list some lemmas that will be needed later.
Lemma 1
([21]) Given \(x\in \mathcal {H}\) and \(z\in \mathscr {C}\). Then
Lemma 2
([21]) Given \(x\in \mathcal {H}\). Then
(1) \(\Vert P_\mathscr {C}x-P_\mathscr {C}y\Vert ^2\le \langle P_\mathscr {C}x-P_\mathscr {C}y,x-y\rangle , \ \forall y\in \mathcal {H}\);
(2) \(\Vert P_\mathscr {C}x-y\Vert ^2\le \Vert x-y\Vert ^2-\Vert x-P_\mathscr {C}x\Vert ^2, \ \forall y\in \mathscr {C}\).
Lemma 3
([22]) Given \(x\in \mathcal {H}\) and \(\alpha \ge \beta >0\). Then
(1) \(\Vert x-P_\mathscr {C}(x-\beta \mathscr {A}x)\Vert \le \Vert x-P_\mathscr {C}(x-\alpha \mathscr {A}x)\Vert \);
(2) \(\dfrac{\Vert x-P_\mathscr {C}(x-\alpha \mathscr {A}x)\Vert }{\alpha }\le \dfrac{\Vert x-P_\mathscr {C}(x-\beta \mathscr {A}x)\Vert }{\beta }\).
Lemma 4
([23]) Let \(\mathcal {H}_1\) and \(\mathcal {H}_2\) be two real Hilbert spaces. Suppose \(\mathscr {A}:\mathcal {H}_1\rightarrow \mathcal {H}_2\) is uniformly continuous on bounded subsets of \(\mathcal {H}_1\) and M is a bounded subset of \(\mathcal {H}_1\). Then \(\mathscr {A}(M)\) (the image of M under \(\mathscr {A}\)) is bounded.
Lemma 5
([24]) Let \(\mathscr {A}:\mathscr {C}\rightarrow \mathcal {H}\) be pseudo-monotone and continuous. Then, \(x^*\) belongs to \(Sol(\mathscr {C},\mathscr {A})\) if and only if
Lemma 6
([25]) Let h be a real-valued function on \(\mathcal {H}\) and defined \(K:=\{x\in \mathscr {C}:h(x)\le 0\}\). If K is nonempty and h is Lipschitz continuous on \(\mathscr {C}\) with modulus \(\theta >0\), then
where dist(x,K) denotes the distance of x to K.
Lemma 7
([26]) Let \(\{d_n\}\) be a sequence of non-negative real number such that there exists a subsequence \(\{d_{n_j}\}\subset \{d_n\}\) such that \(d_{n_j}<d_{n_j+1}\) for all \(j\in \mathbb {N}\). Then there exists a non-decreasing sequence \(\{m_k\}\subset \mathbb {N}\) such that \(\lim _{k\rightarrow \infty }m_k=\infty \) and the following properties are satisfied by all (sufficiently large) number \(k\in \mathbb {N}\):
In fact, \(m_k=\max \{n\in \mathbb {N}:d_n<d_{n+1}, n\le k\}\).
Lemma 8
([27]) Let \(\{d_n\}\) be a sequence of non-negative real numbers such that
where \(\{a_n\}\subset (0,1)\) and \(\{b_n\}\) is a real sequence such that
(a) \(\sum _{n=1}^\infty a_n=\infty \);
(b) \(\lim \sup _{n\rightarrow \infty }b_{n}\le 0\).
Then \(\lim _{n\rightarrow \infty }d_n=0\).
Lemma 9
([12]) Let \(\mathscr {C}\) be a nonempty closed convex subset of a real Hilbert space \(\mathcal {H}\) and \(\mathscr {A}:\mathscr {C}\rightarrow \mathcal {H}\) be pseudo-monotone, sequentially weakly continuous and uniformly continuous on the bounded subsets of \(\mathscr {C}\). There exists a subsequence \(\{w_{n_k}\}\subset \{w_{n}\}\) such that \(w_{n_k}\rightharpoonup q\in \mathscr {C}\) and \(\lim _{k\rightarrow \infty }\Vert w_{n_k}-P_\mathscr {C}(w_{n_k}-\tau _{n_k}\mathscr {A}w_{n_k})\Vert =0\), where \(\tau _{n}\) is a positive sequence. If
Then \(q\in Sol(\mathscr {C},\mathscr {A})\).
3 Main results
3.1 Strong convergence
In this section, we propose an improved projection-type method for solving variational inequality problems in Hilbert spaces. First, we give the following conditions:
(C1) The feasible set \(\mathscr {C}\) is a nonempty closed convex subset of a real Hilbert space \(\mathcal {H}\). The solution set \(Sol(\mathscr {C},\mathscr {A})\) is nonempty.
(C2) The operator \(\mathscr {A}:\mathscr {C}\rightarrow \mathcal {H}\) is uniformly continuous, pseudo-monotone and sequentially weakly continuous on \(\mathscr {C}\).
(C3) Let \(f:\mathscr {C}\rightarrow \mathscr {C}\) be a contraction mapping with a coefficient \(\rho \in [0,1)\) and \(\beta _n\) be a sequence in (0, 1) such that
The sequence \(\{\alpha _n\}\) satisfies \(\lim _{n\rightarrow \infty } \frac{\alpha _n}{\beta _n}=0\).
Next, we introduce our new algorithm.
Remark 1
-
1.
Since \(\mathscr {C}\) is a convex set, it can be deduced that all iterates \(\{x_n\}\), \(\{w_n\}\), \(\{y_n\}\) and \(\{z_n\}\) generated by Algorithm 3 belong to \(\mathscr {C}\). Therefore, the operator \(\mathscr {A}\) needs to be defined only on \(\mathscr {C}\), rather than necessarily on the entire space \(\mathcal {H}\).
-
2.
The operator \(\mathscr {A}\) is pseudo-monotone and uniformly continuous, which allows us to prove the strong convergence theorem without Lipschitz prior knowledge.
-
3.
The sequence \(\{\alpha _n\}\) could be chosen such that
$$\begin{aligned} \alpha _n= {\left\{ \begin{array}{ll} \min \{\frac{\theta _n}{\left\| x_1-x_{n}\right\| },\frac{\alpha }{2}\}, &{}\text { if }x_n\ne x_{1},\\ \frac{\alpha }{2}, &{}\text { otherwise },\\ \end{array}\right. } \end{aligned}$$where \(\alpha \) is a constant such that \(0\le \alpha <1\) and \(\left\{ \theta _n\right\} \) is a positive sequence such that \(\lim _{n\rightarrow \infty }\frac{\theta _n}{\beta _n}=0\). At this time, it is easy to see that \(\lim _{n\rightarrow \infty } \frac{\alpha _n}{\beta _n}=0\). In addition, if \(\alpha =0\), then \(w_n=x_n\), this is a trivial case. In the following text we will mainly study non-trivial situations.
-
4.
Compared to Algorithms 1 and 2, Algorithm 3 introduces a step size rule with parameter \(\gamma >0\) in (7), which allows the trial step \(\tau _n\) to start from a value other than 1 in each outer iteration. This modification makes sense because there are cases where a larger step size \(\tau _n>1\) may be acceptable (resulting in faster convergence), and cases where \(\tau _n<1\) may be smaller (in which case it is beneficial to choose \(\gamma <1\) to avoid unnecessary evaluations of \(\mathscr {A}\) in the inner loop). In addition, during the inner loop, we initially set \(m=0\), which gives \(\tau _{n}=\gamma \). We then substitute this value into (7) to see if it satisfies the condition. If it does, we output the result, otherwise we set \(m=1\) and continue iterating until the condition in (7) is satisfied. Once satisfied, we output the corresponding \(y_n\) value and continue with further iterations.
-
5.
In Algorithm 2, \(h_n(x)\) is defined as \(h_n(x)=\langle \mathscr {A}y_n,x-x_n\rangle +\dfrac{\tau _n}{2\lambda }\Vert r(x_n)\Vert ^2\), where \(\tau _n\Vert r(x_n)\Vert ^2\rightarrow 0\) is a crucial assumption for proving \(r(x_n)\rightarrow 0\). However, in the original proof, it was necessary to consider two separate cases: \(\liminf _{n\rightarrow \infty }\tau _n>0\) and \(\liminf _{n\rightarrow \infty }\tau _n=0\), as detailed in [15, Lemma 3.5]. To streamline the proof, we have made an improvement to \(h_n(x)\) in (8). This adjustment eliminates the need to consider separate cases and leads directly to the conclusion, thus simplifying the overall proof.
Lemma 10
Assume that conditions (C1-C3) hold. Then the line-search rule (7) is well defined.
Proof
If \(w_n\in Sol(\mathscr {C},\mathscr {A})\), then \(w_n = P_\mathscr {C}(w_n-\gamma \mathscr {A}w_n)\), thus (7) holds with \(m_n=0\). Next, we suppose that \(w_n\notin Sol(\mathscr {C},\mathscr {A})\) and assume the contrary. That is, for all m we have
By the Cauchy-Schwarz inequality, then
Combining (9) and (10), we have
It follows that
Since \(w_n\in \mathscr {C}\), \(P_\mathscr {C}\) is continuous and \(\mathscr {A}\) is uniformly continuous, we obtain
and
Noticing (11), we get
Let \(z_m=P_\mathscr {C}(w_n-\gamma l^m \mathscr {A} w_n)\). By Lemma 1, we have
That is
Consequently
Taking \(m\rightarrow \infty \) in (14) and using (12) and (13), we obtain
it follows that \(w_n\in Sol(\mathscr {C},\mathscr {A})\). This contradiction implies that (7) is well defined. \(\square \)
Lemma 11
Assume that conditions (C1-C3) hold. Let the function \(h_n\) be defined by (8) and \(p\in Sol(\mathscr {C},\mathscr {A})\). Then \(h_n(w_n)=\dfrac{\mu }{2}\Vert r(w_n)\Vert ^2\) and \(h_n(p)\le 0\). Particularly, \(h_n(w_n)>0\) whenever \(r(w_n)\ne 0\).
Proof
According to (8), \(h_n(w_n)=\dfrac{\mu }{2}\Vert r(w_n)\Vert ^2\) is obvious. On the other hand, from the pseudomonotonicity of \(\mathscr {A}\) and \(p\in Sol(\mathscr {C},\mathscr {A})\), we have \(\langle \mathscr {A}y_n,y_n-p\rangle \ge 0\) and
From the rule (7) and the definition of \(\{y_n\}\), we obtain
that is
According to Lemma 2 (1), we can conclude that
or equivalently
Substituting (16) and (17) into (15), we get that
Since \(\mu \in (0,1]\), then \(h_n(p)\le 0\). \(\square \)
Lemma 12
Assume that conditions (C1-C3) hold. Let \(\{w_n\}\) be a sequence generated by Algorithm 3. If there exists a subsequence \(\{w_{n_k}\}\subset \{ w_n\}\) converges weakly to \(q\in \mathscr {C}\) and \(\lim _{k\rightarrow \infty }\Vert w_{n_k}-z_{n_k}\Vert =0\), then \(q\in Sol(\mathscr {C},\mathscr {A})\).
Proof
Since \(z_{n_k}=P_\mathscr {C}(w_{n_k}-\tau _{n_k}\mathscr {A}w_{n_k})\), we have
or equivalently
Consequently
Next, we prove that
We consider the following two possible cases.
Case 1 Suppose that \(\liminf _{k\rightarrow \infty }\tau _{n_k}>0\). Since \(w_{n_k} \rightharpoonup q \in \mathscr {C}\), then \(\{w_{n_k}\}\) is bounded. As \(\mathscr {A}\) is uniformly continuous on bounded subsets of \(\mathscr {C}\), by Lemma 4, we get that \(\{\mathscr {A}w_{n_k}\}\) is bounded. Taking \(k\rightarrow \infty \) in (18), since \(\Vert w_{n_k}-z_{n_k}\Vert \rightarrow 0\), we get
Case 2 Assume that \(\liminf _{k\rightarrow \infty }\tau _{n_k}=0\). Let \(s_{n_k}=P_\mathscr {C}(w_{n_k}-\tau _{n_k}l^{-1}\mathscr {A}w_{n_k})\), we have \(\tau _{n_k}l^{-1}>\tau _{n_k}\). Applying Lemma 3, we obtain
that is
Therefore, \(s_{n_k}\rightharpoonup q\in \mathscr {C}\), it folllows that \(\{s_{n_k}\}\) is bounded. In addition,
By \(\tau _{n_k}l^{-1}>\tau _{n_k}\), we know that \(\tau _{n_k}l^{-1}\) does not satisfy (7), owing to (11), then
which implies that
Furthermore, it follows from the definition of \(\{s_{n_k}\}\) and Lemma 1 that
It follows that
Taking \(k\rightarrow \infty \) in (23), we get
By Lemma 9, \(q\in Sol(\mathscr {C},\mathscr {A})\) and the proof is completed. \(\square \)
Theorem 1
Assume that conditions (C1-C3) hold and the sequence \(\{\alpha _n\}\) is chosen such that
Then the sequence \(\{x_n\}\) generated by Algorithm 3 converges strongly to an element \(p\in Sol(\mathscr {C},\mathscr {A})\), where \(p=P_{Sol(\mathscr {C},\mathscr {A})}\circ f(p)\).
Proof
We divide the proof into five claims.
Claim 1. We prove that the \(\{x_n\}\) is bounded. Indeed, let \(s_n:=P_{C_n}(w_n)\), then
That is
Thus
It follows from the definition of \(\left\{ w_n\right\} \) that
Since
there exists \(N\in \mathbb {N}\) and a constant \(M_1>0\) such that
Combining (26) and (27), we have
Substituting (28) into (25), we get
This implies that the sequence \(\{x_n\}\) is bounded. Consequently, the sequences \(\{y_n\},\{f(w_n)\}\) and \(\{\mathscr {A}y_n\}\) are bounded too.
Claim 2. We prove that
for some \(M_2>0\). Ineeed, from (2) and (25), we have
Using (28), we get
Substituting (30) into (29), then
where \(M_2:=\sup _{n \in \mathbb {N}}\{2M_1\Vert x_n-p\Vert +\beta _nM_1^2\}\). Thus,
Claim 3. We prove that
for some \(M_3>0\). Indeed, from \(\{\mathscr {A}y_n\}\) is bounded, there exists \(L>0\) such that \(\Vert \mathscr {A}y_n\Vert \le L\). By the definition of \(h_n(x)\), for all \(u,v\in C_n\),
which implies that \(h_n(\cdot )\) is L-Lipschitz continuous on \(C_n\). From Lemma 6, we have
Applying Lemma 11, we get
Thus,
On the other hand,
where \(M_3:=\sup _{n \in \mathbb {N}}\{\Vert f(w_n)-p\Vert ^2+2M_1\Vert x_n-p\Vert +\beta _nM_1^2\}\). Therefore,
Claim 4. We prove that
for some \(M>0\). In fact, using (2), we have
By (2), (3) and (24), we obtain
Substituting (33) into (34), we get that
where \(M:=\sup _{n \in \mathbb {N}}\{\Vert w_n-p\Vert \}\).
Claim 5. We prove that \(\{\Vert x_n-p\Vert \}\) converges to zero by considering two possible cases.
Case 1 There exists \(N\in \mathbb {N}\) such that \(\Vert x_{n+1}-p\Vert ^2\le \Vert x_n-p\Vert ^2\) for all \(n\ge N\), it follows that \(\lim _{n\rightarrow \infty }\Vert x_n-p\Vert ^2\) exists. Next we prove that
Indeed, by Claim 2 we have
In addition,
and
Combining (36), (37) and (38), we obtain
This implies that (35) holds. On the other hand, since the sequence \(\{x_{n}\}\) is bounded, it follows that there exists a subsequence \(\{x_{n_{k}}\}\) of \(\{x_{n}\}\), which converges weakly to \(q \in \mathscr {C}\) and
According to Claim 3, we get
That is
Using the fact that \( x_{n_k}\rightharpoonup q \ (k\rightarrow \infty )\), (40) and Lemma 12, we have \(q \in Sol(\mathscr {C},\mathscr {A})\). Since \(p=P_{Sol(\mathscr {C},\mathscr {A})}\circ f(p)\), combining (35) and (39), we obtain
Hence, by Claim 4, (41) and Lemma 8, we conclude that \(\lim _{n\rightarrow \infty }\Vert x_n-p\Vert =0\).
Case 2 There exists a subsequence \(\{\Vert x_{n_j}-p\Vert ^2\}\subset \{\Vert x_n-p\Vert ^2\}\) such that
From Lemma 7, there exists a non-decreasing sequence \(\{m_k\}\) of \(\mathbb {N}\) such that \(\lim _{k\rightarrow \infty }m_k=\infty \) and
From Claim 2,
As proved in Case 1, we can conclude that
and
Applying Claim 4, we have
It follows that
Hence, \(x_k\rightarrow p\) as \(k\rightarrow \infty \). This completes the proof. \(\square \)
3.2 Convergence rate
Next, we assume that the operator \(\mathscr {A}\) in Algorithm 3 is \(\eta \)-strongly pseudo-monotone. In this case, by modifying Algorithm 3, we obtain that the sequence \(\{x_n\}\) strongly converges with a linear rate. The modified algorithm is as follows:
Remark 2
Similar to the proof of Lemma 10, it is easy to get that (44) is well defined.
Theorem 2
Assume that \(\mathscr {A}:\mathscr {C}\rightarrow \mathscr {C}\) is \(\eta \)-strongly pseudo-monotone and L-Lipschitz continuous. Then the sequence \(\{x_n\}\) generated by Algorithm 4 converges strongly with a Q-linear rate to the unique element p in Sol\((\mathscr {C},\mathscr {A})\).
Proof
According to the conditions of Theorem 2, \(\langle \mathscr {A}p,x_{n+1}-p\rangle \ge 0\) and thus
From the definition of \(\{x_{n+1}\}\), we have
It follows from (44) that
Moreover,
which implies that
Since \(\mu \in (0,\min \{\frac{1}{\sqrt{2}},\frac{2l\eta }{L}\}) \), we have \(1+\mu ^2<2-\mu ^2\). Next we show that \(\tau _{n}>\frac{\mu l}{L}\). Indeed, we know that \(\frac{\tau _{n}}{l}\) must violate (44), thus,
that is \(\tau _{n}>\frac{\mu l }{L}\). It follows that \(2\tau _{n}\eta> 2\frac{\mu l \eta }{L}> \mu ^2\) and \(\frac{\mu ^2}{\min \{1-\mu ^2,2\tau _{n}\eta \}}<1\). Fix \(\delta \in (\frac{1+\mu ^2}{2\mu },\frac{2-\mu ^2}{2\mu })\) and \(\varepsilon \in (\frac{\mu ^2}{\min \{1-\mu ^2,2\tau _{n}\eta \}},\frac{2(1-\delta \mu )}{\min \{1-\mu ^2,2\tau _{n}\eta \}})\). Since \(\frac{\mu ^2}{\min \{1-\mu ^2,2\tau _{n}\eta \}}<1\), we can choose \(\varepsilon \in (0,1)\) such that
Let \(\nu :=\frac{1}{2}\varepsilon \min \{1-\mu ^2,2\tau _{n}\eta \}\), we see that
Therefore,
where \(\gamma :=2\varepsilon \tau _{n}\eta -\mu ^2>0\). It follows that
that is,
This implies that the sequence \(\{x_n\}\) generated by Algorithm 4 converges strongly to p with a Q-linear rate. \(\square \)
4 Numerical experiments
We give some numerical examples to show performances of our proposed Algorithm 3 and compare with [15, Algorithm SD], [14, Algorithm PY], [28, Algorithm DY] and [12, Algorithm DA].
Example 1
Assume that \(\mathscr {A}: \mathbb {R}^m \rightarrow \mathbb {R}^m\) is defined by \(\mathscr {A}(x) : = Mx+q\) with \(M = NN^T + S + D\), N is an \(m\times m\) matrix, S is an \(m\times m\) skew-symmetric matrix, D is an \(m\times m\) diagonal matrix, whose diagonal entries are positive (so \(\mathscr {A}\) is positive definite). All the entries of N, S and D are randomly generated in the interval (0, 20). Consider \(Sol(\mathscr {C},\mathscr {A})\) with the feasible set \(\mathscr {C}=\{x\in \mathbb {R}^m: -10\le x_i \le 10, i=1,2,\cdots , m\}\). The parameters are taken as follows.
-
Algorithm 3: \(\alpha _n = \left\{ \begin{array}{ll} \frac{1}{(n+1)^2}, &{}\text {if}\,\, \Vert x_{n} - x_1\Vert = 0 \\ \min \left\{ \frac{1}{(n+1)^2}, \frac{1}{(n+1)^2\times \Vert x_{n} - x_1\Vert }\right\} , &{}\text {otherwise} \\ \end{array} \right. \), \(\mu = 0.9\), \(l = 0.9\), \(\gamma = 0.05\), \(\beta _n = \frac{1}{n+1}\) and \(\rho = 0.5\);
-
Algorithm SD: \(\alpha _n = \frac{1}{n+1}\), \(\mu = 0.9\), \(l = 0.9\), \(\rho = 0.5\) and \(\lambda \) is randomly generated in the interval \((0, 1/\mu )\);
-
Algorithm PY: \(\alpha _n = \frac{1}{n+1}\), \(\mu = 0.9\) and \(l= 0.9\);
-
Algorithms DY, DA: \(\alpha _n = \frac{1}{n+1}\), \(\mu = 0.9\), \(l = 0.9\), \(\gamma = 0.05\).
In our experiment, the starting point \(x_1\) is generated randomly in \((-10, 10)^m\), where \(m = 5, 10\). We use the stopping rule \(E_n = \Vert x_n - P_\mathscr {C}(I-\mathscr {A})x_n\Vert \le 10^{-1}\) and we also stop if the number of iterations \(N = 60000\) for all algorithms. Figures 1, 2 and Tables 1, 2 show the computational results for Example 1 by using Algorithms 3, SD, PY, DY and DA.
Let “Iter” denote number of iterations, “CPU” denote the CPU time seconds. Example 1 shows that our proposed Algorithm 3 is fast, efficient and easy to implement. We notice that from Figs. 1, 2 and Tables 1, 2 that our Algorithm 3 outperforms SD, PY, DY and DA, in terms of CPU time and required number of iterations for each case of different dimensions given as follows: \(m=5, 10\).
Example 2
Consider \(\mathcal {H}=L^2([0,1])\) with inner product \(\langle x,y\rangle :=\int _{0}^1 x(t)y(t)dt\) and norm \(\Vert x\Vert _2 : = (\int _0^1 |x(t)|^2 dt) ^{\frac{1}{2}}\). Suppose \(\mathscr {C}: = \{x\in \mathcal {H}: \Vert x\Vert _2\le 2\}\). Let \(g:\mathscr {C}\rightarrow \mathbb {R}\) be defined by \(g(u) := \frac{1}{1+\Vert u\Vert ^2_2}.\) Observe that g is \(L_g\)-Lipschitz continuous with \(L_g = \frac{16}{25}\) and \(\frac{1}{5}\le g(u) \le 1, \forall u\in \mathscr {C}\). Define the Volterra integral mapping \(F: \mathcal {H}\rightarrow \mathcal {H}\) by
Then F is bounded linear monotone, see [29]. Now define \(\mathscr {A}: \mathscr {C}\rightarrow \mathcal {H}\) by
As given in [30], the mapping \(\mathscr {A}\) is pseudo-monotone but not monotone. Now take
where \(a\in \mathcal {H}\) and \(b\in R\). Then we define the metric projection \(P_\mathscr {C}\) as
Let \(Q = \{x\in \mathcal {H}: \Vert x\Vert _2\le r\}\) be a closed ball centered at 0 with radius \(r = 2\), then Q is a nonempty closed and convex subset of \(L^2([0, 1])\). Thus, the projection onto Q is easily computed as
During this experiment, the parameters are taken as follows.
-
Algorithm 3: \(\alpha _n = \left\{ \begin{array}{ll} \frac{1}{(n+1)^2}, &{}\text {if}\,\, \Vert x_{n} - x_1\Vert _2 = 0\\ \min \left\{ \frac{1}{(n+1)^2}, \frac{1}{(n+1)^2\times \Vert x_{n} - x_1\Vert _2 }\right\} , &{}\text {otherwise} \\ \end{array} \right. \), \(\mu = 0.4\), \(l = 0.1\), \(\gamma = 0.5\), \(\beta _n = \frac{1}{n+1}\) and \(\rho = 0.5\);
-
Algorithm SD: \(\alpha _n = \frac{1}{n+1}\), \(\mu = 0.4\), \(l = 0.1\), \(\rho = 0.5\) and \(\lambda \) is randomly generated in the interval \((0, 1/\mu )\);
-
Algorithm PY: \(\alpha _n = \frac{1}{n+1}\), \(\mu = 0.4\) and \(l = 0.1\);
-
Algorithms DY, DA: \(\alpha _n = \frac{1}{n+1}\), \(\mu = 0.4\), \(l = 0.1\), \(\gamma = 0.5\).
We test Algorithms 3, PY, DY for different cases of the initial point \(x_1 \in L^2([0,1])\). Let the initial point be \(x_1 = t\). We take \(E_n = \Vert x_{n+1} - x_n\Vert _2\le 4 \times 10^{-3}\) as a termination criterion. The results of this test are displayed in Fig. 3 and Table 3.
Let the initial point be \(x_1 = t^3\). We terminate the iterations if \(E_n = \Vert x_{n+1} - x_n\Vert _2\le 10^{-3}.\) The numerical results are presented in Fig. 4 and Table 3.
Let “Iter” denote number of iterations, “CPU” denote the CPU time seconds. Figures 3, 4 and Table 3 show that the numerical results of Example 2 with initial points \(x_0 = t\) and \(x_1= t^3\), respectively. They show that the performance of Algorithm 3 is better than that of Algorithms PY, DY, in terms of the number of iterations and CPU time required to reach the stopping criterion.
Example 3
Let us consider the variational inequality problem. Let
and \(\mathscr {C}:=\{x \in \mathbb {R}^2:\ -10\le x_i\le 10,\ i=1,2\}.\) This problem has unique solution \(x^*=(0,-1)^T\). It is easy to see that \(\mathscr {A}\) is not a monotone map on \(\mathscr {C}\). However, using the Monte Carlo approach (see [31]), it can be shown that \(\mathscr {A}\) is pseudo-monotone on \(\mathscr {C}\). Let \(f(x)=20/x\). Parameters in different algorithms are selected as follows:
Algorithm 3: \(\beta _n=\frac{1}{50n+1},\,\gamma =1.35,\,l=0.22,\,\mu =0.99\) and \(\alpha =0.8\);
Algorithm SD: \(\alpha _n=\frac{1}{50n+1},\,l=0.22,\,\mu =0.99\) and \(\lambda =0.9/\mu \);
Algorithm PY: \(\alpha _n=\frac{1}{50n+1},\, l=0.22\) and \(\mu =0.99\).
We take some initial points and different stopping criteria conditions in the Table 4. The following statistical data are obtained by averaging the number of iterations and CPU costs from 10 independent trials. In the Fig. 5, the initial point is \([10,10]^T\) and the number of steps to stop criteria is 1000.
5 Conclusions
In this paper, we have given an improved projection-type method for solving classical variational inequalities in Hilbert spaces. We have proposed a novel line-search rule that removes the reliance on Lipschitz continuity. Furthermore, a strong convergence theorem is obtained through the combination of viscosity iteration and the projection method. In numerical experiments, we have compared our Algorithm 3 with some recent related results, and it can be found from the figures and tables that our new scheme has better convergence performance.
Availability of supporting data
Data sharing not applicable to this article as no datasets were generated or analyzed during the current study.
References
Korpelevich, G.M.: An extragradient method for finding saddle points and for other problems. Ekonom. i Mat. Metody 12, 747–756 (1976)
Yao, Y., Postolache, M., Yao, J.C.: Strong convergence of an extragradient algorithm for variational inequality and fixed point problems. Politehn. Univ. Bucharest Sci. Bull. Ser. A Appl. Math. Phys. 82, 3–12 (2020)
Yao, Y., Postolache, M., Yao, J.C.: Convergence of an extragradient algorithm for fixed point and variational inequality problems. J. Nonlinear Convex Anal. 20, 2623–2631 (2019)
Vuong, P.T.: On the weak convergence of the extragradient method for solving pseudo-monotone variational inequalities. J. Optim. Theory Appl. 176, 399–409 (2018)
Thong, D.V.: Extragradient method with a new adaptive step size for solving non-Lipschitzian pseudo-monotone variational inequalities. Carpathian J. Math. 38, 503–516 (2022)
Tan, B., Liu, L., Qin, X.: Self adaptive inertial extragradient algorithms for solving bilevel pseudomonotone variational inequality problems. Jpn. J. Ind. Appl. Math. 38, 519–543 (2021)
Tan, B., Li, S., Qin, X.: An accelerated extragradient algorithm for bilevel pseudomonotone variational inequality problems with application to optimal control problems. Rev. R. Acad. Cienc. Exactas Fís. Nat. Ser. A Mat. RACSAM 115(4), 19 (2021)
Duvocelle, B., Meier, D., Staudigl, M., Vuong, P.T.: Strong Convergence of Forward-Backward-Forward Methods for Pseudo-monotone Variational Inequalities with Applications to Dynamic User Equilibrium in Traffic Networks (2019)
Iusem, A.N.: An iterative algorithm for the variational inequality problem. Mat. Apl. Comput. 13, 103–114 (1994)
Iusem, A.N., Jofré, A., Oliveira, R.I., Thompson, P.: Variance-based extragradient methods with line search for stochastic variational inequalities. SIAM J. Optim. 29, 175–206 (2019)
Thong, D.V., Hieu, D.V.: Inertial subgradient extragradient algorithms with line-search process for solving variational inequality problems and fixed point problems. Numer. Algorithms 80, 1283–1307 (2019)
Thong, D.V., Gibali, A.: Extragradient methods for solving non-Lipschitzian pseudo-monotone variational inequalities. J. Fixed Point Theory Appl. 21(20), 19 (2019)
Xie, Z., Cai, G., Li, X., Dong, Q.L.: Strong convergence of the modified inertial extragradient method with line-search process for solving variational inequality problems in Hilbert spaces. J. Sci. Comput. 88(3), 19 (2021)
Vuong, P.T., Shehu, Y.: Convergence of an extragradient-type method for variational inequality with applications to optimal control problems. Numer. Algorithms 81, 269–291 (2019)
Reich, S., Thong, D.V., Dong, Q.L., Li, X.H., Dung, V.T.: New algorithms and convergence theorems for solving variational inequalities with non-Lipschitz mappings. Numer. Algorithms 87, 527–549 (2021)
Censor, Y., Gibali, A., Reich, S.: The subgradient extragradient method for solving variational inequalities in Hilbert space. J. Optim. Theory Appl. 148, 318–335 (2011)
Censor, Y., Gibali, A., Reich, S.: Extensions of Korpelevich’s extragradient method for the variational inequality problem in Euclidean space. Optimization 61, 1119–1132 (2011)
Tseng, P.: A modified forward-backward splitting method for maximal monotone mappings. SIAM J. Control. Optim. 38, 431–446 (2000)
Reich, S., Thong, D.V., Cholamjiak, P., Van Long, L.: Inertial projection-type methods for solving pseudomonotone variational inequality problems in Hilbert space. Numer. Algorithms 88, 813–835 (2021)
Yao, Y., Iyiola, O.S., Shehu, Y.: Subgradient extragradient method with double inertial steps for variational inequalities. J. Sci. Comput. 90(2), 29 (2022)
Goebel, K., Reich, S.: Uniform Convexity, Hyperbolic Geometry, and Nonexpansive Mappings. Marcel Dekker, New York (1984)
Denisov, S.V., Semenov, V.V., Chabak, L.M.: Convergence of the modified extragradient method for variational inequalities with non-Lipschitz operators. Cybern. Syst. Anal. 51, 757–765 (2015)
Iusem, A.N., Gárciga, O.R.: Inexact versions of proximal point and augmented Lagrangian algorithms in Banach spaces. Numer. Funct. Anal. Optim. 22, 609–640 (2001)
Cottle, R.W., Yao, J.C.: Pseudo-monotone complementarity problems in Hilbert space. J. Optim. Theory Appl. 75, 281–295 (1992)
He, Y.R.: A new double projection algorithm for variational inequalities. J. Comput. Appl. Math. 185, 166–173 (2006)
Maingé, P.E.: A hybrid extragradient-viscosity method for monotone operators and fixed point problems. SIAM J. Control. Optim. 47, 1499–1515 (2008)
Xu, H.K.: Iterative algorithms for nonlinear operators. J. Lond. Math. Soc. 66, 240–256 (2002)
Thong, D.V., Shehu, Y., Iyiola, O.S.: Weak and strong convergence theorems for solving pseudo-monotone variational inequalities with non-Lipschitz mappings. Numer. Algorithms 84, 795–823 (2020)
Bauschke, H.H., Combettes, P.L.: Convex analysis and monotone operator theory in Hilbert spaces. Springer, New York (2011)
Shehu, Y., Dong, Q.L., Jiang, D.: Single projection method for pseudo-monotone variational inequality in Hilbert spaces. Optimization 68, 385–409 (2019)
Hu, X., Wang, J.: Solving pseudo-monotone variational inequalities and pseudo-convex optimization problems using the projection neural network. IEEE Trans. Neural Netw. 17, 1487–1499 (2006)
Acknowledgements
The authors wish to thank the anonymous referees for their valuable comments and suggestions which lead to an improvement of this paper.
Funding
This work was supported by National Natural Science Foundation of China (Grant No. 12171373).
Author information
Authors and Affiliations
Contributions
Xie and Wu wrote the main manuscript text and Liu finished the numerical examples. All authors reviewed the manuscript.
Corresponding author
Ethics declarations
Ethical Approval and Consent to participate
Not Applicable.
Consent for publication
Not Applicable.
Human and Animal Ethics
Not Applicable.
Competing interests
The authors declare no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Xie, Z., Wu, H. & Liu, L. Modified projection method and strong convergence theorem for solving variational inequality problems with non-Lipschitz operators. Numer Algor (2024). https://doi.org/10.1007/s11075-024-01851-7
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11075-024-01851-7