Abstract
Several efficient methods have been developed in the literature for approximating solutions of fixed point and optimization problems. However, the S-iteration process has been shown to outperform many of these existing methods. In this paper, we study the problem of finding the common solution of split variational inclusion problem, equilibrium problem and common fixed point of nonexpansive mappings. We introduce an improved S-iteration method, which combines inertial and viscosity techniques with self-adaptive step size for approximating the solution of the problem in the framework of Hilbert spaces. Moreover, under some mild conditions we prove strong convergence theorem for the proposed algorithm without the knowledge of the operator norm and we apply our result to study split minimization problem, split feasibility problem and relaxed split feasibility problem. Finally, we present some numerical experiments with graphical illustrations to demonstrate the implementability and efficiency of our proposed method in comparison with some existing state of the art methods in the literature.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Throughout this paper, let \({\mathbb {R}}\) denote the set of all real numbers and \({\mathbb {N}}\) denote the set of all positive numbers. Let H be a real Hilbert space with inner product \(\langle \cdot , \cdot \rangle \) and induced norm \(||\cdot ||,\) and let C be a nonempty closed convex subset of H. Let \(S: C\rightarrow C\) be a nonlinear mapping. A point \({\hat{x}}\in C\) is called a fixed point of S if \(S{\hat{x}} = {\hat{x}}.\) We denote by F(S), the set of all fixed points of S, i.e.
A mapping \(S: C\rightarrow C\) is called a nonexpansive mapping if
The study of fixed point theory for nonexpansive mappings has flourished in recent years due to its vast applications in fields like economics, compressed sensing, and other applied sciences. In particular, certain problems such as variational inequalities problems, convex optimization problems, convex feasibility problems, monotone inclusion problems and image restoration problems can be formulated as finding the fixed points of nonexpansive mapping (see Bauschke and Borwein 1996; Chen et al. 2013). Several researchers have put considerable efforts in the study and formulation of iterative methods to approximate the fixed points of nonexpansive mappings (for example, see Halpern 1967; Moudafi 2000 and the reference therein).
In 2007, Agarwal et al. (2007) introduced the following iterative method known as the S-iteration. Let C be a convex subset of a linear space X and S a mapping of C into itself. The sequence \(\{x_n\}\) is generated as follows:
where \(\{\alpha _n\}\) and \(\{\beta _n\}\) are sequences in (0, 1). The authors showed that the S-iteration process has a better rate of convergence than Mann and Ishikawa iteration processes.
Some authors have used the S-iteration process and its modifications to find common fixed points of two mappings; see for example (Bussaban and Kettapun 2018; Pandey et al. 2019) and the references therein.
Let \( F:C\times C\rightarrow {\mathbb {R}}\) be a bifunction. The equilibrium problem (EP) for the bifunction F on C is formulated as finding a point \({\hat{x}}\in C\) such that
The solution set of EP (1.2) is denoted by EP(F). The EP covers a wide range of topics that have emerged from the social sciences, economics, finance, image restoration, ecology, transport, networking, elasticity and optimization problems (see Olona et al. 2021a; Patriksson 2015). The problem is a generalized concept that unifies several mathematical problems as special cases, namely minimization problems, variational inequality problems, mathematical programming problems, complementarity problems, saddle point problems, Nash equilibrium problems in noncooperative games, minimax inequality problems, fixed point problems, scalar and vector minimization problems, and others; see (Alakoya et al. 2021; Blum 1994) and the references therein. Recently, the EP and its various generalizations have attracted considerable research efforts and various iterative methods have been proposed for approximating their solutions (see Alakoya et al. 2021; Jolaoso et al. 2020; Ogwo et al. 2021; Olona et al. 2021b; Oyewole et al. 2021; Taiwo et al. 2021a, b and the references therein).
In the recent time, the Split Inverse Problem (SIP) has attracted the attention of several authors (see Alakoya et al. 2021; Oyewole et al. 2021; Taiwo et al. 2021 and the references therein) due to its wide areas of applications, for example, in phase retrieval, image recovery, signal processing, data compression, intensity-modulated radiation therapy, among others (see Censor et al. 2006; Censor and Elfving 1994 and the references therein). The SIP model is formulated as follows: Find a point
such that
where \(H_1\) and \(H_2\) are real Hilbert spaces, \(\hbox {IP}_1\) denotes an inverse problem formulated in \(H_1\) and \(\hbox {IP}_2\) denotes an inverse problem formulated in \(H_2,\) and \(A : H_1 \rightarrow H_2\) is a bounded linear operator.
Censor and Elfving (1994) introduced the first instance of the SIP called the split feasibility problem (SFP) in 1994 for modelling inverse problems which arise from medical image reconstruction. Since then, the SFP has been studied intensively by several authors due to its wide areas of application such as in signal processing, control theory, approximation theory, geophysics, biomedical engineering, communications, etc (Byrne 2002; Censor et al. 2006; Godwin et al. 2020). Let C and Q be nonempty closed convex subsets of Hilbert spaces \(H_1\) and \(H_2,\) respectively, and \(A:H_1\rightarrow H_2\) be a bounded linear operator. The SFP is formulated as follows:
Moudafi (2011) introduced another instance of the SIP known as the split monotone variational inclusion problem (SMVIP). Let \(H_1, H_2\) be real Hilbert spaces, \(f_1:H_1\rightarrow H_1, f_2:H_2\rightarrow H_2,\) are inverse strongly monotone mappings, \(A:H_1\rightarrow H_2\) is a bounded linear operator, \(B_1:H_1\rightarrow 2^{H_1}, B_2:H_2\rightarrow 2^{H_2}\) are multivalued maximal monotone mappings. The SMVIP is formulated as follows:
and
If \(f_1\equiv 0\equiv f_2,\) then the SMVIP (1.6)-(1.7) reduces to the following split variational inclusion problem (SVIP):
and
The SVIP (1.8)-(1.9) constitutes a pair of variational inclusion problems which have to be solved so that the image \({\hat{y}} = A{\hat{x}}\) under a given bounded linear operator A of the solution of the SVIP (1.8) in \(H_1\) is the solution of the other SVIP (1.9) in another Hilbert space \(H_2.\) Moudafi (2011), showed that the SVIP (1.8)-(1.9) includes as a special case the SFP (1.11). The SVIP is at the core of modelling many inverse problems arising from phase retrieval and other real world problems, for instance, in sensor networks in computerized and data compression (Byrne 2002; Combettes 1996). We denote the solution set of SVIP (1.8) by SOLVIP(\(B_1\)) while the solution set of SVIP (1.9) is denoted by SOLVIP(\(B_2\)). Hence, the solution set of the SVIP (1.8)–(1.9) is denoted by
To solve the SVIP for two maximal monotone operators \(B_1\) and \(B_2\) in Hilbert spaces, Byrne et al. (2012) proposed the following algorithm:
for \(\lambda >0\) and \(A^*\) is the adjoint operator of the bounded linear operator \(A, \gamma \in (0, \frac{2}{L}), L = ||A^*A||, J_\lambda ^{B_1}:= (I +\lambda B_1)^{-1}, J_\lambda ^{B_2}:= (I +\lambda B_2)^{-1}\) are the resolvent operators of \(B_1\) and \(B_2\) respectively, and \(\{\alpha _n\}\) is a sequence in [0, 1] satisfying \(\lim _{n\rightarrow \infty }\alpha _n=0\) and \(\sum _{n=1}^{\infty }\alpha _n=\infty .\) The authors obtained a strong convergence result for the proposed algorithm under some mild conditions.
In the recent time, the problems of finding common solutions of the set of fixed points of nonlinear mappings and the set of solutions of optimization problems have been considered by some authors (for instance, see Alakoya et al. 2021; Cholamjiak and Suantai 2013; Khan et al. 2020 and the references therein). The importance and motivation for studying such a common solution problem lies in its potential application to mathematical models whose constraints can be expressed as fixed point problems and optimization problems. In this article we will be studying the problems of finding common solutions of the set of fixed points of nonlinear mappings and the set of solutions of certain optimization problems.
Recently, Wangkeeree et al. (2018) introduced the following general iterative scheme for approximating a common solution of SVIP and FPP for a nonexpansive mapping in the setting of real Hilbert spaces.
where \(f: H_1\rightarrow H_1\) is a contraction with constant \(k\in (0,1),~ S: H_1\rightarrow H_1\) is a nonexpansive mapping, \(D: H_1\rightarrow H_1\) is a strongly positive bounded linear operator with constant \(\mu \) and \(0<\beta <\frac{\mu }{k}, \lambda >0, \gamma \in (0, \frac{1}{L}),\) where L is the spectral radius of the operator \(A^*A,\) \(\{\alpha _n\}\subset (0,1)\) and \(B_1: H_1\rightarrow 2^{H_1}, B_2: H_2\rightarrow 2^{H_2}\) are two multi-valued maximal monotone operators on \(H_1\) and \(H_2\) respectively. Under certain conditions, the sequence generated by the proposed Algorithm 1.3 was proved to converge strongly to a common solution of split variational inclusion problem and fixed point problem for a nonexpansive mapping.
Remark 1.4
Here, we remark that the step size \(\gamma \) of Algorithms 1.2 and 1.3 above plays an essential role in the convergence properties of the algorithms. Many of the existing iterative methods for solving SVIP involve step size that depends on the norm of the bounded linear operator A. Such algorithms are usually not easy to implement because they require computation of the operator norm which oftentimes is difficult to compute. In addition, the step size defined by these methods are often very small and deteriorates the convergence rate of the algorithm. A larger step size can often be used in practice to yield better numerical results.
Very recently, Tang (2020) proposed the following Halpern-type algorithm with self-adaptive step size for solving SVIP in the framework of Hilbert spaces.
where \(g(x) = \frac{1}{2}||(I - J_\lambda ^{B_2})Ax||^2, G(x) = A^*(I - J_\lambda ^{B_2})Ax, H(x) = (I - J_\lambda ^{B_1})x,\) and \(\{\alpha _n\}\) is a sequence in (0, 1) satisfying \(\lim _{n\rightarrow \infty }\alpha _n=0\) and \(\sum _{n=1}^{\infty }\alpha _n=\infty .\) Under mild conditions, strong convergence result was obtained for the proposed algorithm.
Based on the heavy ball methods of a two-order time dynamical system, Polyak (1964) first introduced an inertial extrapolation as an acceleration process to solve the smooth convex minimization problem. The inertial algorithm is a two-step iteration where the next iterate is defined by making use of the previous two iterates. In the recent time, there has been an increasing interest in the study of inertial type algorithms and several authors have constructed some fast iterative methods by employing the inertial technique (see, e.g., Alakoya et al. 2021, 2020, 2021; Izuchukwu et al. 2020; Ogwo et al. 2021; Owolabi et al. 2021).
For approximating the zero point of a maximal monotone operator B, Alvarez and Attouch (2001) introduced the following inertial proximal algorithm:
The authors obtained a weak convergence result for the algorithm under the following conditions:
-
(B1)
There exists \(\lambda >0\) such that for all \(n\in {\mathbb {N}}, ~ \lambda _n\ge \lambda .\)
-
(B2)
There exists \(\alpha \in [0,1)\) such that for all \(n\in {\mathbb {N}},~ 0\le \alpha _n\le \alpha .\)
-
(B3)
\(\sum _{n=1}^\infty \alpha _n\Vert x_n - x_{n-1}\Vert ^2< \infty .\)
Recently, authors have pointed out one of the drawbacks of the summability condition (B3) of the Algorithm 1.6, that is, to satisfy the summability condition, it is necessary to first calculate \(\alpha _n\) at each step (see Moudafi and Oliny 2003).
Very recently, Shehu et al. (2021), Iyiola et al. (2018), Shehu et al. (2020), and Shehu and Iyiola (2020) proposed some efficient inertial iterative methods with self-adaptive step size for approximating solutions of certain classes of optimization problems and the authors were able to establish convergence of the proposed methods under some mild conditions imposed on the control parameters.
From the above review, the following natural question arises:
Question: Can we construct an inertial iterative method with self-adaptive step size for approximating a common solution of split variational inclusion problem, equilibrium problem and fixed point problem in Hilbert spaces such that condition (B3) of Algorithm 1.6 is dispensed with?
Our interest in this paper is to provide an affirmative answer to the above question.
Inspired by the above results and the ongoing research interest in this direction, in this paper, we introduce a new inertial iterative scheme which employs the viscosity S-iteration technique with self-adaptive step size for approximating a common element of the set of solutions of split variational inclusion problem, equilibrium problem and common fixed point problem for nonexpansive mappings in Hilbert spaces. Our motivation for studying such a common solution problem lies in its potential application to mathematical models whose constraints can be expressed as split variational inclusion problem, equilibrium problem and common fixed point problem. This arises in practical problems such as signal processing, network resource allocation, image recovery. A scenario is in network bandwidth allocation problem for two services in a heterogeneous wireless access networks in which the bandwidth of the services are mathematically related (see, for instance, Iiduka 2012; Luo et al. 2009 and the references therein). Unlike in Algorithms 1.2 and 1.3 and several other algorithms in the literature, our algorithm is designed such that its implementation does not require the knowledge of the norm of the bounded linear operator. Moreover, our work extend the results in Agarwal et al. (2007), Wangkeeree et al. (2018), Tang (2020), Alvarez and Attouch (2001) to the problem of finding a common solution of split variational inclusion problem, equilibrium problem and common fixed point problem for nonexpansive mappings and the inertial technique employed is more efficient than that used in Alvarez and Attouch (2001). Under some mild conditions, we prove strong convergence theorem for the proposed algorithm. Furthermore, we apply our result to study other optimization problems and we provide some numerical experiments with graphical illustrations to demonstrate the efficiency of the proposed algorithm in comparison with some existing state of the art algorithms in the literature.
The outline of the paper is as follows: In Sect. 2, we recall some basic definitions and existing results which are needed for the convergence analysis of the proposed algorithm. In Sect. 3, we present the proposed algorithm and highlight some of its important features while in Sect. 4 we discuss its convergence. In Sect. 5 we apply our results to study split minimization problem and split feasibility problems. In Sect. 6, numerical examples and comparison with some related algorithms are presented to demonstrate the performance of our new algorithm. Finally, we give the concluding remarks in Sect. 7.
2 Preliminaries
In this section, we recall some concepts and results which will be employed in the sequel. Let H be a real Hilbert space, for a nonempty closed and convex subset C of H, the metric projection \(P_C: H\rightarrow C\) is defined, for each \(x\in H,\) as the unique element \(P_Cx\in C\) such that
It is known that \(P_C\) is firmly nonexpansive, i.e.,
for all \(x,y\in H.\) Moreover, for any \(x\in H\) and \(z\in C, z=P_Cx\) if and only if (see [?])
In what follows, we denote the weak and strong convergence of a sequence \(\{x_n\}\) to a point \(x \in H\) by \(x_n \rightharpoonup x\) and \(x_n \rightarrow x\), respectively and \(w_\omega (x_n)\) denotes set of weak limits of \(\{x_n\},\) that is,
Definition 2.1
Let H be a real Hilbert space, \(C\subset H\) be a subset of H and \(h:C\rightarrow H\) be an operator from C onto H. The operator h is said to be
-
(1)
firmly nonexpansive if
$$\begin{aligned} \langle h(x)-h(y),x-y \rangle \ge ||h(x)-h(y)||^2,\quad \forall x,y\in C; \end{aligned}$$ -
(2)
L-Lipschitz continuous, where \(L>0,\) if
$$\begin{aligned} ||hx - hy|| \le L||x-y||,\quad \forall ~~x,y\in C; \end{aligned}$$if \(L\in [0,1),\) then T is called a contraction mapping;
-
(3)
nonexpansive if T is \(1-\)Lipschitz continuous;
-
(4)
hemicontinuous if it is continuous along each line segment in C.
Lemma 2.2
(Zhao et al. 2018) (Demiclosedness Principle). Let T be a nonexpansive mapping on a closed convex subset C of a real Hilbert space H. Then \(I-T\) is demiclosed at any point \(y\in H,\) that is, if \(x_n\rightharpoonup x\) and \(x_n-Tx_n\rightarrow y\in H,\) then \(x-Tx=y.\)
Definition 2.3
Let H be a real Hilbert space. A function \(f:H\rightarrow {\mathbb {R}}\cup \{+\infty \}\) is said to be weakly lower semicontinuous (w-lsc) at \(x\in H,\) if
holds for an arbitrary sequence \(\{x_n\}_{n=0}^\infty \) in H satisfying \(x_n\rightharpoonup x.\)
Definition 2.4
Let H be a real Hilbert space and \(\lambda >0.\) The operator \(B: H\rightarrow 2^{H}\) is said to be
-
monotone if
$$\begin{aligned} \langle u -v, x-y \rangle \ge 0~ \text {for all}~ u\in B(x), v\in B(y). \end{aligned}$$ -
maximal monotone mapping if the graph G(B) of B,
$$\begin{aligned} G(B) := \{(x,u)\in H \times H| u\in B(x)\}, \end{aligned}$$is not properly contained in the graph of any other monotone mapping.
-
The resolvent of B with parameter \(\lambda > 0\) denoted by \(J_\lambda ^B\) is given by
$$\begin{aligned} J_\lambda ^B := (I + \lambda B)^{-1}, \end{aligned}$$where I is the identity operator.
Remark 2.5
For \(\lambda > 0,\) the following results hold (Tang 2020):
-
(1)
B is maximal monotone if and only if \(J_\lambda ^B\) is single-valued, firmly nonexpansive and \(dom(J_\lambda ^B) = H,\) where \(dom(B): = \{x\in H| B(x) \ne \emptyset \}.\)
-
(2)
The point \(x^*\in B^{-1}(0)\) if and only if \(x^* = J_\lambda ^Bx^*.\)
-
(3)
The solution set \({\mathcal {F}}\) of the SVIP (1.8)-(1.9) is equivalent to the following:
$$\begin{aligned} \text {Find}~~ x^*\in H_1~~ \text {with}~~ x^* = J_\lambda ^{B_1}x^*~~ \text {such that}~~ y^* = Ax^*\in H_2 ~\text {and}~ y^* = J_\lambda ^{B_2}y^* \end{aligned}$$(2.3)
Assumption 2.6
For solving the EP, we assume that the bifunction \(F: C \times C \rightarrow {\mathbb {R}}\) satisfies the following conditions:
-
(A1)
\(F(x,x) = 0\) for all \(x \in C;\)
-
(A2)
F is monotone, that is, \(F(x,y) + F(y,x) \le 0\) for all \(x,y \in C;\)
-
(A3)
F is upper hemicontinuous, that is, for all \( x,y,z \in C\), \(\lim _{t \downarrow 0} F\big (tz + (1-t)x,y \big )\le F(x,y);\)
-
(A4)
for each \(x \in C, y \mapsto F(x,y)\) is convex and lower semicontinuous.
Lemma 2.7
Ma et al. (2015) Let C be a nonempty closed convex subset of a Hilbert space H and \(F:C \times C \rightarrow {\mathbb {R}}\) be a bifunction satisfying Assumption 2.6. For \(r > 0\) and \(x \in H,\) define a mapping \( T^F_r : H \rightarrow C\) as follows:
Then \(T^F_r\) is well defined and the following hold:
-
(1)
for each \(x \in H, T^F_r(x) \ne \emptyset ;\)
-
(2)
\(T^F_r\) is single-valued;
-
(3)
\(T^F_r\) is firmly nonexpansive, that is, for any \(x,y \in H,\)
$$\begin{aligned} \Vert T^F_r x - T^F_r y\Vert ^2 \le \langle T^F_r x - T^F_r y, x - y \rangle ; \end{aligned}$$ -
(4)
\(F(T^F_r) = EP(F);\)
-
(5)
EP(F) is closed and convex.
Lemma 2.8
Chuang (2013), Ogwo et al. (2021) Let H be a real Hilbert space. Then the following results hold for all \(x,y\in H\) and \(\delta \in {\mathbb {R}}:\)
-
(i)
\(||x + y||^2 \le ||x||^2 + 2\langle y, x + y \rangle ;\)
-
(ii)
\(||x + y||^2 = ||x||^2 + 2\langle x, y \rangle + ||y||^2;\)
-
(iii)
\(||x - y||^2 = ||x||^2 - 2\langle x, y \rangle + ||y||^2.\)
-
(iv)
\(||\delta x + (1-\delta ) y||^2 = \delta ||x||^2 + (1-\delta )||y||^2 -\delta (1-\delta )||x-y||^2.\)
Lemma 2.9
Maingé (2007) Let \(\{a_n\}, \{c_n\}\subset \mathbb {R_+}, \{\sigma _n\}\subset (0,1)\) and \(\{b_n\}\subset {\mathbb {R}}\) be sequences such that
Assume \(\sum _{n=0}^{\infty }|c_n|<\infty .\) Then the following results hold:
-
(1)
If \(b_n\le \beta \sigma _n\) for some \(\beta \ge 0,\) then \(\{a_n\}\) is a bounded sequence.
-
(2)
If we have
$$\begin{aligned} \sum _{n=0}^\infty \sigma _n = \infty ~~ \text {and}~~ \limsup _{n\rightarrow \infty }\frac{b_n}{\sigma _n}\le 0, \end{aligned}$$then \(\lim _{n\rightarrow \infty }a_n =0.\)
Lemma 2.10
Saejung and Yotkaew (2012) Let \(\{a_n\}\) be a sequence of non-negative real numbers, \(\{\alpha _n\}\) be a sequence in (0, 1) with \(\sum _{n=1}^\infty \alpha _n = \infty \) and \(\{b_n\}\) be a sequence of real numbers. Assume that
if \(\limsup _{k\rightarrow \infty }b_{n_k}\le 0\) for every subsequence \(\{a_{n_k}\}\) of \(\{a_n\}\) satisfying \(\liminf _{k\rightarrow \infty }(a_{n_{k+1}} - a_{n_k})\ge 0,\) then \(\lim _{n\rightarrow \infty }a_n =0.\)
3 Proposed method
In this section, we present the proposed algorithm. First, we define the following functions:
and
From Aubin (1993), it can easily be verified that g and h are weak lower semi-continuous and convex differentiable. Moreover, G and H are Lipschitz continuous (see Tang 2020).
In what follows, we assume that C and Q are nonempty closed convex subsets of real Hilbert spaces \(H_1\) and \(H_2,\) respectively, \(B_1 : H_1 \rightarrow 2^{H_1}\) and \(B_2 : H_2 \rightarrow 2^{H_2}\) are maximal monotone mappings, \(F:C\times C\rightarrow {\mathbb {R}}\) is a bifunction satisfying Assumption 2.6, and \(A : H_1 \rightarrow H_2\) is a bounded linear operator with \(A^*\) being its adjoint operator (\(A^*=A^T\) in finite dimensional spaces). Let \(S,T:H_1\rightarrow H_1\) be nonexpansive mappings and \(f:H_1\rightarrow H_1\) be a contraction with coefficient \(k\in (0,1).\) We denote the solution set by \(\Omega = F(S)\cap F(T)\cap {\mathcal {F}} \cap EP(F) \ne \emptyset \). We establish the convergence of the algorithm under the following conditions on the control parameters:
-
(C1)
Let \( \{\alpha _n\} \subset (0,1)\) such that \(\lim _{n\rightarrow \infty }\alpha _n=0\) and \(\sum _{n=0}^\infty \alpha _n = \infty ;\)
-
(C2)
Let \(\{\beta _n\}, \{\sigma _n\}, \{\delta _n\}, \{\xi _n\}\subset [a,b], a,b\in (0,1)\) and such that \(\alpha _n+\delta _n+\xi _n=1;\)
-
(C3)
Let \(\theta >0, \{\mu _n\}\) be a positive sequence such that \(\lim _{n\rightarrow \infty }\frac{\mu _n}{\alpha _n}=0;\)
-
(C4)
\(0< a \le \rho _n \le b < 4, \{r_n\}\subset (0,\infty )\) such that \(\liminf _{n\rightarrow \infty }r_n>0,\) and \(\lambda _i>0,~~ i=1,2.\)
Now, our main algorithm is presented as follows:
Remark 3.2
By conditions (C1) and (C3), one can easily verify from (3.1) that
Remark 3.3
Observe that the step size of our proposed Algorithm (3.1) is constructed such that its implementation does not require knowledge of the operator norm. Moreover, the implementation of the inertial term does not require the very stringent summability condition (B3) of Algorithm 1.6 (Alvarez and Attouch 2001). These important features make our proposed method easily implementable.
Remark 3.4
We remark that since the split variational inclusion problem generalizes the split feasibility problem and variational inequality problem, our proposed method could be viewed as an extension of the methods proposed in Shehu and Ogbuisi (2015), Cai et al. (2018), Gibali and Shehu (2019).
Remark 3.5
Observe that according to Lemma 2.7, the operator defined in (2.4) is single-valued. To compute \(u_n\) in (3.2) is equivalent to evaluating \(T^F_{r_n} (w_n).\) In doing this, we find an element \(u_n\in C\) such that inequality (3.2) holds. By applying the given definition of \(F(u_n,y)\) and simplifying the resulting inequality, we will obtain a quadratic function in variable y. Since the operator \(T^F_{r_n}\) is single-valued, then the quadratic function will have at most one solution in \({\mathbb {R}}.\) Hence, the value of \(u_n\) for which the discriminant of the quadratic function is zero is determined and this gives the value of the operator \(T^F_{r_n}\) at \(w_n.\)
4 Convergence analysis
Next, we state the strong convergence theorem for the proposed algorithm as follows.
Theorem 4.1
Let \(H_1\) and \(H_2\) be real Hilbert spaces, and \(A:H_1\rightarrow H_2\) be a bounded linear operator with adjoint \(A^*.\) Suppose \(S,T:H_1\rightarrow H_1\) are nonexpansive mappings, and \(f:H_1\rightarrow H_1\) is a contraction with coefficient \(k\in (0,1).\) Let \(\{x_n\}\) be the sequence generated by Algorithm 3.1 such that conditions (A1)-(A4) and (C1)-(C4) are satisfied. Then \(\{x_n\}\) converges strongly to a point \({\hat{x}}\in \Omega ,\) where \({\hat{x}}=P_\Omega \circ f({\hat{x}}).\)
We divide the proof of the strong convergence theorem into the following lemmas.
Lemma 4.2
Suppose that \(\{x_n\}\) is the sequence generated by Algorithm 3.1. Then \(\{x_n\}\) is bounded.
Proof
Observe that by (3.2), \(u_n=T_{r_n}^Fw_n.\) Also, we note that the mapping \(P_\Omega \circ f\) is a contraction. Then by the Banach Contraction Principle, there exists \(p\in H_1\) such that \(p=P_\Omega \circ f(p)\) and in particular \(p\in \Omega .\) Thus, it follows that \(Sp=p=Tp, T_{r_n}p=p, J_{\lambda _1}^{B_1}p=p\) and \(J_{\lambda _2}^{B_2}(Ap)=Ap.\) Since \(T_{r_n}^F\) is nonexpansive, then we have
Applying (4.1), we get
Next, by the definition of G(x) and the firmly nonexpansivity of \(I-J_{\lambda _2}^{B_2},\) we have
Then by Lemma 2.8(iii) and applying (4.3) together with the nonexpansivity of \(J_{\lambda _1}^{B_1},\) it follows that
By the condition on \(\rho _n,\) we have that
Applying (4.5), we obtain
Next, applying the triangle inequality, we get
By Remark 3.2, \(\lim _{n\rightarrow \infty }\frac{\theta _n}{\alpha _n}||x_n - x_{n-1}|| = 0.\) Then, it follows that there exists a constant \(M_1 > 0\) such that \(\frac{\theta _n}{\alpha _n}||x_n - x_{n-1}||\le M_1\) for all \(n\ge 1.\) Hence, from (4.7) we obtain
Applying (4.2), (4.5), (4.6) and (4.8), we have
where \(M^* :=\sup _{n\in {\mathbb {N}}}\Big \{\frac{||f(p)-p||}{1-k} + \frac{(1-\alpha _n)M_1}{1-k}\}.\) Setting \(a_n:=||x_n-p||;~ b_n:=\alpha _n(1-k)M^*;~ c_n:=0,\) and \(\sigma _n:=\alpha _n(1-k).\) By invoking Lemma 2.9 together with the assumptions on the control parameters, we have that \(\{||x_n-p||\}\) is bounded and this implies that \(\{x_n\}\) is bounded. Consequently, \(\{w_n\}, \{u_n\}, \{v_n\}, \{t_n\}\) and \(\{z_n\}\) are all bounded. \(\square \)
Lemma 4.3
Let \(\{x_n\}\) be the sequence generated by Algorithm 3.1 and \(p\in \Omega .\) Then, under conditions (C1)-(C4) and for all \(n\in {\mathbb {N}},\) we have
Proof
Let \(p\in \Omega \). Then, applying Lemma 2.8(iv) and (4.1), we get
Again, by invoking Lemma 2.8(ii) and applying Cauchy-Schwartz inequality we have
where \(M_2:= \sup _{n\in {\mathbb {N}}}\{||x_n - p||, \theta _n||x_n - x_{n-1}||\}>0.\)
Next, applying Lemma 2.8(iv), (4.4), (4.9) and (4.10), we obtain
Now, invoking Lemma 2.8 and applying (4.2), (4.5) and (4.11) we have
From this, we obtain
where \(M_3:=\sup \{||||x_n - p||^2 : n\in {\mathbb {N}}\}.\) This completes the proof. \(\square \)
Lemma 4.4
Let \(\{x_n\}\) be a sequence generated by Algorithm 3.1 such that conditions (C1)-(C4) hold. Then, the following inequality holds for all \(p\in \Omega \) and \(n\in {\mathbb {N}}:\)
Proof
Let \(p\in \Omega .\) From (4.4), observe that
Applying Lemma 2.8 and the firmly nonexpansivity of \(J_{\lambda _1}^{B_1},\) we have
Consequently, we have that
where \(M_4:=\sup \{\gamma _n||v_n-t_n||: n\in {\mathbb {N}}\}.\)
Next, by invoking Lemma 2.8(iv) and applying (4.2), (4.6), (4.10) and (4.12) we obtain
which is the required inequality. \(\square \)
Lemma 4.5
Let \(\{x_n\}\) be the sequence generated by Algorithm 3.1 such that conditions (A1)-(A4) and (C1)-(C4) are satisfied. Then \(\{x_n\}\) converges strongly to a point \({\hat{x}}\in \Omega ,\) where \({\hat{x}}=P_\Omega \circ f({\hat{x}}).\)
Proof
Let \({\hat{x}}=P_\Omega \circ f({\hat{x}}).\) It then follows from Lemma 4.3 that
Now, we claim that the sequence \(\{||x_n - {\hat{x}}||\}\) converges to zero. To do this, by Lemma 2.10 it suffices to show that \(\limsup _{k\rightarrow \infty }\langle f({\hat{x}}) - {\hat{x}}, x_{n_k+1} -{\hat{x}} \rangle \le 0\) for every subsequence \(\{||x_{n_k} - {\hat{x}}||\}\) of \(\{||x_n - {\hat{x}}||\}\) satisfying
Now, suppose that \(\{||x_{n_k} - {\hat{x}}||\}\) is a subsequence of \(\{||x_n - {\hat{x}}||\}\) such that
From Lemma 4.3 we have
By (4.14) together with the fact that \(\lim _{k\rightarrow \infty }\alpha _{n_k}=0,\) we obtain
Consequently, we have
Following similar argument, from Lemma 4.3 we obtain
and
Since G and H are Lipschitz continuous, then by the condition on \(\rho _n\) it follows that
From this, we obtain
Consequently, we have
From this, we get
Also, from Lemma 4.4 we have
By (4.14), and by applying (4.19) together with Remark 3.2 and the fact that \(\lim _{k\rightarrow \infty }\alpha _{n_k}=0,\) we obtain
Similarly, from Lemma 4.4 we get
By Remark 3.2, we obtain
Applying (4.15) and (4.22), we get
On the other hand, by applying (4.16), (4.20), (4.21) and (4.23) we obtain
Also, by applying (4.23)–(4.25) we get
Now, by using (4.24) together with the fact that \(\lim _{k\rightarrow \infty }\alpha _{n_k}=0,\) we have
To complete the proof, we need to show that \(w_\omega (x_n)\subset \Omega .\) First, we claim that \(w_\omega (x_n)\subset EP(F).\) Since \(\{x_n\}\) is bounded, then \(w_\omega (x_n)\) is nonempty. Let \(x^*\in w_\omega (x_n)\) be an arbitrary element. Then there exists a subsequence \(\{x_{n_k}\}\) of \(\{x_n\}\) such that \(x_{n_k}\rightharpoonup x^*\) as \(k\rightarrow \infty .\) By (4.23), it follows that \(u_{n_k}\rightharpoonup x^*\) as \(k\rightarrow \infty .\) By the definition of \(T_{r_{n_k}}^Fw_{n_k},\) we have that
It follows from the monotonicity of F that
By (4.15), \(\liminf _{k \rightarrow \infty }r_{n_k}>0,\) and condition (A4), we have that
Let \(y_t = ty + (1 - t)x^*,~~\forall ~ t \in (0, 1]\) and \(y \in C.\) This implies that \(y_t\in C,\) and it follows from (4.27) that \(F(y_t, x^*)\le 0.\) So, by applying conditions (A1)-(A4), we have
Hence, we have
Letting \(t\rightarrow 0,\) by condition (A3), we get
This implies that \(x^*\in EP(F).\)
Next, we show that \(x^* \in {\mathcal {F}}.\) By the lower semicontinuity of g, it follows from (4.17) that
which implies that
Thus, by Remark 2.5 we have that
Since \(t_{n_k} = J_{\lambda _1}^{B_1}(v_{n_k} - \gamma _{n_k}A^*(I - J_{\lambda _2}^{B_2})Av_{n_k})\) can be rewritten as
or, equivalently
By passing to limit as \(k\rightarrow \infty \) in (4.29), applying (4.19), (4.20) and (4.24), and taking into consideration the fact that the graph of a maximal monotone operator is weakly-strongly closed, we obtain \(0\in B_1(x^*).\) This together with (4.28) implies that \(x^*\in \Gamma .\)
Next, we show that \(x^*\in F(S)\cap F(T).\) By (4.24) and (4.25), we have \(t_{n_k}\rightharpoonup x^* \) and \(z_{n_k}\rightharpoonup x^* \) as \(k\rightarrow \infty .\) Since S and T are nonexpansive, by the demiclosedness principle, it follows from (4.25) that \(x^*\in F(S)\cap F(T).\) Consequently, we have that \(w_\omega (x_n)\subset \Omega .\)
Moreover, from (4.24) and (4.25) it follows that \(w_\omega \{t_n\} = w_\omega \{x_n\} =w_\omega \{z_n\}.\) By the boundedness of \(\{x_{n_k}\},\) there exists a subsequence \(\{x_{n_{k_j}}\}\) of \(\{x_{n_k}\}\) such that \(x_{n_{k_j}}\rightharpoonup x^\dagger \) and
Since \({\hat{x}}=P_\Omega \circ f({\hat{x}}),\) then it follows that
Now, from (4.26) and (4.30), we obtain
Applying Lemma 2.10 to (4.13), and using (4.31) together with the fact that \(\lim _{n\rightarrow \infty }\frac{\theta _n}{\alpha _n}||x_n - x_{n-1}|| =0\) and \(\lim _{n\rightarrow \infty }\alpha _n = 0,\) we deduce that \(\lim _{n\rightarrow \infty }||x_n - {\hat{x}}||=0\) as desired. \(\square \)
5 Applications
In this section we apply our results to study some related optimization problems.
5.1 Split minimization problem
Let \(H_1\) and \(H_2\) be real Hilbert spaces, \(A:H_1 \rightarrow H_2\) be a bounded linear operator. Given some proper, lower semicontinuous and convex functions \(f_1:H_1 \rightarrow {\mathbb {R}}\cup \{+\infty \}\) and \(f_2:H_2 \rightarrow {\mathbb {R}} \cup \{+\infty \},\) the Split Minimization Problem (SMP) is defined as
We denote the set of solution of the SMP (5.1) by \(\Gamma _{SMP}.\) The SMP was first introduced by Moudafi and Thakur (2014). It has attracted lots of attention in recent years and has been applied in the study of many applied science problems such as multi-resolution sparse regularization, Fourier regularization, hard-constrained inconsistent feasibility and alternating projection signal synthesis problems (see (Abbas et al. 2018) and the references therein).
In a real Hilbert space H, the proximal operator of f is defined by
It is well known that
where \(\partial f\) is the subdifferential of f defined by
for each \(x\in H.\) From [?], \(\partial f\) is a maximal monotone operator and \(prox_{\lambda ,f}\) is firmly nonexpansive.
By setting \(B_1 = \partial f_1\) and \(B_2 = \partial f_2\) in Theorem 4.1, we obtain the following result for approximating a common solution of split minimization problem, equilibrium problem and and common fixed point of nonexpansive mappings in Hilbert spaces.
Theorem 5.1
Let \(H_1\) and \(H_2\) be real Hilbert spaces, and \(A:H_1\rightarrow H_2\) be a bounded linear operator with adjoint \(A^*.\) Let \(f_1:H_1 \rightarrow {\mathbb {R}}\cup \{+\infty \}\) and \(f_2:H_2 \rightarrow {\mathbb {R}} \cup \{+\infty \}\) be proper, lower semicontinuous and convex functions, \(S,T:H_1\rightarrow H_1\) be nonexpansive mappings, and \(f:H_1\rightarrow H_1\) be a contraction with coefficient \(k\in (0,1).\) Suppose that \(\Omega = F(S)\cap F(T)\cap \Gamma _{SMP}\cap EP(F) \ne \emptyset ,\) and conditions (A1)-(A4) and (C1)-(C4) are satisfied. Then the sequence \(\{x_n\}\) generated by the following algorithm converges strongly to a point \({\hat{x}}\in \Omega ,\) where \({\hat{x}}=P_\Omega \circ f({\hat{x}}).\)
where
and
5.2 Split feasibility problem
Let \(H_1\) and \(H_2\) be two real Hilbert spaces and let C and Q be nonempty closed convex subsets of \(H_1\) and \(H_2,\) respectively. The Split Feasibility Problem (SFP) is defined as follows:
where \(A : H_1 \rightarrow H_2\) is a bounded linear operator. Let the solution set of SFP (5.5) be denoted by \(\Gamma _{SFP}\).
In 1994, the SFP was introduced by Censor and Elfving (1994) in finite dimensional Hilbert spaces for modelling inverse problems which arise from phase retrievals and in medical image reconstruction (Byrne 2004). Furthermore, the problem (5.5) is also useful in various disciplines such as computer tomography, image restoration, and radiation therapy treatment planning (Censor et al. 2006, 2005). The problem has been studied by numerous researchers, (see Byrne 2004; Censor et al. 2006, 2005 and the references therein).
Let C be a nonempty closed convex subset of a real Hilbert space H and \(i_c\) be the indicator function on C, that is
Moreover, we define the normal cone \(N_Cu\) of C at \(u \in C\) as follows:
It is known that \(i_C\) is a proper, lower semicontinuous and convex function on H. Hence, the subdifferential \(\partial i_C\) of \(i_C\) is a maximal monotone operator. Therefore, we define the resolvent \(J^{\partial i_C}_r\) of \(\partial i_C\), \(\forall r > 0\) as follows:
Moreover, for each \(x \in C\), we have
Hence, for all \(\alpha > 0,\) we derive
Now, by applying Theorem 4.1 we obtain the following result for approximating a common solution of split feasibility problem, equilibrium problem and common fixed point of nonexpansive mappings in Hilbert spaces.
Theorem 5.3
Let C and Q be nonempty closed convex subsets of real Hilbert spaces \(H_1\) and \(H_2,\) respectively, \(A:H_1\rightarrow H_2\) be a bounded linear operator with adjoint \(A^*.\) Let \(S,T:H_1\rightarrow H_1\) be nonexpansive mappings, and \(f:H_1\rightarrow H_1\) be a contraction with coefficient \(k\in (0,1).\) Suppose that \(\Omega = F(S)\cap F(T)\cap \Gamma _{SFP}\cap EP(F) \ne \emptyset ,\) and conditions (A1)-(A4) and (C1)-(C4) are satisfied. Then the sequence \(\{x_n\}\) generated by the following algorithm converges strongly to a point \({\hat{x}}\in \Omega ,\) where \({\hat{x}}=P_\Omega \circ f({\hat{x}}).\)
where
and
5.3 Relaxed split feasibility problem
Next, we study the Relaxed Split Feasibility Problem (RSFP) which is a special case of the split feasibility problem when the sets C and Q are defined as follows:
where \(c:H_1\rightarrow {\mathbb {R}}\) and \(q:H_2\rightarrow {\mathbb {R}}\) are convex and lower semicontinuous functions such that \(\partial c\) and \(\partial q\) are bounded on bounded sets. We denote the solution set of the RSFP by \(\Gamma _{RSFP}.\) Now, by applying Theorem 4.1 we obtain the following result for approximating a common solution of relaxed split feasibility problem, equilibrium problem and common fixed point of nonexpansive mappings in Hilbert spaces.
Theorem 5.5
Let \(H_1\) and \(H_2\) be real Hilbert spaces \(H_1\) and \(H_2,\) and let \(A:H_1\rightarrow H_2\) be a bounded linear operator with adjoint \(A^*.\) Let \(S,T:H_1\rightarrow H_1\) be nonexpansive mappings, and \(f:H_1\rightarrow H_1\) be a contraction with coefficient \(k\in (0,1).\) Suppose that \(\Omega = F(S)\cap F(T)\cap \Gamma _{RSFP}\cap EP(F) \ne \emptyset ,\) and conditions (A1)-(A4) and (C1)-(C4) are satisfied. Then the sequence \(\{x_n\}\) generated by the following algorithm converges strongly to a point \({\hat{x}}\in \Omega ,\) where \({\hat{x}}=P_\Omega \circ f({\hat{x}}).\)
where
and
6 Numerical examples
In this section, we present some numerical experiments to illustrate the performance of our method, Algorithm 3.1 in comparison with Algorithms 1.2, Algorithm 1.3, Algorithm 1.5 and Algorithm 7.2 in the literature. All numerical computations were carried out using Matlab version R2019(b).
In our computations, we choose \(\rho _n = 3-\frac{1}{2n+1}, \beta _n=\frac{n}{n+3}, \sigma _n=\frac{n}{2n+1}, \alpha _n = \frac{1}{2n+3}, \delta _n=\xi _n = \frac{n+1}{2n+3}, \mu _n=\frac{1}{(2n+3)^3}, r_n=\frac{n}{n+3}, \lambda = \lambda _1=\lambda _2=0.5, \theta =0.8\) for each \(n\in {\mathbb {N}}\). It can easily be checked that all the conditions of Theorem 4.1 are satisfied. We take \(\gamma =0.0001\) in Algorithms 1.2, 1.3 and 7.2, and \(Dx=\frac{1}{3}x, \beta =0.5\) in Algorithm 1.3.
Example 6.1
Let \(H_1={\mathbb {R}}^3=H_2\) and \(C=\{x = (x_1,x_2,x_3)\in {\mathbb {R}}^3| \langle a,x\rangle \ge b\}.\) For \(r>0,\) \(T_r^Fx = P_Cx = \frac{b - \langle a,x\rangle }{\Vert a\Vert _{2}^{2}}a + x.\) Here we take \(a = (8,-3,1), b = -1\) and define \(S,T:H_1\rightarrow H_1\) by \(Sx=\frac{1}{2}x, Tx=\frac{1}{3}x\) for all \(x\in H_1.\) Define the operators \(A, B_1\) and \(B_2\) as follows:
where \(x=(x_1,x_2,x_3)\in {\mathbb {R}}^3\) and set \(f(x) = \frac{1}{5}x.\)
We choose different initial values as follows:
Case I: \(x_0 = (-3,2,5)^T, x_1 = (2,1,-1)^T;\)
Case II: \(x_0 = (7,2.1,3.5)^T, x_1 = (5,1,2)^T;\)
Case III: \(x_0 = (2.3,4.7,-3.5)^T, x_1 = (3,1,0)^T;\)
Case IV: \(x_0 = (8,2,5)^T, x_1 = (-5,1,-1)^T.\)
Using MATLAB 2019(b), we compare the performance of Algorithm 3.1 with Algorithm 1.2, Algorithm 1.3, Algorithm 1.5 and Algorithm 7.2. The stopping criterion used for our computation is \(||x_{n+1}-x_{n}||< 10^{-3}\). We plot the graphs of errors against the number of iterations in each case. The numerical results are reported in Fig. 1 and Table 1.
Example 6.2
Let \(H_1=( l _2({\mathbb {R}}), \Vert \cdot \Vert _2)=H_2,\) where \( l _2({\mathbb {R}}):=\{x=(x_1,x_2,\ldots ,x_n,\ldots ), x_j\in {\mathbb {R}}:\sum _{j=1}^{\infty }|x_j|^2<\infty \}, ||x||_2=(\sum _{j=1}^{\infty }|x_j|^2)^{\frac{1}{2}}\) for all \(x\in l _2({\mathbb {R}}).\) We set \(f(x) = \frac{1}{2}x\) and define \(S,T:H_1\rightarrow H_1\) by \(Sx=\frac{1}{3}x, Tx=\frac{1}{5}x,\) \(A:H_1\rightarrow H_2\) be defined by \(Ax = \frac{x}{3}~~ \text {for all}~~x\in H_1,\) then \(A^*y = \frac{y}{3} ~~ \text {for all}~~ y\in H_2.\) Define \(B_1:H_1\rightarrow H_1\) by \(B_1x = \frac{5}{2}x,\) and \(B_2:H_2\rightarrow H_2\) by \(B_2x = \frac{3}{2}x.\) Then \(B_1\) and \(B_2\) are maximal monotone operators. Define the bifunction F by \(F(x,y)=x(y-x).\) It can be verified that
We choose different initial values as follows:
Case I: \(x_0 = (0, -3, 7, \dots ),\) \(x_1 = (-1, 2, 3,\dots ),\)
Case II: \(x_0 = (5, -1, \frac{1}{5}, \dots ),\) \(x_1 = (3, 0.3, 0.03, \dots ),\)
Case III: \(x_0 = (0, 3, -7, \dots ),\) \(x_1 = (1, -2, 3, \dots ),\)
Case IV: \(x_0 = (7, -3, -\frac{1}{7},\dots ),\) \(x_1 = (4, 0.4, 0.04, \dots ).\)
Using MATLAB 2019(b), we compare the performance of Algorithm 3.1 with Algorithm 1.2, Algorithm 1.3, Algorithm 1.5 and Algorithm 7.2. The stopping criterion used for our computation is \(||x_{n+1}-x_{n}||< 10^{-3}\). We plot the graphs of errors against the number of iterations in each case. The numerical results are reported in Fig. 2 and Table 2.
Remark 6.3
By using different starting points and plotting the graphs of errors against the number of iterations in each example (Examples 6.1–6.2), we obtain the numerical results displayed in Tables 1 and 2 and Figs. 1 and 2. We compared our proposed Algorithm 3.1 with Algorithm 1.2, Algorithm 1.3, Algorithm 1.5 and Algorithm 7.2. Furthermore, we note the following from our numerical experiments:
-
We observe that the different choices of the starting point and key parameters does not have a significant effects on the output of our method with respect to the performance of the proposed algorithm.
-
In all the examples, we can see from the tables and figures that the number of iterations for our proposed method remain consistent (well-behaved).
-
From the Table 2 and Fig. 2, we can see clearly that in terms of number of iterations, our proposed Algorithm 3.1 outperforms the other four existing methods. Table 1 and Fig. 1 also show that our method performs favourable well compared with the four existing methods.
7 Conclusion
In this paper, we studied the problem of finding the common solution of split variational inclusion problem, equilibrium problem and common fixed point of nonexpansive mappings. We introduced a new inertial viscosity S-iteration method for approximating the solution of the problem and we proved strong convergence theorem for the proposed algorithm without the knowledge of the operator norm. Finally, we applied our results to study other optimization problems and provided some numerical experiments with graphical illustrations to demonstrate the efficiency of our method in comparison with some existing methods in the current literature.
where \(f: H_1\rightarrow H_1\) is a contraction with constant \(k\in (0,1),~ S: H_1\rightarrow H_1\) is a nonexpansive mapping, \(\gamma \in (0, \frac{1}{L}),\) where L is the spectral radius of the operator \(A^*A,\) and \(A^*\) is the adjoint of \(A, \{\alpha _n\}\subset (0,1)\) and \(B_1: H_1\rightarrow 2^{H_1}, B_2: H_2\rightarrow 2^{H_2}\) are two multi-valued maximal monotone operators on \(H_1\) and \(H_2,\) respectively.
References
Abbas M, Al SM, Ansari QH, Iyiola GS, Shehu Y (2018) Iterative methods for solving proximal split minimization problem. Numer Algorithms 78(1):193–215
Agarwal RP, O’Regan D, Sahu DR (2007) Iterative construction of fixed point of nearly asymptotically nonexpansive mappings. J Nonlinear Convex Anal 8(1):61–79
Alakoya TO, Jolaoso LO, Mewomo OT (2021) Strong convergence theorems for finite families of pseudomonotone equilibrium and fixed point problems in Banach spaces. Afr Mat 32:897–923
Alakoya TO, Jolaoso LO, Mewomo OT (2020) Two modifications of the inertial Tseng extragradient method with self-adaptive step size for solving monotone variational inequality problems. Demonstr Math 53:208–224
Alakoya TO, Owolabi AOE, Mewomo OT (2021) An inertial algorithm with a self-adaptive step size for a split equilibrium problem and a fixed point problem of an infinite family of strict pseudo-contractions. J Nonlinear Var Anal 5:803–829
Alakoya TO, Jolaoso LO, Taiwo A, Mewomo OT (2021) Inertial algorithm with self-adaptive stepsize for split common null point and common fixed point problems for multivalued mappings in Banach spaces. Optimization. https://doi.org/10.1080/02331934.2021.1895154
Alakoya TO, Taiwo A, Mewomo OT, Cho YJ (2021) An iterative algorithm for solving variational inequality, generalized mixed equilibrium, convex minimization and zeros problems for a class of nonexpansive-type mappings. Ann Univ Ferrara Sez VII Sci Mat 67(1):1–31
Alvarez F, Attouch H (2001) An inertial proximal method for maximal monotone operators via discretization of a nonlinear oscillator with damping. Set-Valued Anal 9:3–11
Aubin JP (1993) Optima and equilibria: an introduction to nonlinear analysis. Springer, New York
Bauschke HH, Borwein JM (1996) On projection algorithms for solving convex feasibility problems. SIAM Rev 38:367–426
Blum E (1994) From optimization and variational inequalities to equilibrium problems. Math Stud 63:123–145
Bussaban L, Kettapun A (2018) Common fixed points of an iterative method for Berinde nonexpansive mappings. Thai J Math 16(1):49–60
Byrne C (2004) A unified treatment of some iterative algorithms in signal processing and image reconstruction. Inverse Probl 20:103–120
Byrne C (2002) Iterative oblique projection onto convex sets and the split feasibility problem. Inverse Probl 18:441–453
Byrne C, Censor Y, Gibali A, Reich S (2012) The split common null point problem. Nonlinear Convex Anal 13:759–775
Cai G, Gibali A, Iyiola OS, Shehu Y (2018) A new double-projection method for solving variational inequalities in Banach spaces. J Optim Theory Appl 178:219–239
Censor Y, Borteld T, Martin B, Trofimov A (2006) A unified approach for inversion problems in intensity-modulated radiation therapy. Phys Med Biol 51:2353–2365
Censor Y, Elfving T (1994) A multiprojection algorithm using Bregman projections in a product space. Numer Algorithms 8:221–239
Censor Y, Elfving T, Kopf N, Bortfeld T (2005) The multiple-sets split feasibility problem and its applications. Inverse Probl 21:2071–2084
Chen P, Huang J, Zhang X (2013) A primal-dual fixed point algorithm for convex separable minimization with applications to image restoration. Inverse Probl 29(2):Art. ID 025011
Cholamjiak P, Suantai S (2013) Iterative methods for solving equilibrium problems, variational inequalities and fixed points of nonexpansive semigroups. J Glob Optim 57:1277–1297
Chuang CS (2013) Strong convergence theorems for the split variational inclusion problem in Hilbert spaces. Fixed Point Theory Appl 2013:350
Combettes PL (1996) The convex feasibility problem in image recovery. Adv Imaging Electron Phys 95:155–270
Gibali A, Shehu Y (2019) An efficient iterative method for finding common fixed point and variational inequalities in Hilbert spaces. Optimization 68(1):13–32
Godwin EC, Izuchukwu C, Mewomo OT (2020) An inertial extrapolation method for solving generalized split feasibility problems in real Hilbert spaces. Boll Unione Mat Ital. https://doi.org/10.1007/s40574-020-00
Halpern B (1967) Fixed points of nonexpanding maps. Bull Am Math Soc 73:957–961
Iiduka H (2012) Fixed point optimization algorithm and its application to network bandwidth allocation. J Comput Appl Math 236:1733–1742
Iyiola OS, Ogbuisi FU, Shehu Y (2018) An inertial type iterative method with Armijo linesearch for nonomonotone equilibrium problems. Calcolo 55(4):1–22
Izuchukwu C, Ogwo GN, Mewomo OT (2020) An inertial method for solving generalized split feasibility problems over the solution set of monotone variational inclusions. Optimization. https://doi.org/10.1080/02331934.2020.1808648
Jolaoso LO, Taiwo A, Alakoya TO, Mewomo OT (2020) Strong convergence theorem for solving pseudo-monotone variational inequality problem using projection method in a reflexive Banach space. J Optim Theory Appl 185(3):744–766
Kazmi KR, Rizvi SH (2014) An iterative method for split variational inclusion problem and fixed point problem for a nonexpansive mapping. Optim Lett 8(3):1113–1124
Khan SH, Alakoya TO, Mewomo OT (2020) Relaxed projection methods with self-adaptive step size for solving variational inequality and fixed point problems for an infinite family of multivalued relatively nonexpansive mappings in Banach spaces. Math Comput Appl 25:Art. 54
Luo C, Ji H, Li Y (2009) Utility-based multi-service bandwidth allocation in the 4G heterogeneous wireless networks. IEEE Wireless Commun Netw Conf 5:1–5. https://doi.org/10.1109/WCNC.2009.4918017
Ma Z, Wang L, Chang SS, Duan W (2015) Convergence theorems for split equality mixed equilibrium problems with applications. Fixed Point Theory Appl. 2015, Art 31
Maingé PE (2007) Approximation methods for common fixed points of nonexpansive mappings in Hilbert spaces. J Math Anal Appl 325:469–479
Moudafi A (2011) Split monotone variational inclusions. J Optim Theory Appl 150:275–283
Moudafi A (2000) Viscosity approximation methods for fixed-point problems. J Math Anal Appl 241:46–55
Moudafi A, Oliny M (2003) Convergence of a splitting inertial proximal method for monotone operators. J Comput Appl Math 155(2):447–454
Moudafi A, Thakur BS (2014) Solving proximal split feasibility problems without prior knowledge of operator norms. Optim Lett 8:2099–2110
Nesterov Y (1983) A method of solving a convex programming problem with convergence rate O(1/sqr(k)). Sov Math Dokl 27:372–376
Muu LD, Oettli W (1992) Convergence of an adaptive penalty scheme for finding constrained equilibria. Nonlinear Anal TMA 18(12):1159–1166
Ogwo GN, Alakoya TO, Mewomo OT (2021) Iterative algorithm with self-adaptive step size for approximating the common solution of variational inequality and fixed point problems. Optimization. https://doi.org/10.1080/02331934.2021.1981897
Ogwo GN, Izuchukwu C, Mewomo OT (2021) Inertial methods for finding minimum-norm solutions of the split variational inequality problem beyond monotonicity. Numer Algorithms. https://doi.org/10.1007/s11075-021-01081-1
Olona MA, Alakoya TO, Owolabi AO-E, Mewomo OT (2021) Inertial shrinking projection algorithm with self-adaptive step size for split generalized equilibrium and fixed point problems for a countable family of nonexpansive multivalued mappings. Demonstr Math. https://doi.org/10.1515/dema-2021-0006
Olona MA, Alakoya TO, Owolabi AO-E, Mewomo OT (2021) Inertial algorithm for solving equilibrium, variational inclusion and fixed point problems for an infinite family of strictly pseudocontractive mappings. J Nonlinear Funct Anal 2021:21 (Art. ID 10)
Owolabi AO-E, Alakoya TO, Taiwo A, Mewomo OT (2021) A new inertial-projection algorithm for approximating common solution of variational inequality and fixed point problems of multivalued mappings. Numer Algebra Control Optim. https://doi.org/10.3934/naco.2021004
Oyewole OK, Abass HA, Mewomo OT (2021) A Strong convergence algorithm for a fixed point constrainted split null point problem. Rend Circ Mat Palermo II 70(1):389–408
Oyewole OK, Aremu KO, Mewomo OT (2021) A multi step inertial algorithm for approximating a common solution of split generalized mixed equilibrium and minimization problems. Ric Mat. https://doi.org/10.1007/s11587-021-00624-x
Pandey R, Pant R, Rakocević V, Shukla R (2019) Approximating fixed points of a general class of nonexpansive mappings in Banach spaces with applications. Result Math 74(7)
Patriksson M (2015) The traffic assignment problem: models and methods. Courier Dover Publications, Mineola, NY
Polyak BT (1964) Some methods of speeding up the convergence of iteration methods. Politehn Univ Bucharest Sci Bull Ser A Appl Math Phys 4(5):1–17
Saejung S, Yotkaew P (2012) Approximation of zeros of inverse strongly monotone operators in Banach spaces. Nonlinear Anal 75:742–750
Shehu Y, Iyiola OS (2020) Projection methods with alternating inertial steps for variational inequalities: weak and linear convergence. Appl Numer Math 157:315–337
Shehu Y, Iyiola OS, Ogbuisi FU (2020) Iterative method with inertial terms for nonexpansive mappings: applications to compressed sensing. Numer Algorithms 83:1321–1347
Shehu Y, Iyiola OS, Thong DV, Van NTC (2021) An inertial subgradient extragradient algorithm extended to pseudomonotone equilibrium problems. Math Meth Oper Res 93:213–242
Shehu Y, Ogbuisi FU (2015) Convergence analysis for proximal split feasibility problems and fixed point problems. J Appl Math Comput 48:221–239
Taiwo A, Alakoya TO, Mewomo OT (2021) Strong convergence theorem for solving equilibrium problem and fixed point of relatively nonexpansive multi-valued mappings in a Banach space with applications. Asian-Eur J Math 14(8):31 (Art. ID 2150137)
Taiwo A, Alakoya TO, Mewomo OT (2021) Halpern-type iterative process for solving split common fixed point and monotone variational inclusion problem between Banach spaces. Numer Algorithms 86(12):1–30
Taiwo A, Mewomo OT, Gibali A (2021) A simple strong convergent method for solving split common fixed point problems. J Nonlinear Var Anal 5:777–793
Tang Y (2020) Convergence analysis of a new iterative algorithm for solving split variational inclusion problems. J Indus Mgt Opt 16(2):945
Wangkeeree R, Rattanaseeha K, Wangkeeree R (2018) The general iterative methods for split variational inclusion problem and fixed point problem in Hilbert spaces. J Comp Anal Appl 25(1):19–31
Zhao J, Liang Y, Liu Y, Cho YJ (2018) Split equilibrium, variational inequality and fixed point problems for multi-valued mappings in Hilbert spaces. Appl Comput Math 17(3):271–283
Acknowledgements
The authors sincerely thank the editor and anonymous referees for the careful reading, constructive comments and fruitful suggestions that substantially improved the manuscript. The research of the first author is wholly supported by the University of KwaZulu-Natal, Durban, South Africa Postdoctoral Fellowship. He is grateful for the funding and financial support. The second author is supported by the National Research Foundation (NRF) of South Africa Incentive Funding for Rated Researchers (Grant Number 119903). Opinions expressed and conclusions arrived are those of the authors and are not necessarily to be attributed to the NRF.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no competing interests.
Additional information
Communicated by Orizon Pereira Ferreira.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Alakoya, T.O., Mewomo, O.T. Viscosity S-iteration method with inertial technique and self-adaptive step size for split variational inclusion, equilibrium and fixed point problems. Comp. Appl. Math. 41, 39 (2022). https://doi.org/10.1007/s40314-021-01749-3
Received:
Revised:
Accepted:
Published:
DOI: https://doi.org/10.1007/s40314-021-01749-3
Keywords
- S-iteration method
- Inertial technique
- Self-adaptive step size
- Split variational inclusion problem
- Equilibrium problem
- Nonexpansive mappings