Abstract
Under the assumption that the sectional curvature of the manifold is bounded from below, we establish convergence result about the cyclic subgradient projection algorithm for convex feasibility problem presented in a paper by Bento and Melo on Riemannian manifolds (J Optim Theory Appl 152, 773–785, 2012). If, additionally, we assume that a Slater type condition is satisfied, then we further show that, without changing the step size, this algorithm terminates in a finite number of iterations. Clearly, our results extend the corresponding ones due to Bento and Melo and, in particular, we solve partially the open problem proposed in the paper by Bento and Melo.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
The convex feasibility problem (CFP, for short), which consists in finding a point in the intersection of finitely many closed and convex sets in Hilbert spaces, is a very broad and adaptive framework, and its study has a long and rich history in applied mathematics. The main approach to solve this problem is offered by the projection-like algorithms, which probably goes back to the work by von Neumann in the early 1930s for the special case, when there are only two sets. For details, we refer the reader to [1–10] and the references therein. For the special situation, where each convex set is given as a sublevel set of a convex function, Censor and Lent proposed in [11] a cyclic subgradient projection algorithm by using, instead of the orthogonal projections to the convex sets, the projections of the corresponding subgradients. As pointed out in [12], this special approach does not lose the generality because one can certainly choose the functions as the squared Euclidean distance functions to the involved convex sets to cover the general case. Since then, some authors have incorporated acceleration procedures in order to improve the speed of convergence; see, e.g., [2, 8, 12–15].
Recent interests are focused on extensions of some important notions and techniques in Hilbert spaces to Riemannian manifolds. The reason is that, as explained in [16], some optimization problems, arising in various applications, cannot be posted in linear spaces and require Riemannian manifold structures for their formalization and study; while some nonconvex and/or nonsmooth problems of constrained optimization in \({\mathbb {R}}^n\) can be reduced to convex and smooth unconstrained optimization problems on appropriate Riemannian manifolds; see, for examples, [17–25]. For developments of optimization techniques on Riemannian manifolds, we refer the reader to [16, 26–37] and the bibliographies therein for various results, examples, discussions, and applications. Here, let us particularly mention the CFP on Riemannian manifolds with each convex set being given as a sublevel set of a convex function and the cyclic subgradient projection algorithm (i.e., Algorithm 3.1) for solving it; this algorithm was first proposed in [38], but the idea comes from the book of Udriste, that is, the general descent and/or gradient algorithms studied in [25], Chap. 7, §3 and §4] for finding a minimizer of a \(C^2\)-function \(f\) on Riemannian manifolds. Convergence properties of the gradient algorithm have also been studied in [24, 30, 34, 37, 39]. The CFP can be reformulated as a convex minimization problem on Riemannian manifolds, but the induced convex function \(f\) on Riemannian manifolds is nonsmooth in general; hence, the known results in [25] and the literature mentioned above cannot be applied to solve the CFP on Riemannian manifolds considered here, nor establish the convergence properties of the cyclic subgradient projection algorithm.
In the special case, when the underlying Riemannian manifold is of non-negative sectional curvature, Bento and Melo proved in [38] that the cyclic subgradient projection algorithm converges to a solution of the CFP, and if, additionally, a Slater type condition is satisfied, then this algorithm has the finite termination property. Moreover, the following open problem was proposed there: How is the CFP solved on Riemannian manifolds with negative sectional curvature?
The main purpose of the present paper is to study the convergence problem of the cyclic subgradient projection algorithm on Riemannian manifolds with possible non-positive sectional curvature. More precisely, under the assumption that the underlying Riemannian manifold is of sectional curvature bounded from below, we show that, employing the same step sizes as in [38], any sequence generated by the cyclic subgradient projection algorithm converges to a solution of the CFP, and if a Slater type condition is additionally satisfied, then this algorithm terminates in a finite number of iterations. Our results extend the corresponding ones in [38], and, in particular, the open problem mentioned above is solved partially.
The paper is organized as follows. We recall some basic notions, notations, and properties of Riemannian geometry, together with some definitions and preliminary results on convex analysis on Riemannian manifolds in Sect. 2 . In Sect. 3 , the cyclic subgradient projection algorithm considered in [38] is introduced and some useful lemmas are shown. The main results of the present paper, including the convergence theorems and the finite termination theorems of this algorithm, are presented in Sects. 4 and 5, respectively.
2 Notations and Preliminaries
The notations used in the present paper are standard; and the readers are referred to some textbooks for more details, for example, [25, 40, 41].
Let \(M\) be a connected \(n\)-dimensional Riemannian manifold. We use \(\nabla \) to denote the Levi-Civita connection on \(M\). Let \(x\in M\), let \(T_{x}M\) denote the tangent space at \(x\) to \(M\). We denote by \(\langle ,\rangle _{x}\) the scalar product on \(T_{x}M\) with the associated norm \(\Vert .\Vert _{x}\), where the subscript \(x\) is sometimes omitted. For \(y\in {M}\), let \(\gamma :[0,1]\rightarrow M\) be a piecewise smooth curve joining \(x\) to \(y\). Then, the arc-length of \(\gamma \) is defined by \(l(\gamma ):=\int _{0}^{1}\Vert \gamma '(t)\Vert \mathrm{d}t\), while the Riemannian distance from \(x\) to \(y\) is defined by \(d(x,y):=\inf _{\gamma }l(\gamma )\), where the infimum is taken over all piecewise smooth curves \(\gamma :[0,1]\rightarrow M\) joining \(x\) to \(y\). A vector field \(V\) is said to be parallel along \(\gamma \) if \(\nabla _{\gamma '}V=0\). In particular, for a smooth curve \(\gamma \), if \(\gamma '\) is parallel along itself, then \(\gamma \) is called a geodesic, that is, a smooth curve \(\gamma \) is a geodesic if and only if \(\nabla _{\gamma '}{\gamma '}=0\). A geodesic \(\gamma :[0,1]\rightarrow M\) joining \(x\) to \(y\) is minimal if its arc-length equals its Riemannian distance between \(x\) and \(y\). By the Hopf–Rinow theorem [41], \((M,d)\) is a complete metric space, and there is at least one minimal geodesic joining \(x\) to \(y\). The closed metric ball in \(M\) centered at the point \(x\in M\) with radius \(r>0\) is denoted by \(\mathbb {B}(x,r)\), i.e.,
A complete simply connected Riemannian manifold of non-positive sectional curvature is called a Hadamard manifold. One important example of Hadamard manifold is the \(m\)-dimensional hyperbolic space \({\mathbb {H}}^m\), which is defined as follows:
where \(\langle \cdot ,\cdot \rangle _{\mathbb {H}}\) denotes the symmetric bilinear form (which is called the Lorentz metric on \({\mathbb {R}}^{m+1}\)), defined by
Then, \({\mathbb {H}}^m\) is a Hadamard manifold with sectional curvature \(-1\) [20, 27]. Let \(\kappa <0\) and let \(M_{\kappa }^{m}\) be the Hadamard manifold obtained from the \(m\)-dimensional hyperbolic space \({\mathbb {H}}^m\) by multiplying the distance function by the constant \(\frac{1}{\sqrt{|\kappa |}}\).
Consider two geodesic segments \(\gamma _1,\gamma _2:[0,+\infty [\rightarrow M\) emanating from \(p\) (i.e., \(\gamma _1(0)=\gamma _2(0)=p\)). Let \(\angle _p(\gamma _1,\gamma _2)\) denote the angle between \(\gamma _1\) and \(\gamma _2\) at \(p\), which is defined to be the angle between the tangent vectors \(\gamma _1'(0)\) and \(\gamma _2'(0)\). By [27, p. 173, Corollary 1A7.], \(\angle _p(\gamma _1,\gamma _2)\) coincides with the Alexandrov angle. Recall that a geodesic triangle \(\triangle (p_{1}p_{2}p_{3})\) in \(M\) is a figure consisting of three points \(p_{1},p_{2},p_{3}\) (the vertices of \(\triangle (p_{1}p_{2}p_{3})\)) and three geodesic segments \(\gamma _{i}\) (the edges of \(\triangle (p_{1}p_{2}p_{3})\)) that join \(p_{i-1}\) to \(p_{i+1}\) with \(i=1,2,3\) (mod3). For each \(i=1,2,3\) (mod3), the inner angle of \(\triangle (p_{1}p_{2}p_{3})\) at \(p_{i}\) is denoted by \(\angle (p_{i-1}p_{i}p_{i+1})\), which equals \(\angle _{p_i}(-\gamma _{i-1},\gamma _{i+1})\). The following proposition is known in [40, p. 138].
Proposition 2.1
Let \(\kappa <0\) and let \(\triangle (p_{1}p_{2}p_{3})\) be a geodesic triangle in \(M_{\kappa }^{2}\). Then, the following “law of cosines” holds
where \(l_i=d(p_{i-1},p_{i+1})\) for each \(i=1,2,3 \) (mod3).
Another important tool that will be used is the comparison theorem described in Proposition 2.2 below, which is known in [40, p. 161, Theorem 4.2].
Following [40, p. 161], a generalized geodesic hinge \(\varLambda (p;\gamma ,\tau )\) in \(M\) is a figure consisting of a point \(p\in M\) (the vertex of the hinge) and two geodesic segments \(\gamma ,\tau \) (the edges of the hinge) emanating from \(p\) with one being minimal. Moreover, a hinge \(\varLambda (\bar{p};\bar{\gamma },\bar{\tau })\) in \(M_{\kappa }^{2}\) is called a comparison hinge of \(\varLambda (p;\gamma ,\tau )\) if it satisfies
Proposition 2.2
Let \(\kappa <0\) and suppose that \(M\) is a complete Riemannian manifold of sectional curvature bounded from below by \(\kappa \). Let \(\varLambda (p;\gamma ,\tau )\) be a generalized geodesic hinge in \(M\). Then, there exists a comparison hinge \(\varLambda (\bar{p};\bar{\gamma },\bar{\tau })\) in \(M_{\kappa }^{2}\) such that \(d(q_\gamma ,q_\tau )\le d(\bar{q}_{\bar{\gamma }},\bar{q}_{\bar{\tau }})\), where \(q_\gamma \), \(q_\tau \), \(\bar{q}_{\bar{\gamma }}\), and \(\bar{q}_{\bar{\tau }}\) denote the end points of \(\gamma \), \(\tau \), \(\bar{\gamma }\), and \(\bar{\tau }\), respectively.
Let \(\varGamma _{xy}\) denote the set of all geodesics \(\gamma :[0,1]\rightarrow M\) with \(\gamma (0)=x\) and \(\gamma (1)=y\). Then, \(\varGamma _{xy}\) is nonempty for all \(x,y\in M\). Consider next a real-valued function \(f:M\rightarrow \mathbb {R}\). We say that \(f\) is convex on \(M\) if, for any \(x,y\in M\) and any \(\gamma \in \varGamma _{xy}\), the composition \((f\circ \gamma ):[0,1]\rightarrow \mathbb {R}\) is a convex function on \([0,1]\), i.e.,
For the remainder, we always assume that \(f:M\rightarrow \mathbb {R}\) is convex on \(M\). Then, as pointed out in [21, p. 642] (see also [25, p. 70]), \(f\) is Lipschitz continuous on any compact subset of \(M\). This, in particular, implies that \(f\) is continuous on \(M\). Let \(x\in M\). We define the subdifferential \(\partial f(x)\) of \(f\) at \(x\) by
Then, \(\partial f(x)\) is a nonempty, closed convex set in the space \(T_{x}M\). Furthermore, by the Lipschitz continuity, one sees that
(see also [38]).
We end this section with the notions of the distance function and the projection associated to a subset \(Q\) of \(M\). The distance function \(d_{Q}(\cdot )\) of \(Q\) is defined by
while the projection \(P(\cdot |Q)\) on \(Q\) is defined by
Note that, in general, \(P(\cdot |Q)\) is a set-valued map even if \(Q\) is convex in the sense that, for any \(y,z\in Q\), any minimal geodesic joining \(y\) to \(z\) is in \(Q\). For example, consider \(M:={\mathbb {S}}^2\), the two dimensional sphere, and
Let \(x_0:=(0,0,-1)\). Then, \(d_{Q}(x_0)=\frac{\pi }{2}\) and \(P(x_0|Q)=\{ (t_1,t_2,0)\in Q\}\).
3 Algorithm and Auxiliary Results
Let \(I:=\{1,2,\ldots ,m\}\), where \(m\in {\mathbb {N}}\), and let \(f_{i}:\ M\rightarrow \mathbb {R}\) be a convex function for each \(i\in I\). Consider a family of closed and convex subsets \(\{Q_i:i\in I\}\) in \(M\) with each \(Q_i\) given by
The CFP on Riemannian manifold \(M\) considered here consists in finding a point \(x^*\) satisfying
Let \(f:\ M\rightarrow \mathbb {R}\) be the sup-function of \(\{f_i:i\in I\}\) defined by
Then, \(f\) is convex on \(M\), and the solution set \(Q\) is re-expressed as
We assume throughout the whole paper that \(Q\ne \emptyset \).
The following algorithm was proposed in [38] by Bento et al. for solving the CFP on Riemannian manifolds, which is actually suggested by the book of Udriste (see [25], Chap. 7, §3 and §4]); in particular, this algorithm coincides with “the gradient method” described in [25, p. 263] in the case when \(f\) is smooth.
Algorithm 3.1
-
Step 1. Select \(x^{0}\in M\) and put \(k:=0\).
-
Step 2. If \(f(x^{k})\le 0\), then stop; otherwise, go to Step 3.
-
Step 3. Put \(i_{k}:=(k\,\mathrm{mod}\,m)+1\).
-
Step 4. If \(f_{i_{k}}(x^{k})\le 0\), then set \(x^{k+1}:=x^{k}\); otherwise, go to Step 5.
-
Step 5. Select \(g^{k}\in \partial f_{i_{k}}({x^{k}})\), \(t_k \in (0,+\infty )\) and construct the geodesic \(\gamma _k\) such that
$$\begin{aligned} \gamma _{k} (0)=x^{k}\quad \mathrm{and}\quad \gamma '_k(0)=-\frac{g^{k}}{\Vert g^{k}\Vert } . \end{aligned}$$(2) -
Step 6. Set \(x^{k+1}:=\gamma _{k}(t_{k})\).
-
Step 7. Replace \(k\) by \(k+1\) and go to Step 2.
Remark 3.1
-
(i)
Note that the selected \(g^{k}\in \partial f_{i_{k}}({x^{k}})\) in Step 5 is nonzero (otherwise, \(x^{k}\) is a minimum of \(f_{i_{k}}\) and then, \(f_{i_{k}}(x^{k})\le 0\)). This means that Algorithm 3.1 is well defined.
-
(ii)
The sequence \(\{t_{k}\}\) is called the step size sequence. By Algorithm 3.1, we have that
$$\begin{aligned} d(x^{k},x^{k+1})\le t_{k}\quad \mathrm{for\,each}\; k\in {\mathbb {N}}. \end{aligned}$$(3)To analyze the convergence of Algorithm 3.1, we first establish the basic inequality on Riemannian manifolds. For simplicity, we define the function \(\hbar :]0,\infty [\rightarrow {\mathbb {R}}\) by
$$\begin{aligned} \hbar (t):=\frac{\tanh t}{t}\quad \mathrm{for\,any}\; t\in ]0,\infty [. \end{aligned}$$It is easy to verify that \(\hbar \) is strictly decreasing on \(]0,\infty [\).
For the remainder of the paper, we assume that the Riemannian manifold \(M\) is of sectional curvature bounded from below by some negative number \(\kappa \).
Lemma 3.1
Let \(\{x^{k}\}\) be a sequence generated by Algorithm 3.1. Let \(z\in M\) and let \(k\ge 0\) be such that \(f_{i_{k}}(x^{k})>0\). Then, the following inequality holds:
Proof
Without loss of generality, we assume that \(\kappa =-1\). Take \(z\in M\) and let \(\tau _{k}:[0,1]\rightarrow M\) be a minimal geodesic segment joining \(x^{k}\) and \(z\) with \(\tau _k(0)=x^{k}\) and \(\tau _k(1)=z\). Let \(\gamma _{k}:[0, t_k]\rightarrow M\) be the geodesic segment defined by Algorithm 3.1. Then,
Consider the generalized geodesic hinge \(\varLambda (x^{k};\gamma _{k},\tau _{k})\) in \(M\), and let \(\alpha := \angle _{x^k}(\gamma _k,\tau _k)\). Then \(\alpha \) coincides with the angle between the tangent vectors \(\gamma '_{k}(0)\) and \(\tau '_{k}(0)\) in the tangent space \(T_{x^k}M\). This, together with (2), implies that
By Proposition 2.2, there exists a comparison geodesic hinge \(\varLambda (\bar{x}^{k};\bar{\gamma }_{k},\bar{\tau }_{k})\) of the hinge \(\varLambda (x^{k};\gamma _{k},\tau _{k})\) in \(M_{-1}^{2}\) such that
where \(\bar{x}^{k+1}\) and \(\bar{z}\) are the end points of \(\bar{\gamma }_{k}\) and \(\bar{\tau }_{k}\), respectively. Consider the geodesic triangle \(\triangle (\bar{x}^{k}\bar{x}^{k+1}\bar{z})\) in \(M_{-1}^{2}\). By the definition of the comparison geodesic hinge and using (5), we have that
Then, applying the “law of cosines” to the geodesic triangle \(\triangle (\bar{x}^{k}\bar{x}^{k+1}\bar{z})\) in \(M_{-1}^{2}\), we get
Combing this and (7) yields that
By elementary calculus, one can easily verify that
It follows from (8) that
Recall that \(g^{k}\in \partial f_{i_{k}}(x^{k})\). By equality (6), one has that
hence,
Substituting this into (9) and recalling the definition of the function \(\hbar \), we see that (4) holds, and the proof is complete. \(\square \)
Set
Then, \(f^{*}\le 0\) by the nonemptiness assumption of \(Q\). Throughout the whole paper, let \(\{f^*_i:\,i\in I\}\subseteq {\mathbb {R}}\) satisfy
and choose the step size sequence \(\{t_k\}\) as follows:
where \(i_{k}=(k\,\mathrm{mod}\,m)+1\). Set
Note that \(C\subseteq Q\) and the equality holds if each \(f_i^*=0\). Furthermore, let \(x^0\in M\) and write \(r_0:= d_C(x^{0})\).
Lemma 3.2
Let \(x^0\in M\) and suppose that
Let \(\{x^{k}\}\) be a sequence generated by Algorithm 3.1 with initial point \(x^0\). Set
and let \(z\in \tilde{C}\). Then, the following estimate holds for each \(k\in {\mathbb {N}}\):
where
Proof
As in the proof of Lemma 3.1, we also assume that \(\kappa =-1\). By the definition of \(\sigma \), assumption (11) implies that \(\sigma \in ]0,1[\) and
Without loss of generality, we assume that \(f_{i_{k}}(x^{k})> 0\) (otherwise, \(x^{k+1}=x^k\) and (12) holds trivially by (10)). Then,
(noting that \(z\in \tilde{C}\)). By the choice of \(t_k\) and noting (13), one sees that
This, together with (14), implies that
Applying (4), we get that
Since \(\sigma >0\), we have that
Below, we use mathematical induction to show that
holds for each \(k\in {\mathbb {N}}\). By the choice of \(z\) and the definition of \( \tilde{C}\), we can choose \(\omega \in P(x^0|C)\) such that \(\mathrm{d}(z,\omega )\le r_0\). Thus, (17) is clear for \(k=0\), because
Now, suppose that (17) is valid for \(k=n\), that is,
Then,
(as \(\hbar \) is decreasing on \(]0,+\infty [\)). It follows from (16) that
hence, \( d(x^{n+1},z)\le d(x^{n},z)\). This, together with (18), yields that (17) holds for \(k=n+1\). Therefore, (17) is checked for all \(k\in {\mathbb {N}}\), and so (19) holds for all \(n\in {\mathbb {N}}\). Thus, combining (15) and (19) (with \(k\) in place of \(n\)), one concludes that
Recalling (17), one can use Taylor expansion for the function \(\cosh (\cdot )\) to verify the following estimate:
where \(\beta :=\frac{\sinh (2r_0)}{4r_0}\). This, together with (20), implies that
Since \(\sinh t_{k}>t_{k}\) and \(\cosh d(x^{k},z)\ge 1\), it follows that
On the other hand, using (10) and (13), we see that
This, together with (21), yields (12). The proof is complete. \(\square \)
Remark 3.2
Suppose that assumption (11) holds. Then, the following assertions hold by Lemma 3.2:
-
(i)
The sequence \(\{x^{k}\}\subseteq {\mathbb {B}}(\omega ,r_{0})\) for any point \(\omega \in P(x^{0}|C)\), and is Fejér convergent to \(\tilde{C}\) in the sense that
$$\begin{aligned} d(x^{k+1},z)\le d(x^{k},z)\quad \mathrm{for\,each}\; k\in {\mathbb {N}}\quad \mathrm{and}\quad z\in \tilde{C}. \end{aligned}$$(22) -
(ii)
By (1), we have that
$$\begin{aligned} L:=\sup \{\Vert g\Vert :\; g\in \partial f_{i}(x^k),\,(i,k)\in I\times {\mathbb {N}}\}<+\infty . \end{aligned}$$(23) -
(iii)
$$\begin{aligned} \lim _{k\rightarrow \infty } d(x^{k+1},x^{k})=\lim _{k\rightarrow \infty }t_{k}=0. \end{aligned}$$(24)
4 Convergence Results
We assume in this section that \(f_i^*=0\) for each \(i\in I\) in the step size sequence \(\{t_k\}\), that is Algorithm 3.1 employs the following step sizes
Then, one has that \(C=Q\). The main result of this section is as follows.
Theorem 4.1
Let \(x^0\in M\) and suppose that assumption (11) holds. Then, any sequence \(\{x^{k}\}\) generated by Algorithm 3.1 with initial point \(x^0\) converges to a point \(x^{\infty }\in Q\).
Proof
By assumption (11), Remark 3.2 is applicable. Then, \(\{x^k\}\) satisfies (22), (24), and that
Thus, there exists a subsequence \(\{x^{k_j}\}\) of \(\{x^{k}\}\) such that \(x^{k_j}\rightarrow x^\infty \in M\). Below, we verify that \(x^\infty \in Q\), that is, \(f_l(x^\infty )\le 0\) for each \(l\in I\). To do this, recall that \(i_{k_j}=(k\,\mathrm{mod}\,m)+1\) and so \(i_k\in I\). Thus, without loss of generality, we assume that \(i_{k_j}=m\) for each \(j\). Let \(l\in I\) and consider the subsequence \(\{x^{k_j+l}\}_{j=0}^\infty \). Then,
where the first assertion is true by (24). Recalling the choice of \(t_k\), we have that, if \(f_l(x^{k_{j}+l})> 0\), then
where \(L>0\) is given by (23); hence,
This inequality holds trivially if \(f_l(x^{k_{j}+l})\le 0\). Thus, (27) holds for each \(j\in {\mathbb {N}}\). Since \(f_l\) is continuous on \(M\), it follows from (26) that
Therefore, \(x^\infty \in Q\) as \(l\in I\) is arbitrary. Moreover, by (25), we get that \( d(x^{\infty },\omega )\le r_0\), and so \(x^\infty \in \tilde{C}\) (noting that \(C=Q\)). Appying (22), one concludes that \(\{x^{k}\}\) converges to \(x^{\infty }\) and the proof is complete. \(\square \)
As a consequence of Theorem 4.1, we get the following corollary, which was proved in [38].
Corollary 4.1
Suppose that the sectional curvature of \(M\) is non-negative and that
Then, any sequence \(\{x^{k}\}\) generated by Algorithm 3.1 with initial point \(x^0\in M\) converges to a point \(x^{\infty }\in Q\).
Proof
Let \(x^0\in M\). Recall that function \(\hbar \) is continuous on \(]0,\infty [\) and \(\hbar (0)=1\). Then, by assumption (28), one can choose \( \kappa <0\) such that assumption (11) holds. Furthermore, by assumption, it is clear that the sectional curvature of \(M\) is bounded from below by \( \kappa \). Thus, Theorem 4.1 is applicable to complete the proof. \(\square \)
5 Finite Convergence
Recall that the step size sequence is given as in (10):
where \(i_{k}=(k\,\mathrm{mod}\,m)+1\) and \( f^*\le f_i^*\le 0\) for each \(i\in I\). We suppose in this section that the CFP on \(M\) satisfies the Slater condition:
and that \(\{f^*_i:i\in I\}\) satisfies
The following theorem shows that Algorithm 3.1 terminates in a finite number of iterations.
Theorem 5.1
Let \(x^0\in M\) and suppose that assumption (11) holds. Then, any sequence \(\{x^{k}\}\) generated by Algorithm 3.1 with initial point \(x^0\) terminates in a finite number of iterations.
Proof
Suppose on the contrary that Algorithm 3.1 generates an infinite sequence \(\{x^k\}\), which satisfies that
This means that there exists a subsequence \(\{x^{k_j}\}\) such that
where \(i_{k_j}=(k_j\,\mathrm{mod}\,m)+1\) as before. Moreover, by assumption (11), Remark 3.2 is applicable. Therefore, \(\{x^k\}\) is bounded and (24) holds. Thus, without loss of generality, we may further assume that \(\{x^{k_j}\}\) converges to some point \(x^\infty \in M\) and \(i_{k_j}=1\) for each \(j\) (as each \(i_{k_j}\) is in the finite set \(I\)). Since \(f_1\) is continuous on \(M\), it follows from (29) that
Furthermore, we have that
where \(L>0\) is given by (23). Passing to the limit and making use of (24), we get that \(f_1(x^{\infty })-f_1^*\le 0\); hence, \(f_1(x^\infty )\le f_1^*<0\), which contradicts (30). Hence, the proof is complete. \(\square \)
With a similar argument that we did for Corollary 4.1, but using Theorem 5.1 in place of Theorem 4.1, we can show the following corollary, which was proved in [38, Theorem 5.1].
Corollary 5.1
Suppose that the sectional curvature of \(M\) is non-negative and that
Then, any sequence \(\{x^{k}\}\) generated by Algorithm 3.1 with initial point \(x^0\in M\) terminates in a finite number of iterations.
6 Conclusions
By using a different approach, we generalize the results in [38] for the Riemannian manifolds with non-negative sectional curvature to those with sectional curvature bounded from below. In particular, the open problem proposed in [38] has been solved partially. It is still an open problem how to establish similar convergence results on Riemannian manifolds without the assumption that the sectional curvature is bounded from below. Note that, in the Hilbert space setting, if the CFP satisfies a Slater type condition, then the cyclic subgradient projection algorithm is linearly convergent [1]. Our future work is to establish the linear convergence result about the algorithm on Riemannian manifolds under a Slater type condition. Moreover, recalling that every complete, connected Riemannian manifold is a special geodesic metric space (as pointed out in [27]), we expect that our approach in the present paper can be used to study similar problems in the setting of geodesic spaces (with curvature bounded from below).
References
Bauschke, H.H., Borwein, J.M.: On projection algorithms for solving convex feasibility problems. SIAM Rev. 38(3), 367–426 (1996)
Bauschke, H.H., Combettes, P.L., Kruk, S.G.: Extrapolation algorithm for affine–convex feasibility problems. Numer. Algorithms 41, 239–274 (2006)
Censor, Y.: Mathematical Optimization for the Inverse Problem of Intensity-Modulated Radiation Therapy. Medical Physics Publisher, Madison (2003)
Censor, Y., Altschuler, M.D., Powlis, W.D.: On the use of Cimmino’s simultaneous projections method for computing a solution of the inverse problem in radiation therapy treatment planning. Inverse Probl. 4, 607–623 (1988)
Censor, Y., Zenios, S.A.: Parallel Optimization: Theory, Algorithms, and Applications. Oxford University Press, New York (1997)
Combettes, P.L.: The foundations of set theoretic estimation. Proc. IEEE 81, 182–208 (1993)
Combettes, P.L.: The convex feasibility problem in image recovery. Adv. Imaging Electron Phys. 95, 155–270 (1996)
Combettes, P.L.: Convex set theoretic image recovery by extrapolated iterations of parallel subgradient projections. IEEE Trans. Image Process. 6(4), 493–506 (1997)
Herman, G.: Image Reconstruction from Projections. The Fundamentals of Computerized Tomography. Academic Press, New York (1980)
Marks, L.D., Sinkler, W., Landree, E.: A feasible set approach to the crystallographic phase problem. Acta Crystallogr. 55(4), 601–612 (1999)
Censor, Y., Lent, A.: A Cyclic Subgradient Projections Method for the Convex Feasibility Problems. Technical Report. University of Haifa, Israel (1980)
Butnariu, D., Censor, Y., Gurfil, P., Hadar, E.: On the behavior of subgradient projections methods for convex feasibility problems in Euclidean spaces. SIAM J. Optim. 19, 786–807 (2008)
Kiwiel, K.C.: The efficiency of subgradient projection. SIAM J. Control Optim. 34, 677–697 (1996)
Kiwiel, K.C.: The efficiency of subgradient projection methods for convex optimization I: General level methods. SIAM J. Control Optim. 34, 660–676 (1996)
dos Santos, L.T.: A parallel subgradient method for the convex feasibility problem. J. Comput. Appl. Math. 18, 307–320 (1987)
Li, C., Mordukhovich, B.S., Wang, J.H., Yao, J.C.: Weak sharp minima on Riemannian manifolds. SIAM J. Optim. 21(4), 1523–1560 (2011)
Rapcsák, T.: Smooth Nonlinear Optimization in \({\mathbb{R}}^n\). Nonconvex Optimization and its Applications, vol. 19. Kluwer, Dordrecht (1997)
Adler, R., Dedieu, J.P., Margulies, J., Martens, M., Shub, M.: Newton’s method on Riemannian manifolds and a geometric model for the human spine. IMA J. Numer. Anal. 22, 359–390 (2002)
Burke, J.V., Lewis, A., Overton, M.: Optimal stability and eigenvalue multiplicity. Found. Comput. Math. 1, 205–225 (2001)
Ferreira, O.P., Pérez, L.R.L., Nemeth, S.Z.: Singularities of monotone vector fields and an extragradient-type algorithm. J. Glob. Optim. 31(1), 133–151 (2005)
Greene, R., Wu, H.: On the subharmonicity and plurisubharmonicity of geodesically convex functions. Indiana Univ. Math. J. 22, 641–653 (1972)
Mahony, R.E.: The constrained Newton method on a Lie group and the symmetric eigenvalue problem. Linear Algebr. Appl. 248, 67–89 (1996)
Miller, S.A., Malick, J.: Newton methods for nonsmooth convex minimization: connections among U-Lagrangian, Riemannian Newton and SQP methods. Math. Program. 104, 609–633 (2005)
Smith, S.T.: Geometric Optimization Methods for Adaptive Filtering. Division of Applied Sciences. PhD Thesis, Harvard University, Cambridge, MA (1993)
Udriste, C.: Convex Functions and Optimization Methods on Riemannian Manifolds. Mathematics and Its Applications, vol. 297. Kluwer, Dordrecht (1994)
Absil, P.A., Baker, C.G., Gallivan, K.A.: Trust-region methods on Riemannian manifolds. Found. Comput. Math. 7(3), 303–330 (2007)
Bridson, M., Haefliger, A.: Metric Spaces of Non-positive Curvature. Springer, Berlin (1999)
Ferreira, O.P., Oliveira, P.R.: Subgradient algorithm on Riemannian manifolds. J. Optim. Theory Appl. 97(1), 93–104 (1998)
Ferreira, O.P., Oliveira, P.R.: Proximal point algorithm on Riemannian Manifold. Optimization 51(2), 257–270 (2002)
Gabay, D.: Minimizing a differentiable function over a differential manifold. J. Optim. Theory Appl. 37(2), 177–219 (1982)
Li, C., López, G., Martín-Márquez, V.: Monotone vector fields and the proximal point algorithm on Hadamard manifolds. J. Lond. Math. Soc. 79, 663–683 (2009)
Li, C., Wang, J.H.: Newton’s method for sections on Riemannian manifolds: generalized covariant \(\alpha \)-theory. J. Complex. 24, 423–451 (2008)
Li, C., Yao, J.C.: Variational inequalities for set-valued vector fields on Riemannian manifolds: convexity of the solution set and the proximal point algorithm. SIAM J. Control Optim. 50(4), 2486–2514 (2012)
Luenberger, D.G.: The gradient projection method along geodesics. Manag. Sci. 18, 620–631 (1972)
Wang, J.H., Huang, S.C., Li, C.: Extended Newton’s algorithm for mappings on Riemannian manifolds with values in a cone. Taiwan J. Math. 13, 633–656 (2009)
Wang, J.H., López, G., Martín-Márquez, V., Li, C.: Monotone and accretive vector fields on Riemannian manifolds. J. Optim. Theory Appl. 146, 691–708 (2010)
Yang, Y.: Globally convergent optimization algorithms on Riemannian manifolds: uniform framework for unconstrained and constrained optimization. J. Optim. Theory Appl. 132(2), 245–265 (2007)
Bento, G.C., Melo, J.G.: Subgradient algorithm for convex feasibility on Riemannian manifolds. J. Optim. Theory Appl. 152(3), 773–785 (2012)
da Cruz Neto, J.X., de Lima, L., Oliverira, P.R.: Geodesic algorithms in Riemannian geometry. Balkan J. Geom. Appl. 3, 89–100 (1998)
Sakai, T.: Riemannian Geometry. Translations of Mathematical Monographs, vol. 149. American Mathematical Society, Providence (1996)
do Carmo, M.P.: Riemannian Geometry. Birkhauser, Boston (1992)
Acknowledgments
Research of the second author is supported in part by the National Natural Science Foundation of China (Grants 11171300, 11371325) and by Zhejiang Provincial Natural Science Foundation of China (Grant LY13A010011). Research of the third author was partially supported by the National Science Council of Taiwan under Grant NSC 99-2115-M-037-002-MY3.
Author information
Authors and Affiliations
Corresponding author
Additional information
Communicated by Sándor Zoltán Németh.
Rights and permissions
About this article
Cite this article
Wang, X.M., Li, C. & Yao, J.C. Subgradient Projection Algorithms for Convex Feasibility on Riemannian Manifolds with Lower Bounded Curvatures. J Optim Theory Appl 164, 202–217 (2015). https://doi.org/10.1007/s10957-014-0568-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10957-014-0568-9
Keywords
- Convex feasibility problem
- Cyclic subgradient projection algorithm
- Riemannian manifold
- Sectional curvature