Abstract
We propose a new numerical method for semi-infinite optimization problems whose objective function is a nonsmooth function. Existing numerical methods for solving semi-infinite programming (SIP) problems make strong assumptions on the structure of the objective function, e.g., differentiability, or are not guaranteed to furnish a feasible point on finite termination. In this paper, we propose a feasible proximal bundle method with convexification for solving this class of problems. The main idea is to derive upper bounding problems for the SIP by replacing the infinite number of constraints with a finite number of convex relaxation constraints, introduce improvement functions for the upper bounding problems, construct cutting plane models of the improvement functions, and reformulate the cutting plane models as quadratic programming problems and solve them. The convex relaxation constraints are constructed with ideas from the α BB method of global optimization. Under mild conditions, we showed that every accumulation point of the iterative sequence is an 𝜖-stationary point of the original SIP problem. Under slightly stronger assumptions, every accumulation point of the iterative sequence is a local solution of the original SIP problem. Preliminary computational results on solving nonconvex, nonsmooth constrained optimization problems and semi-infinite optimization problems are reported to demonstrate the good performance of the new algorithms.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
In this paper, we consider the following nonsmooth, nonconvex optimization problem
where \(f:\mathbb {R}^{n}\rightarrow \mathbb {R}\) is locally Lipschitz but potentially nonsmooth and nonconvex, \(g:\mathbb {R}^{n}\times T\rightarrow {\mathbb {R}}\) is twice continuously differentiable on X × T. In addition, T is a fixed, nonempty compact subset of \(\mathbb {R}^{p}\) and X is a convex compact subset of \(\mathbb {R}^{n}\). This problem is called the semi-infinite programming (SIP) problem. For simplicity of notation, we consider the semi-infinite problem with one semi-infinite constraint and a one-dimensional index set T = [a,b], that is p = 1. However, the algorithm proposed in this paper might be helpful to deal with more general cases.
SIP has recently drawn a lot of attention due to its application in many different fields, including portfolio problem, robot trajectory planning, vibrating membrane problem, minimal cost control of air pollution, and optimization over probability measures (see [11, 15, 22]).
There exist a wide range of numerical methods for SIP problems. Traditional methods for SIP like discretization, exchange, and reduction-based [14, 21, 32, 39, 44] solve a sequence of finitely nonlinear programming subproblems, that is, subproblems with finitely many constraints chosen from the original constraints. In particular, in the reduction-based methods, the finite subset of constraints is generated from finding implicitly all the local maxima t(x) of g(x,⋅) on T for each x ∈ X. Based on this reduction principle, a number of nonsmooth approaches have been presented [10, 24, 31]. These traditional methods suffer from the major drawback that they generally do not provide a feasible point in a finite number of iterations.
A breakthrough for SIP was achieved by Bhattacharjee et al. [4], who replaced the semi-infinite constraint by an overestimation obtained via interval extensions and presented a convergent upper bounding approach to SIP. It is the first algorithm with feasible iterates for SIP, to our knowledge. Based on this upper bounding approach, a number of algorithms for finding the global solution of SIP problems have been presented [5, 8, 27, 29, 30]. Under the assumptions that the functions f and g(⋅,t) are continuously differentiable on X for each t ∈ T, g(x,⋅) is continuous on T for each x ∈ X, there exist SIP Slater points arbitrarily close to all global minima, and the resulting finite nonlinear programs (NLPs) are solved to local optimality; the method using interval extensions with feasible iterates generates an 𝜖-optimal estimate of the global solution of the SIP on finite termination [5]. The main drawback of this method is the rapid increase in the size of the lower and upper bounding problems as nodes of increasing depth are visited by the branch-and-bound procedure. Other methods to generate the global solution of the SIP problems were proposed by Mitsos and coworkers [8, 27, 30]. The upper bounding problems of these methods are not generated by interval extensions, but rather by a restriction of the constraints right-hand side by a negative value (< −𝜖g, for some 𝜖g > 0). For fixed point x, these methods use three subproblems, including the lower bounding problem, the upper bounding problem, and the lower level problem. Under the assumptions that f is continuous on X, g is continuous on X × T, there exists an 𝜖f-optimal SIP Slater point, and the resulting finite NLPs are solved to global optimality, the algorithms terminate finitely with a feasible point and solve the SIP approximately to global optimality. We point out that the authors in [8, 27, 29, 30] consider global solution of all subproblems, which is computationally expensive.
In contrast to these methods, the adaptive convexification method discussed in [9] does not focus on the global solution of the SIP problem, but global solutions of the lower level problem. The authors construct convex relaxations of the lower level problem using the α BB method [1, 2] of global optimization, replace the convex lower level problems with their Karush-Kuhn-Tucker (KKT) conditions, and solve the resulting mathematical programs with complementarity constraints. Under the assumptions that \(f:\mathbb {R}^{n}\rightarrow \mathbb {R}\) and \(g:\mathbb {R}^{n}\times \mathbb {R}^{p}\rightarrow \mathbb {R}^{m}\) are twice continuously differentiable, the algorithm generates a feasible point and reaches a 𝜖-stationary point of the original SIP problem on finite termination. Such method requires slightly stronger assumption on the constraints, but can give tighter bounds than interval extensions. Based on the method discussed in [9], a number of convexification algorithms have been presented [36, 38, 41]. We point out that the method discussed in [41] relies on constructing concave relaxations of the lower level problem and solving the resulting approximate problems with finitely many constraints.
Bundle methods are among the most efficient methods for solving nonsmooth optimization problems. See [18, 20, 25, 35] for a more detailed discussion and comparison of some bundle methods for nonsmooth convex constrained optimization. When the objective function is nonconvex, the recent results [12, 13] make assumptions on the structure of the objective function and propose a redistributed bundle method for unconstrained nonsmooth, nonconvex optimization problems. Based on this redistributed method, many nonconvex bundle approaches have been suggested. The authors [24, 43] propose proximal bundle methods for constrained nonsmooth, nonconvex optimization problems, where the objective and constraint functions are lower-C2. Lv et al. [23] propose an inexact proximal bundle method for constrained nonsmooth, nonconvex optimization problems, where the objective and constraint functions are lower-C1. Hoseini and Nobakhtian [16, 17] propose bundle methods for constrained nonsmooth nonconvex optimization problems, where the objective and constraint functions are regular. The methods presented in [23, 24, 31] can be used to solve the SIP problem, but the calculation of the lower level problem in these methods is costly since the lower level problem needs to be solved within a given fixed accuracy at each point. Moreover on finite termination, these methods are not guaranteed to furnish a feasible point.
In this paper, our aim is to design a feasible algorithm for nonsmooth, nonconvex SIP that combines some of the ideas of the proximal bundle methods and of the convexification methods. The main idea is to generate upper bounding problems for the SIP by replacing the lower level problem with its concave relaxations using the α BB method, and solve the upper bounding problems by a new feasible proximal bundle method. The relaxed lower level problem is a finite nonsmooth problem. Under certain assumptions given in Theorem 6 below, it can be shown that every accumulation point of the iterative sequence is an 𝜖-stationary point of the original SIP problem. Under slightly stronger assumptions given in Theorem 7 below, every accumulation point of the iterative sequence is a local solution of the original SIP problem. These assumptions appeared in nonsmooth, nonconvex bundle methods cited above. Compared to the related works [23, 24, 31], our algorithm starts with a coarse subset of T and is advantageous to save computational time on constraint calculations. Moreover, we adopt an aggregation technique to control the size of subproblems flexibly. In particular, our algorithm can obtain a feasible point in finite iterations. Compared to [9, 36, 38, 41], the assumptions made on the structure of the objective function in this paper are weaker. We point out that the subproblems of our algorithm are solved by existing QP solver and do not depend on the number of the constraints of the upper bounding problem, whereas the subproblems of algorithm from [9, 36, 38, 41] are solved by the existing NLP solver and depend on the number of the constraints of the upper bounding problem. Thus, our algorithm is, naturally, significantly less time-consuming when the number of subintervals/nodes increases. The results of numerical experiment also show the good performance of the new method.
The remainder of this paper is organized as follows. The main concepts used throughout the paper are described in the next section, where we also state the methods to construct upper bounding functions of the constraints. Section 3 presents our feasible proximal bundle algorithm with convexification. Convergence of the proximal bundle method with convexification is addressed in Section 4. Section 5 presents some numerical results and Section 6 gives conclusions.
Our notation is fairly standard. The Euclidean inner product of two vectors \(x,y\in \mathbb {R}^{n}\) is denoted by 〈x,y〉 := xTy, and the associated norm by ∥⋅∥. The positive-part function is denoted by \(x^{+} := \max \limits \{x, 0\}\). For a set \(X\in \mathbb {R}^{n}\), convX denotes its convex hull. The closed ball with center \(x\in \mathbb {R}^{n}\) and radius r > 0 is denoted by B(x;r). That is, \({B}(x; r):=\{y\in \mathbb {R}^{n} | \|y-x\|\leq r\}\).
2 Preliminaries
In this section, we first recall some concepts and results related to nonsmooth analysis. Next, we give the necessary condition of the optimization measure. Then, we introduce the convexification method to construct the concave relaxation of the lower level problem.
2.1 Background
Definition 1
(Lipschitz continuity) A function \(f:\mathbb {R}^{n}\rightarrow \mathbb {R}\) is Lipschitz continuous on a set \(X\subset \mathbb {R}^{n}\) if there exists L > 0 such that
Then, L is called a Lipschitz constant for f on X.
Definition 2
The directional derivative of f at x in the direction \(d\in \mathbb {R}^{n}\) is defined by
Let \(f:\mathbb {R}^{n}\rightarrow \mathbb {R}\) be a locally Lipschitz continuous function at a point \(x\in \mathbb {R}^{n}\). Then, the generalized directional derivative (Clarke) of f at x in the direction of \(d\in \mathbb {R}^{n}\) is defined by
Definition 3
A function \(f:\mathbb {R}^{n}\rightarrow \mathbb {R}\) is called regular at \(x\in \mathbb {R}^{n}\) if it is locally Lipschitz continuous at x and for all \(d\in \mathbb {R}^{n}\) the classical directional derivative \(f^{\prime }(x;d)\) exists and we have \(f^{\prime }(x;d)=f^{\circ }(x;d)\).
Definition 4
A function \(f:\mathbb {R}^{n}\rightarrow \mathbb {R}\) is called semismooth at \(x\in \mathbb {R}^{n}\) if f is Lipschitz on a ball about x and for each \(d\in \mathbb {R}^{n}\) and for any sequences \(\{h_{k}\}\subset \mathbb {R}^{+}\), HCode \(\{\theta _{k}\}\subset \mathbb {R}^{n}\) and \(\{\xi _{k}\}\subset \mathbb {R}^{n}\) such that {hk}↓ 0, HCode \(\{\theta _{k}/ h_{k}\}\rightarrow 0\in \mathbb {R}^{n}\) and ξk ∈ ∂f(x + hkd + 𝜃k), the sequence {〈ξk,d〉} has exactly one accumulation point.
Definition 5
A locally Lipschitz function \(f:O\rightarrow \mathbb {R}\), where O is an open subset of \(\mathbb {R}^{n}\), is called lower-C1 (lower-C2) on O, if on some neighborhood V of each x0 ∈ O there is a representation \(f(x)=\max \limits _{\omega \in {{\varOmega }}}f_{\omega }(x)\) in which the functions fω are of class (twice) continuously differentiable on V and the index set Ω is a compact space such that fω(x) and ∇fω(x) (∇2fω(x)) depend continuously not just on x ∈ V but jointly on (ω,x) ∈Ω× V.
Definition 6
Let \(f:\mathbb {R}^{n}\rightarrow \mathbb {R}\) be a locally Lipschitz continuous function at a point \(x\in \mathbb {R}^{n}\). Then, the subdifferential of f at x is defined by
Each element ξ ∈ ∂f(x) is called a subgradient of f at x. If \(f:\mathbb {R}^{n}\rightarrow \mathbb {R}\) is a convex function, then the subdifferential of f at \(x\in \mathbb {R}^{n}\) is defined by
Theorem 1
[3, Theorem 3.3] Let \(f:\mathbb {R}^{n}\rightarrow \mathbb {R}\) be a locally Lipschitz continuous function at \(x\in \mathbb {R}^{n}\) with a Lipschitz constant L. Then, the subdifferential ∂f(x) is a nonempty, convex, and compact set such that
Theorem 2
[3, Theorem 3.23] Let \(f_{i}:\mathbb {R}^{n}\rightarrow \mathbb {R}\) be locally Lipschitz continuous at x for all i = 1,...,m. Then, the function
is locally Lipschitz continuous at x and
where
In addition, if fi is regular at x for all i = 1,...,m, then f is also regular at x and equality holds in (2).
Now, we introduce the indicator function IX of a set \(X\subset \mathbb {R}^{n}\) as follows:
The indicator function IX of a set \(X\subset \mathbb {R}^{n}\) is convex if and only if X is convex. Let the point x belong to X, then
From [26, Proposition 4] and [3, Theorem 3.13], we have the following:
Theorem 3
If \(f:X\rightarrow \mathbb {R}\) is continuously differentiable then f is semismooth and regular on X.
From [37, Proposition 2.4] and [7, Theorem 2], we have the following:
Proposition 1
Given an open set O containing X. The following statements are equivalent:
- 1.:
-
f is semismooth and regular on O.
- 2.:
-
For all x ∈ O, for all 𝜖 > 0, there exists ρ > 0 such that for all y ∈ B(x;ρ) and ξ ∈ ∂f(y)
$$ f(z)-f(y)\geq\langle\xi ,z-y\rangle-\epsilon\|z-y\|,\qquad z\in B(x;\rho). $$ - 3.:
-
f is lower-C1 on O.
Now we return to our basic problem (1), and give the following assumption.
Assumption 1
The Slater constraint qualification holds for problem (1), i.e., there exists \(\hat {x}\in \mathbb {R}^{n}\) such that \(g(\hat {x},t)<0\) for all t ∈ T.
Assumption 1 was used in [41]. In this paper, we always assume that the Slater constraint qualification holds.
2.2 Necessary conditions
To solve problem (1), we introduce the improvement function
where
Now, the subdifferential of G at a point x is defined by [6]
where A(x) = {t ∈ T|G(x) = g(x,t)}.
Definition 7
[42](EMFCQ) A feasible point x is said to satisfy the extended Mangasarian-Fromovitz constraint qualification for SIP problem (1) if there exists a feasible direction d of X at x such that
where Tact(x) = {t ∈ T|g(x,t) = 0}.
Theorem 4
Let x∗ be a local minimizer of (1). Then the following statements hold:
- (a):
-
0 ∈ ∂H(x∗,x∗) + ∂IX(x∗).
- (b):
-
There exist nonnegative multipliers λi, i = 0,...,n + 1, and indices ti ∈ A(x∗),i = 1,...,n + 1 such that \({{\sum }_{i=0}^{n+1}} \lambda _{i}=1\), λig(x∗,ti) = 0,i = 1,...,n + 1 and
$$ 0 \in\lambda_{0}\partial f(x^{*})+\sum\limits_{i=1}^{n+1}\lambda_{i}\nabla_{x} g(x^{*},t_{i})+\partial I_{X}(x^{*}). $$ - (c):
-
If EMFCQ holds at x∗ for problem (1), then there exist nonnegative multipliers λi and indices ti ∈ Tact(x∗),i = 1,...,n + 1 such that
$$ 0 \in \partial f(x^{*})+\sum\limits_{i=1}^{n+1}\lambda_{i}\nabla_{x} g(x^{*},t_{i})+\partial I_{X}(x^{*}). $$
Proof
-
(a)
Let F be the feasible region of the problem (1), namely, F = {x ∈ X|G(x) ≤ 0}. Since \(x^{*}\in \mathbb {R}^{n}\) is a local optimum of the problem (1), then G(x∗) ≤ 0 and there exists a neighborhood U of x∗ such that f(x∗) ≤ f(x) for all x ∈ U ∩ F. Then, for all \(x\in U\cap F\subseteq X\)
$$ H(x,x^{*})=\max\{f(x)-f(x^{*}) , G(x)\}\geq f(x)-f(x^{*})\geq 0=H(x^{*},x^{*}). $$This means that x∗ is a local minimum of H(⋅,x∗) on X, which implies 0 ∈ ∂H(x∗,x∗) + ∂IX(x∗).
-
(b)
If G(x∗) < 0, we have \(\partial H(x^{*},x^{*})\subseteq \partial f(x^{*})\). Then, the assertion of the theorem is proved by choosing λ0 = 1 and λi = 0 for i = 1,...,n + 1. If G(x∗) = 0, we have \(\partial H(x^{*},x^{*})\subseteq \text {conv}\{\partial f(x^{*})\cup \partial G(x^{*})\}\), where ∂G(x∗) = conv{∇g(x∗,ti),ti ∈ A(x∗)}. Then due to the definition of convex hull and the Caratheodory’s Theorem, there exist λi ≥ 0, \(i=0,1,\dots ,n+1\), such that
$$ 0 \in\lambda_{0}\partial f(x^{*})+\sum\limits_{i=1}^{n+1}\lambda_{i}\nabla_{x} g(x^{*},t_{i})+\partial I_{X}(x^{*}). $$ -
(c)
Suppose that λ0 = 0, then we deduce that G(x∗) = 0 and A(x∗) = Tact(x∗). Moreover, \(-{\sum }_{i=1}^{n+1}\lambda _{i}\nabla _{x} g(x^{*},t_{i})\in \partial I_{X}(x^{*})\). For any feasible direction d of X at x∗, we have
$$ \left\langle-\sum\limits_{i=1}^{n+1}\lambda_{i}\nabla_{x} g(x^{*},t_{i}),d\right\rangle\leq 0, \ \ t_{i}\in T_{act}(x^{*}), $$which is an contradiction with (5). Thus, λ0 is strictly positive, and the result item (c) holds by setting λi = λi/λ0.
□
Definition 8
x∗ is called an 𝜖-KKT point of (1) if there exist some multipliers λi, i = 1,...,n + 1 such that
where \(T_{act}^{\epsilon }(x^{*})=\{t\in T | g(x^{*},t)\in [-\epsilon ,0]\}\).
2.3 The concave relaxation of the lower level problem
In this subsection, we consider the following lower level problem
We note that the lower level problem is actually a concave optimization problem for all x ∈ X if g(x,⋅) is convex on T for these x ∈ X.
For \(N\in \mathbb {N}\), let a = t0 < t1 < ... < tN = b define a subdivision E of T:
Then, we have
where U = {1,2,...,N} and Tu = [tu− 1,tu]. Tu is called a subinterval of T and satisfies
The length of the subinterval Tu is defined by |Tu| = |tu − tu− 1| and the length of the subdivision E is defined by \(|E|=\max \limits _{u\in U}|T_{u}|\).
Following the approach of [41], for all u ∈ U, we generate the upper bound function gu of g on Tu,
In this case, the function gu is twice continuously differentiable with respect to t and the second-order derivative with respect to t of gu is
It follows that gu(x,t) is convex with respect to t on Tu if the parameter αu satisfies
On the other hand, we require to construct an upper bound function of g. Thus,
Lemma 1
[41] Let f be a convex function and C = convS, where S is an arbitrary set of points. Then
where the first supremum is attained only when the second supremum is attained.
We know from Lemma 1 that the maximum of gu(x,t)(∀ u ∈ U) must be attained on the boundary of Tu, that is,
Then, we have
The convex upper bound function G(⋅,α,E) of G on X is defined by
It can be rewritten as
where
where the parameter αu satisfies (6). We define
and
Then, we can obtain that
Now the upper bounding problem of the original problem is defined by
Using the improvement function mentioned in the preceding section, we define \(H(\cdot ,x,\alpha ,E):\mathbb {R}^{n}\rightarrow \mathbb {R}\) as follows:
where α and E vary along the iterations.
Lemma 2
[41] If EMFCQ holds for all feasible points of the original problem (1), then the MFCQ holds for all feasible points of the approximation problem (12) when |E| is sufficiently small.
Theorem 5
Suppose that \(f: X\rightarrow \mathbb {R}\) and \(g: X\times T\rightarrow \mathbb {R}\) are locally Lipschitz continuous, then the following statements hold:
- 1.:
-
If x∗ is a local minimizer of (12), then 0 ∈ ∂H(x∗,x∗,α,E) + ∂IX(x∗).
- 2.:
-
If 0 ∈ ∂H(x∗,x∗,α,E) + ∂IX(x∗) and G(x∗,α,E) ≤ 0, then there exist nonnegative multipliers λi, i = 0,...,n + 1, and indices \(t_{i}\in \bar {A}(x^{*},\alpha ,E), i=1,...,n+1\), such that \({{\sum }_{i=0}^{n+1}} \lambda _{i}=1\), λi(g(x∗,ti) + h(ti,αi)) = 0,i = 1,...,n + 1, and
$$ 0 \in\lambda_{0}\partial f(x^{*})+\sum\limits_{i=1}^{n+1}\lambda_{i}\nabla_{x} g(x^{*},t_{i})+\partial I_{X}(x^{*}). $$If, in addition, MFCQ holds at x∗ for problem (12), then there exist nonnegative multipliers λi and indices \(t_{i}\in \bar {T}_{act}(x^{*},\alpha ,E), i=1,...,n+1\), such that
$$ 0 \in \partial f(x^{*})+\sum\limits_{i=1}^{n+1}\lambda_{i}\nabla_{x} g(x^{*},t_{i})+\partial I_{X}(x^{*}), $$where \(\bar {T}_{act}(x^{*},\alpha ,E)=\{t_{u}\in E | g(x^{*},t_{u})+h(t_{u},\alpha _{u})=0\}.\) If, in addition, D(E,α) ≤ 𝜖, then x∗ is an 𝜖-KKT point of the problem (1).
Proof
The proof of this theorem is analogous to Theorem 4; we only require to show that x∗ is an 𝜖-KKT point of the original problem if D(E,α) ≤ 𝜖. It suffices to prove that if \(t_{i}\in \bar {T}_{act}(x^{*},\alpha ,E)\), i = 1,...,n + 1, then \(t_{i}\in T_{act}^{\epsilon }(x^{*})\), i = 1,...,n + 1, provided that condition D(E,α) ≤ 𝜖 holds. For each \(t_{i}\in \bar {T}_{act}(x^{*},\alpha ,E)\), we have
where
where T0 = 0 and TN+ 1 = 0. It implies that g(x∗,ti) ≤ 0 and
It follows that \(t_{i}\in T_{act}^{\epsilon }(x^{*})\), i = 1,...,n + 1, if the condition D(E,α) ≤ 𝜖 holds. □
3 The numerical approach
Let k be the current iteration index and xk be the stability center (or serious iterate) at iteration k, where iterations generate a sequence of trial points \(\{y^{j}\}_{j\in J_{k}^{ora}}\), with \(J_{k}^{ora}\subseteq \{0,1,...,k\}\). In particular, \(x^{k}\in \{y^{j}\}_{j\in J_{k}^{ora}}\). Let \(E^{k}=\{t_{u} | u\in U^{\prime }_{k}=\{0,1,...,N_{k}\}\}\) be the current subdivision of T and αk be the corresponding convexification parameter.
Following the redistributed proximal approach [12, 13], we generate the augmented functions of f and G(⋅,αk,Ek) defined by
The idea is to utilize the augmented functions in the model construction. We can first construct a convex piecewise linear approximation defined by
where \({\xi ^{j}_{f}}\in \partial f(y^{j})\) and \(\xi ^{k,j}_{g}\in \partial G(y^{j},\alpha ^{k},E^{k})\). It can be rewritten in an equivalent form
where
with
Any choice for the parameters \({\eta _{1}^{k}}\), HCode \({\eta _{2}^{k}}\) that keeps \(c_{k,j}^{f},c_{k,j}^{g}\) in (13) nonnegative is acceptable. In our method, we take
where γ is a small positive parameter. In particular, since \(x^{k}\in \{y^{j}\}_{j\in J_{k}^{ora}}\), we obtain that
Then, we seek for the new point yk+ 1 as a solution of
where μk > 0 is a proximal parameter.
The following lemma introduces the characterization of the solution of the (18).
Lemma 3
Let yk+ 1 be the unique solution to (18) and assume μk > 0. Then, we have that
Lemma 3 implies that
The linearization error of \(\hat {H}_{k}(\cdot )\) is defined by
Since the piecewise linear model \(\hat {H}_{k}\) is convex, we can construct a convex linear approximation of \(\hat {H}_{k}\) by
Hence, the cutting plane model can be redefined in the form
where \(J_{k}^{ora}\subseteq \{0,1,...,k\}\) and \(J_{k}^{agg}\subseteq \{1,...,k\}\). Then, (18) can be stated as a quadratic programming problem in \(\mathbb {R}^{1}\times X\):
From the KKT conditions for (23), one easily obtains a representation for Sk+ 1 and Ck+ 1 (defined by (19) and (21)):
and (λ1,k+ 1,λ2,k+ 1,λ3,k+ 1) is a solution to
To measure the quality of the candidate, the nominal decrease δk+ 1 is defined by:
According to (20), (21), and Lemma 3, we can obtain that
We now present the criterion that determines whether a new stability center is generated:
where m is a positive descent parameter. If (28) holds, the trial point yk+ 1 brings a sufficient descent for the objective or constraint function. Then, the stability center will be moved to yk+ 1, i.e., xk+ 1 := yk+ 1. This case is called a serious step (or descent step). Otherwise, the trial point yk+ 1 is used to enrich the model and the stability center is not changed. This case is called a null step.
In the following, we introduce a subdivision strategy to refine the current subdivision and construct a refined approximation of the lower level problem by the following procedure.
Refinement procedure
-
1.
For a given subset Tu = [tu− 1,tu], let αu be the corresponding convexification parameter.
-
2.
Define
$$ E^{\alpha BB}(T_{{u}})=\left\{t_{{u},i} | t_{{u},i}=t_{{u}-1}+\frac{i}{2s+1}(t_{{u}}-t_{{u}-1}),i=1,...,2s, s\in\mathbb{N}\right\}. $$(29) -
3.
Put Tu,i = [tu,i− 1,tu,i], i = 1,..., 2s + 1, where tu,0 = tu− 1 and tu,2s+ 1 = tu.
-
4.
Determine the corresponding convexification parameters αu,i, i = 1,..., 2s + 1, such that αu,i ≤ αu and
$$ \alpha_{{u},i}\geq\max\left\{0,\underset{(x,t)\in X\times T_{{u},i}}{\max} \left\{-{\nabla_{t}^{2}} g(x,t)\right\}\right\}. $$(30)
Remark 1
Note that the computation of αu,i involves a global optimization problem. However, we can use an upper bound for the right-hand side in (30). In particular, we can choose αu,i = αu. Therefore, the computation of the globally optimal value of the lower level problem \(\max \limits _{t\in T}\{g(x,t)\}\) uses more computation efforts than the computation of αu,i.
Now, we show that our refinement procedure brings a sufficient descent for the new constraint function.
Lemma 4
Assume that Tu,i = [tu,i− 1,tu,i] and αu,i, i = 1,..., 2s + 1, are obtained by taking the refinement procedure on subset Tu = [tu− 1,tu]. Then, we have
where
Proof
The first inequality can be obtained directly by the definitions of g1,i,i = 1,..., 2s + 1. We set \(\psi _{i}(t)=\frac {\alpha _{u,i}}{2}(t-\frac {t_{u,i-1}+t_{u,i}}{2})^{2}-\frac {\alpha _{u}}{2}(t-\frac {t_{u-1}+t_{u}}{2})^{2}\). Then, we only need to prove that ψi(t) ≤ 0 for any i = 1,..., 2s + 1. We note that \(|t-\frac {t_{u,i-1}+t_{u,i}}{2}|\leq \frac {t_{u}-t_{u-1}}{2(2s+1)}\leq \frac {t_{u-1}+t_{u}}{2}-t\) for all t ∈ [tu,i− 1,tu,i],i = 1,...,s. Then, for any i ≤ s, we have
Therefore, we conclude that ψi(t) is nondecreasing on [tu,i− 1,tu,i],i = 1,...,s. It implies that ψi(t) ≤ ψi(tu,i) for all t ∈ [tu,i− 1,tu,i],i = 1,...,s. For any i ≤ s, we have
Then, we have ψi(t) ≤ 0 for any [tu,i− 1,tu,i],i = 1,...,s. Similarly, we can prove that ψi(t) ≤ 0 for any [tu,i− 1,tu,i],i = s + 2,..., 2s + 1. For i = s + 1, we have
The proof is complete. □
Now, we present the algorithm in full detail.
Algorithm 1
(Feasible proximal bundle method with convexification for SIP problems) Data. m ∈ (0, 1), tol1 ≥ 0, tol2 > 0, tol3 ≥ 0, \(\mu _{\max \limits }>\mu _{\min \limits }>0\), \(\mu _{0}\in [\mu _{\min \limits },\mu _{\max \limits }]\), \({\eta _{1}^{0}}\geq 0\), \({\eta _{2}^{0}}\geq 0\), γ > 0, s > 0, \(i_{n\max \limits }>0\), \(\tilde {X}\subset \mathbb {R}\). Set k = 0,N0 = 2s + 1, and
Determine the parameter α0 with (6) on E0.
Step 0 (Initialization)
Solve the problem
and from the solution \(\tilde {x}\) compute \(G(\tilde {x},\alpha ^{0},E^{0})\). Set \(y^{0}=\tilde {x}\), \(G(y^{0},\alpha ^{0},E^{0})=G(\tilde {x},\alpha ^{0},E^{0})\). If z > 0, set iz = 1, \(J_{1}^{ora}=\{0\}\), and go to Step 7. Otherwise, compute \(f(y^{0}),\ {\xi _{f}^{0}}\in \partial f(y^{0}),\ \xi _{g}^{0,0}\in \partial G(y^{0},\alpha ^{0},E^{0})\) and set x0 = y0, f(x0) = f(y0), G(x0,α0,E0) = G(y0,α0,E0). Define the associated \(e_{0,0}^{f}, e_{0,0}^{g}, {b^{0}_{0}}, {{{\varDelta }}_{0}^{0}}\) by (15). Set iz = 0, ix = 0, in = 0, \(J_{0}^{ora}=\{0\}\), \(J_{0}^{agg}=\emptyset \) and go to Step 1.
Step 1 (Quadratic programming subproblem)
Given the model \(\hat {H}_{k}\) defined by (14) and (22), compute Sk+ 1, Ck+ 1, yk+ 1, (λ1,k+ 1,λ2,k+ 1,λ3,k+ 1) by solving the subproblem (23) and its dual (25).
Step 2 (Stopping test)
Compute δk+ 1 by (26). If δk+ 1 ≤ tol1 and D(Ek,αk) ≤ tol2, STOP.
Step 3 (Bundle management)
Set
Step 4 (Descent test)
Compute
Define the associated \(e_{k,{k+1}}^{f}\), \(e_{k,{k+1}}^{g}\), \(b_{k}^{k+1}\), and \({{\varDelta }}_{k}^{k+1}\) by (15). If the descent condition (28) holds, then set in = 0 and go to Step 6. Otherwise, set in = in + 1 and go to Step 5.
Step 5 (Null step update)
Choose \(\mu _{k+1}\in [\mu _{k},\mu _{\max \limits }]\) and set
If \(i_{n}>i_{n\max \limits }\) and D(Ek,αk) > tol2, then go to Step 7. Otherwise, set
and go to Step 8.
Step 6 (Serious step update)
Choose \(\mu _{k+1}\in [\mu _{\min \limits },\mu _{\max \limits }]\) and set
For all \(j\in J_{l+1}^{ora}\), we set
If D(Ek,αk) > tol2, go to Step 7. Otherwise, set
For all \(j\in J_{k+1}^{ora}\), we set \(\xi ^{k+1,j}_{g}=\xi ^{k,j}_{g}\) and
Then, go to Step 8.
Step 7 (Subdivision update)
Select the index set Wk as follows:
and
Take the Refinement procedure on subsets \(T_{\bar {u}}\), u ∈ Wk, where \(\bar {u}\) is defined by
where T0 = 0 and \(T_{N_{k}+1}=0\). Put
If iz = 1, set E0 = Ek+ 1, α0 = αk+ 1, k = 0 and go to Step 0. Otherwise, for all \(j\in J_{k+1}^{ora}\), compute \(G(y^{j},\alpha ^{k+1},E^{k+1}),\ \xi _{g}^{k+1,j}\in \partial G(y^{j},\alpha ^{k+1},E^{k+1})\) and define the associated \(e_{k+1,j}^{g}\) by (15). Then, go to Step 8.
Step 8 (Parameter update)
Compute the parameters \(\eta _{1}^{k+1}\) and \(\eta _{2}^{k+1}\) by (16). Set k = k + 1 and go to Step 1.
Remark 2
Under Assumption 1, there exist E,α, and a point x such that
Therefore, Algorithm 1 cannot pass infinitely many times through step 0, i.e., Algorithm 1 generates a feasible initial point x0 such that G(x0,α0,E0) ≤ z ≤ 0 at step 0. According to (28), if the descent test is satisfied, then it holds that G(xk+ 1,αk,Ek) ≤ G+(xk,αk,Ek) − mδk+ 1 ≤ G+(xk,αk,Ek). By Lemma 4, we have G(xk+ 1,αk+ 1,Ek+ 1) ≤ G+(xk,αk,Ek). If xk is infeasible, then we can obtain that G(x0,α0,E0) > 0, which contradicts the fact that G(x0,α0,E0) ≤ z ≤ 0. Therefore, if a serious step is declared, then the following conditions hold:
Lemma 5
Algorithm 1 cannot pass infinitely many times through step 7, i.e., there exist \(k_{E}, \bar {E}, \bar {\alpha }\) such that
4 Convergence
Firstly, we define the index set of serious step as follows:
We start with the following lemma.
Lemma 6
Assume that f is bounded below on X and Algorithm 1 generates an infinite number of serious steps. Then \(C^{k}\rightarrow 0\) and \(S^{k}+v^{k}\rightarrow 0\) as \(k\rightarrow \infty \) in \(\mathcal {K}_{s}\).
Proof
We first show that
Making use of (28) and Remark 2, we conclude that, for all \(k\in \mathcal {K}_{s}\),
Thus, the sequence {f(xk)} is nonincreasing and bounded below on X. Let \(\bar {f}=\inf \{f(x) | x\in X\}\). Therefore, we obtain that
Now, we obtain that \(\{\delta _{k}\}_{k\in \mathcal {K}_{s}}\rightarrow 0\). Making use of (26), we conclude that
It immediately follows that \(C^{k}\rightarrow 0\) and \(S^{k}+v^{k}\rightarrow 0\) as \(k\rightarrow \infty \) in \(\mathcal {K}_{s}\) (a consequence of \(\mu _{k-1}\in [\mu _{\min \limits },\mu _{\max \limits }]\)). □
Let Ψk(yk+ 1) denote the objective function optimal value of the subproblem (18), i.e., \({{\varPsi }}_{k}(y^{k+1})=\hat {H}_{k}(y^{k+1})+\frac {\mu _{k}}{2}\|y^{k+1}-x^{k}\|^{2}\).
Lemma 7
Assume that Algorithm 1 takes a finite number of serious steps and \(\bar {k}\) is the last serious iteration, i.e., for all \(k\geq \bar {k}\) we have \(x^{k}=x^{\bar {k}}\). Then we can obtain that
Proof
In what follows, we consider \(k>\max \limits \{\bar {k},k_{E}\}\). Using Lemma 3 and the convexity of \(\hat {H}_{k}\), we obtain that
where the first equality is by (17), the second inequality is due to (20). Since \(k\in J_{k}^{agg}\), (21) and (22) imply that
Moreover,
where the second inequality is by the fact that 〈vk,yk+ 1 − yk〉≤ 0. □
Lemma 8
Assume that Algorithm 1 takes a finite number of serious steps. Suppose that \(\{{\eta ^{k}_{1}}\}\) and \(\{{\eta _{2}^{k}}\}\) are bounded above. Then, \(C^{k}\rightarrow 0\) and \(S^{k}+v^{k}\rightarrow 0\) as \(k\rightarrow \infty \).
Proof
Let \(\bar {k}\) be the last serious iteration. In what follows, we consider \(k>\{\bar {k},k_{E}\}\). By the definition (22) of \(\hat {H}_{k}\) and the fact \(k\in J_{k}^{ora}\), we obtain that
It can be rewritten as
Since {yk}∈ X, the functions f and g are Lipschitz continuous on X and X × T respectively and X is compact, Theorem 1 implies that \(\{{\xi _{f}^{k}}\}\) and \(\{\xi ^{k,k}_{g}\}\) are bounded on that set. Therefore by the boundedness of \(\{{\eta _{1}^{k}}\}\) and \(\{{\eta _{2}^{k}}\}\), we get \(\{{\xi _{f}^{k}}+{\eta _{1}^{k}}(y^{k}-x^{k})\}\) and \(\{\xi ^{k,k}_{g}+{\eta _{2}^{k}}(y^{k}-x^{k})\}\) are bounded. We set
Then, we can obtain
According to (27) and (28), we can obtain that
where the third inequality holds by \(k>\bar {k}\). Combining this relation with (26) yields
According to Lemma 7, we infer that
which implies \(\lim _{k\rightarrow \infty }\delta _{k}=0\) [33, Lemma 3, p. 45]. Making use of (26), we conclude that
It immediately follows that \(C^{k}\rightarrow 0\) and \(S^{k}+v^{k}\rightarrow 0\) as \(k\rightarrow \infty \). □
Let
Lemma 9
Let xk(k ≥ kE) be the current stability center. Suppose that \(\{{\eta ^{k}_{1}}\}\) and \(\{{\eta _{2}^{k}}\}\) are bounded above, then there exist \(\hat {\lambda }_{1,k+1},\hat {\lambda }_{2,k+1}\in \mathbb {R}^{k+1}\) and \(\hat {J}_{k}\) with
such that
and
where \(\hat {\eta }_{i}^{j}\in [\eta _{\min \limits }^{i,j},\eta _{\max \limits }^{i}],~i\in \{1,2\}\).
Proof
In what follows, we consider k ≥ kE. If \(k\in \mathcal {K}_{s}\) or k = kE, then we have \(J^{agg}_{k}=\emptyset \) and
where \(\lambda _{1,k+1}^{j}, \lambda _{2,k+1}^{j}\geq 0\) for all \(j\in J^{ora}_{k}\) and \({\sum }_{j\in J_{k}^{ora}}(\lambda _{1,k+1}^{j} +\lambda _{2,k+1}^{j})=1\). Then, the result holds by setting \(\hat {J}_{k}=J_{k}^{ora}\setminus {J^{C}_{k}}\), \(\hat {\lambda }_{1,k+1}^{j}=\lambda _{1,k+1}^{j}\), \(\hat {\lambda }_{2,k+1}^{j}=\lambda _{2,k+1}^{j}\), \(\hat {\eta }_{1}^{j}={\eta _{1}^{k}}\) and \(\hat {\eta }_{2}^{j}={\eta _{2}^{k}}\).
If \(k\notin \mathcal {K}_{s}\) and k > kE, then there exists \(\tilde {k}\in \{k_{E}+1,...,k-1\}\) such that \(x^{k}=x^{\tilde {k}}\) and \(J^{agg}_{\tilde {k}}=\emptyset \). It holds that \({{\varDelta }}^{j}_{\tilde {k}}={{{\varDelta }}^{j}_{k}}\), \(b^{j}_{\tilde {k}}={b^{j}_{k}}\), \(\xi _{g}^{\tilde {k},j}=\xi _{g}^{k,j}\) for all \(j\in J_{k}^{ora}\) and \(G(x^{\tilde {k}},\bar {\alpha },\bar {E})=G(x^{k},\bar {\alpha },\bar {E})\). Hence,
where \({\sum }_{j\in J_{\tilde {k}}^{ora}}\left (\lambda _{1,\tilde {k}+1}^{j} +\lambda _{2,\tilde {k}+1}^{j} \right )=1\). Since \(J^{agg}_{\tilde {k}}=\emptyset \), then \(J^{agg}_{\tilde {k}+1}=\tilde {k}+1\). Hence,
and
It is easily seen that
We set
For any \(j\in \hat {J}_{\tilde {k}+1}\) and i ∈{1, 2}, we set
where
Combining these, we have
and
By the same argument, we can obtain that there exist \(\hat {\lambda }_{1,k+1},\hat {\lambda }_{2,k+1}\in \mathbb {R}^{k+1}\) and \(\hat {J}_{k}\) satisfying (34) such that (35) and (36) hold. □
From Theorems 2 and 3, we can obtain that G(⋅,αk,Ek) is regular on X. The authors [17] state that the following lemma holds. We give the proof since the cutting plane model and the bundle update are different from [17]. In addition, the proof of [17] does not consider the aggregate information of the cutting plane model.
Lemma 10
Assume that f is regular on X. Suppose that \(\{{\eta ^{k}_{1}}\}\) and \(\{{\eta _{2}^{k}}\}\) are bounded above. Let xk (k ≥ kE) be the current stability center and \(\bar {\eta }=\max \limits _{j\in \{0,1,...,k\}}\{{\eta _{1}^{j}},{\eta _{2}^{j}}\}\). The upper envelope model is defined as
where
B(xk; R) is a fixed ball large enough to contain all possible trial steps during the iteration k. Then, the following statements hold.
-
1.
Φ(⋅,xk,αk,Ek) is a convex function and \(\hat {H}_{k}(\cdot )\leq {{\varPhi }}(\cdot ,x^{k},\alpha ^{k},E^{k})\).
-
2.
Φ(xk,xk,αk,Ek) = G+(xk,αk,Ek).
-
3.
\(\partial {{\varPhi }}(x^{k},x^{k},\alpha ^{k},E^{k})\subseteq \partial H(x^{k},x^{k},\alpha ^{k},E^{k})\).
Proof
We only show that \(\hat {H}_{k}(\cdot )\leq {{\varPhi }}(\cdot ,x^{k},\alpha ^{k},E^{k})\). For the remainder of the proofs, we refer the reader to [17].
Using (14) together with (15), for \(j\in J_{k}^{ora}\) we obtain that
On the other hand, from Lemma 9, we can obtain that for \(j\in J_{k}^{agg}\) there exist \(\hat {\lambda }_{1,j},\hat {\lambda }_{2,j}\in \mathbb {R}^{j}\), and \(\hat {J}_{j-1}\) satisfying (34) such that
where \(\hat {\eta }_{i}^{l}\geq {\eta }_{\min \limits }^{i,l},\ i=1,2\). According to (16) and the definition of \({\eta }_{\min \limits }^{i,l}\), we can obtain that
Hence,
By the definition (22) of \(\hat {H}_{k}(\cdot )\), we can obtain that \(\hat {H}_{k}(\cdot )\leq {{\varPhi }}(\cdot ,x^{k},\alpha ^{k},E^{k})\). □
Theorem 6
Assume that f is bounded below and regular on X. Suppose that \(\{{\eta ^{k}_{1}}\}\) and \(\{{\eta _{2}^{k}}\}\) are bounded above. Consider the stopping parameter tol1 = 0 and suppose there is no termination. Then, the following mutually exclusive situations hold:
- (i):
-
Algorithm 1 generates an infinite sequence of serious steps. Let \(\bar {x}\) be the accumulation point of {xk}, then \(0\in \partial H(\bar {x},\bar {x},\bar {\alpha },\bar {E})+\partial I_{X}(\bar {x})\).
- (ii):
-
Algorithm 1 generates a finite sequence of serious steps. Let \(\bar {x}\) be the last stability center, then \(0\in \partial H(\bar {x},\bar {x},\bar {\alpha },\bar {E})+\partial I_{X}(\bar {x})\).
In addition, if MFCQ holds at \(\bar {x}\) for problem (12) with \(E=\bar {E}\), then \(\bar {x}\) is a tol2-KKT point of the problem (1).
Proof
In what follows, we consider k ≥ kE. From Theorem 1 and Lemma 9, we can obtain that {Sk} is bounded; thus, we can assume that \(S^{k+1}\rightarrow \bar {S}\). Hence, \(v^{k+1}\rightarrow -\bar {S}\) as \(k\rightarrow \infty \). Using (21) together with Lemma 10, we can obtain that
By (19) and the convexity of IX, IX(y) ≥ IX(yk+ 1) + 〈vk+ 1,y − yk+ 1〉. As a result, it holds that
-
(i)
First, note that {xk} is in the compact set X, so it has an accumulation point, say for some infinite set \(K\subseteq \mathcal {K}_{s}\), \(x^{k}\rightarrow \bar {x}\) as \(K\ni k\rightarrow \infty \). For all k + 1 ∈ K, we have \(x^{k}-y^{k+1}=\frac {1}{\mu _{k}}(S^{k+1}+v^{k+1})\). By Lemma 6, \(C^{k+1}\rightarrow 0\), HCode \(S^{k+1}+v^{k+1}\rightarrow 0\), HCode \(x^{k}-y^{k+1}\rightarrow 0\), \(x^{k}\rightarrow \bar {x}\) as \(K\ni k+1\rightarrow \infty \). Passing onto the limit in (38) as \(K\ni k+1\rightarrow \infty \), we obtain that
$$ {{\varPhi}}(y,\bar{x},\bar{\alpha},\bar{E})+I_{X}(y)\geq G^{+}(\bar{x},\bar{\alpha},\bar{E})+\langle 0,y-\bar{x}\rangle={{\varPhi}}(\bar{x},\bar{x},\bar{\alpha},\bar{E})+\langle 0,y-\bar{x}\rangle. $$Using the convexity of \({{\varPhi }}(\cdot ,\bar {x},\bar {\alpha },\bar {E})+I_{X}(\cdot )\), we have \(0\in \partial {{\varPhi }}(\bar {x},\bar {x},\bar {\alpha },\bar {E})+\partial I_{X}(\bar {x}).\) According to Lemma 10, we can conclude that \(0\in \partial H(\bar {x},\bar {x},\bar {\alpha },\bar {E})+\partial I_{X}(\bar {x}).\)
-
(ii)
Let \(\bar {k}\) be the last serious iteration, i.e., \(x^{k}=\bar {x}\) for all \(k\geq \bar {k}\). For \(k\geq \bar {k}\), we have that \(\bar {x}-y^{k+1}=x^{k}-y^{k+1}=\frac {1}{\mu _{k}}(S^{k+1}+v^{k+1})\). By Lemma 8, \(C^{k+1}\rightarrow 0\), HCode \(S^{k+1}+v^{k+1}\rightarrow 0\), HCode \(y^{k+1}-x^{k}\rightarrow 0\) as \(k\rightarrow \infty \). Passing onto the limit in (38) as \(k\rightarrow \infty \), we obtain that
$$ {{\varPhi}}(y,\bar{x},\bar{\alpha},\bar{E})+I_{X}(y)\geq G^{+}(\bar{x},\bar{\alpha},\bar{E})+\langle 0,y-\bar{x}\rangle={{\varPhi}}(\bar{x},\bar{x},\bar{\alpha},\bar{E})+\langle 0,y-\bar{x}\rangle. $$By the same discussion, we deduce that \(0\in \partial H(\bar {x},\bar {x},\bar {\alpha },\bar {E})+\partial I_{X}(\bar {x})\).
From Theorem 5 and Remark 2, we obtain that \(\bar {x}\) is a tol2-KKT point of the original problem (1). The proof is complete.
□
Theorem 7
Assume that f is bounded below and regular on X. Suppose that \(\{{\eta ^{k}_{1}}\}\) and \(\{{\eta _{2}^{k}}\}\) are bounded above. Consider the stopping parameter tol1 = 0 and suppose that algorithm loops forever. If f is semismooth, then for any accumulation point \(\bar {x}\) of the sequence {xk}, for each 𝜖 > 0 there exists ρ > 0 such that
If, in addition, the set \(X_{\rho ,\epsilon ,tol_{2}}=\{y\in X\cap B(\bar {x};\rho ): G(y)<-(\rho \epsilon +tol_{2})\}\) is not empty, then
Proof
Let \(\hat {\lambda }^{j}=\hat {\lambda }_{1,k+1}^{j}+\hat {\lambda }_{2,k+1}^{j}\), then we can obtain that
Equation (39) together with Lemma 6 and 8 implies that \(\|y^{j}-x^{k}\|\rightarrow 0\) for all \(j\in \hat {J}_{k}\). Moreover, \(\|y^{j}-\bar {x}\|\leq \|y^{j}-x^{k}\|+\|x^{k}-\bar {x}\|\rightarrow 0\) for all \(j\in \hat {J}_{k}\). From Proposition 1 and Theorem 3, we can obtain that to fix any 𝜖 > 0 there exist ρ > 0, \(\hat {\lambda }_{1,k+1},\hat {\lambda }_{2,k+1}\in \mathbb {R}^{k+1}\) and \(\hat {J}_{k}\) satisfying (34) such that \(y^{j}\in B(\bar {x};\rho )\), for all \(j\in \hat {J}_{k}\) and
Using (35) together with (36), for \(j\in \hat {J}_{k}\) we see that
From Theorem 1 and Lemma 9, we can obtain that {Sk} is bounded; thus, we can assume that \(S^{k+1}\rightarrow \bar {S}\). By noting Lemma 6 and Lemma 8, and passing to the limit as \(k\rightarrow \infty \), we obtain that
As already seen, \(S^{k+1}+v^{k+1}\rightarrow 0\) implies that \(-\bar S\in \partial I_{X}(\bar x)\), thus \(\langle -\bar S,y-\bar x\rangle \leq 0\). Hence
By (11), we have \(G(y,\bar {\alpha },\bar {E})\leq G(y)+D(\bar {E},\bar {\alpha })\leq G(y)+tol_{2}\). Since also \(y\in B(\bar {x};\rho )\), then \(f(y)\geq f(\bar {x})-\epsilon \|y-\bar {x}\|\) for all \(y\in X_{\rho ,\epsilon ,tol_{2}}=\{y\in X\cap B(\bar {x};\rho ): G(y)<-(\rho \epsilon +tol_{2})\}\). The proof is complete. □
5 Computational results
From Section 2.3, we find that the SIP can be rewritten as a finite constrained optimization problem when the lower level problem is a concave optimization problem. Hence, we consider the nonconvex nonsmooth constrained optimization problems and the nonconvex nonsmooth SIP problems.
In this section, we report some numerical results to verify the practical efficiency of our approach. We consider two classes of test problems and those test problems are introduced in Appendix Appendix. Solvers. For comparison purposes, we use the following solvers:
-
PBM-inexact, the Proximal Bundle Method from [23].
-
ICPBM, the Infeasible Constrained Proximal Bundle Method from [17].
-
SolvOpt, the pubic software for local nonlinear optimization problems. The code is available at https://imsc.uni-graz.at/kuntsevich/solvopt/index.html.
-
ASA, the Adaptive Subdivision Algorithm from [41].
-
fseminf, Matlab Optimization Toolbox for finding minimum of semi-infinitely constrained multivariable nonlinear function.
Solvers ICPBM and SolvOpt are designed for nonsmooth nonconvex constrained problems; these methods are only applicable to Problems 1–8. Solvers ASA and fseminf are designed for semi-infinite programming problems; these methods are only applicable to Problems 9–23. All solvers were implemented in MATLAB R2018a (Windows 10 64-bit). The experiments were performed on a PC with Intel Core i5, 2.4 GHz. We solve the quadratic subproblems using quadprog, which is available in the Matlab Optimization Toolbox. For solving the NLP subproblems, ASA and Algorithm 1 use fmincon of Matlab Optimization Toolbox. Parameters. Parameters of Algorithm 1 are set as follows: HCode \(tol_{1}=10^{-6},\ tol_{2}=10^{-6},\ s=2,\ \gamma =2,\ i_{n\max \limits }=100.\) We choose the descent parameter as follows:
The proximal parameter is only updated at serious steps. We set μ0 = 5 and
where
In ICPBM, the stopping tolerance tol has been set to 10− 6. For all the other parameters, we have used the default values suggested by the respective codes [17]. Furthermore, in the parameters selection of PBM-inexact and ASA we have used the default values [23, 41]. In ASA, the maximum number of iterations is fixed to 10,000. In fseminf, the initial sampling interval is set as follows:
The stopping criterions of fseminf are set as follows:
Results. We have summarized the results of our numerical experiments in Tables 1 and 2, where we have used the following notations:
-
P is the number of problems,
-
NF is the number of function evaluations,
-
NT is the number of nodes in the final subdivision,
-
time is the CPU time in seconds,
-
f∗ is the objective function value when the algorithm terminates,
-
g∗ is the constraint function value when the algorithm terminates for constrained optimization problems,
-
G∗ measures the feasibility of x∗ when the algorithm terminates for SIP problems, i.e.,
$$ G^{*}=\underset{t\in\bar{E}}{\max} g(x^{*},t), $$where x∗ is the obtained solution of the problem when the algorithm terminates and
$$ \bar{E}= \left\{ \begin{aligned} & a :\frac{b-a}{10^{6}}:b, & p=1, \\ & a_{i} :\frac{b_{i}-a_{i}}{10^{4}}:b_{i},i=1,2, & p=2. \end{aligned} \right. $$
The results presented in Table 1 show that the new solver Algorithm 1 is more reliable to find an optimal solution to each problem than the other solvers. The solvers Algorithm 1 and PBM-inexact reach the same optimal value in most problems, only in some of the problems (Problem 3, 7 (n= 100)) Algorithm 1 reaches lower objective value. Despite being more unreliable than Algorithm 1, the solver PBM-inexact has also quite a good performance and yields feasible points in all problems. In some of the problems (Problem 7 (n= 20,100,200), 8 (n= 100,200)) Algorithm 1 reaches a lower objective value than the solver ICPBM. The solver SolvOpt fails to find the optimal value in Problems 7 and 8 (n≥ 20).
Figure 1 depicts the performance plot comparing the solvers Algorithm 1, PBM-inexact, ICPBM and SolvOpt on Problems 1–8. It can be observed that Algorithm 1 performs much better in terms of CPU time, compared to the other solvers. The solver SolvOpt is the faster one for n ≤ 10. However, it uses more computational efforts than the other solvers in Problems 7 and 8 (n > 10).
All in all, we can conclude that the new solver Algorithm 1 is really efficient to solve the nonconvex, nonsmooth constrained optimization problems.
The results presented in Table 2 show that the solver Algorithm 1 is efficient to find an optimal solution to each SIP problem. The solvers Algorithm 1 and PBM-inexact find a feasible point to each SIP problem. The solver ASA fails to find the optimal solution approximately in 40% of the cases while fseminf fails in 80% of the cases. From Table 2, it observes that Algorithm 1 requires a lot more nodes than ASA. However, Algorithm 1 is much faster than ASA. This is expected because of the subproblems of ASA use more computational efforts than the subproblems of Algorithm 1 and the subproblems of our method do not depend on the number of the nodes in the subdivision.
Figure 2 depicts the performance plot comparing the solvers Algorithm 1, PBM-inexact, ASA, and fseminf on Problems 9–23. Algorithm 1 performs much better in terms of CPU time, compared to the solvers PBM-inexact and ASA. The solver fseminf is the faster one in some cases. However, it cannot find the optimal value in most problems.
All in all, we can conclude that Algorithm 1 is really efficient to solve the nonconvex, nonsmooth SIP problems.
6 Conclusions
In this paper, we design a feasible proximal bundle method with convexification for SIP problems. We consider SIP problems that involve nonsmooth, nonconvex functions and present upper bounding problem. The upper bounding problem is constructed based on a concave relaxation of the lower level problem which results in a restriction of the SIP. The relaxed lower level problem is a finite nonsmooth problem and constructed with ideas from the α BB method of global optimization. Then, we introduce an improvement function of the upper bounding problem, construct a cutting plane model of the improvement function, and reformulate the cutting plane model as a quadratic programming problem and solve it. In contrast to the traditional bundle method, the cutting plane model involves new parameters E, α which vary along the iterations. If \(D(E,\ \alpha )\rightarrow 0\), the upper bounding problems converge to the original SIP problem. We perform an initialization to generate a feasible initial point of our proximal bundle method. Under our refinement procedure and descent condition, we can obtain that each iteration point is feasible. The reasonable convergence properties of the our algorithms are obtained under mild assumptions. The presented results of numerical experiments confirm that Algorithm 1 is efficient for solving nonsmooth, nonconvex SIP problems. The feasibility of Algorithm 1 is demonstrated through theoretical analysis and numerical experiments.
References
Adjiman, C.S., Androulakis, I.P., Floudas, C.A.: A global optimization method, α BB, for general twice-differentiable constrained NLPs-I: theoretical advances. Comput. Chem. Eng. 22, 1137–1158 (1998)
Adjiman, C.S., Androulakis, I.P., Floudas, C.A.: A global optimization method, α BB, for general twice-differentiable constrained NLPs-II: implementation and computational results. Comput. Chem. Eng. 22, 1159–1179 (1998)
Bagirov, A., Karmitsa, N., Mäkelä, M. M.: Introduction to Nonsmooth Optimization: Theory, Practice and Software. Springer, Cham (2014)
Bhattacharjee, B., Green, Jr.W. H., Barton, P.I.: Interval methods for semi-infinite programs. Comput. Optim. Appl. 30, 63–93 (2005)
Bhattacharjee, B., Lemonidis, P., Green, W.H. Jr., Barton, P.I.: Global solution of semi-infinite programs. Math. Program. 103, 283–307 (2005)
Clarke, F.H., Ledyaev, Y.S., Stern, R.J., Wolenski, P.R.: Nonsmooth Analysis and Control Theory. Springer, New York (1998)
Daniilidis, A., Georgiev, P.: Approximate convexity and submonotonicity. J. Math. Anal. Appl. 291 (1), 292–301 (2004)
Djelassi, H., Mitsos, A.: A hybrid discretization algorithm with guaranteed feasibility for the global solution of semi-infinite programs. J Glob Optim. 68, 227–253 (2017)
Floudas, C.A., Stein, O.: The adaptive convexification algorithm: a feasible point method for semi-infinite programming. SIAM J. Optim. 18, 1187–1208 (2007)
Fuduli, A., Gaudioso, M., Giallombardo, G., Miglionico, G.: A partially inexact bundle method for convex semi-infinite minmax problems. Commun. Nonlinear Sci. Numer. Simul. 21, 172–180 (2015)
Gustafson, S.-Å., Kortanek, K.: Semi-infinite programming and applications. In: Mathematical Programming The State of the Art, pp 132–157. Springer, Berlin (1983)
Hare, W., Sagastizábal, C.: A redistributed proximal bundle method for nonconvex optimization. SIAM J. Optim. 20 (5), 2442–2473 (2010)
Hare, W., Sagastizábal, C., Solodov, M.: A proximal bundle method for nonsmooth nonconvex functions with inexact information. Comput. Optim. Appl. 63, 1–28 (2016)
Hettich, R.: An implementation of a discretization method for semi-infinite programming. Math. Program. 34 (3), 354–361 (1986)
Hettich, R., Kortanek, K.O.: Semi-infinite programming: theory, methods, and applications. Soc. Ind. Appl. Math. 35 (3), 380–429 (1993)
Hoseini Monjezi, N., Nobakhtian, S.: A filter proximal bundle method for nonsmooth nonconvex constrained optimization. J. Glob. Optim. (2020)
Hoseini Monjezi, N., Nobakhtian, S.: A new infeasible proximal bundle algorithm for nonsmooth nonconvex constrained optimization. Comput Optim Appl. 74, 443–480 (2019)
Karas, E., Ribeiro, A., Sagastizábal, C., Solodov, M.: A bundle-filter method for nonsmooth convex constrained optimization. Math. Program. 116, 297–320 (2009)
Karmitsa, N.: Test problems for large-scale nonsmooth minimization. Reports of the Department of Mathematical Information Technology, Series B, Scientific computing, No. B 4/2007 (2007)
Kiwiel, K.C.: Methods of Descent for Nondifferentiable Optimization, vol. 1133. Springer, Berlin, Heidelberg (2006)
Kortanek, K.O., No, H.: A central cutting plane algorithm for convex semi-infinite programming problems. SIAM J. Optim. 3 (4), 901–918 (1993)
López, M., Still, G.: Semi-infinite programming. Eur. J. Oper. Res. 180 (2), 491–518 (2007)
Lv, J., Pang, L.P., Meng, F.Y.: A proximal bundle method for constrained nonsmooth nonconvex optimization with inexact information. J. Global Optim. 70, 517–549 (2018)
Lv, J., Pang, L.P., Xu, N., Xiao, Z.H.: An infeasible bundle method for nonconvex constrained optimization with application to semi-infinite programming problems. Numer. Algorithms 80, 397–427 (2019)
Meng, F.Y., Pang, L.P., Lv, J., Wang, J.H.: An approximate bundle method for solving nonsmooth equilibrium problems. J. Global Optim. 68, 537–562 (2017)
Mifflin, R.: Semismooth and semiconvex functions in constrained optimization. SAIM J. Control Optim. 15 (6), 959–972 (1977)
Mitsos, A.: Global optimization of semi-infinite programs via restriction of the right-hand side. Optimization 60 (10-11), 1291–1308 (2011)
Mitsos, A., Djelassi, H.: A Test Set of Semi-Infinite Programs. Process Systems Engineering (AVT.SVT), RWTH Aachen University, Aachen, Germany. http://web.mit.edu/mitsos/www/pubs/siptestset.pdf (2016)
Mitsos, A., Lemonidis, P., Lee, C.K., Barton, P.I.: Relaxation-based bounds for semi-infinite programs. SIAM J. Optim. 19, 77–113 (2007)
Mitsos, A., Tsoukalas, A.: Global optimization of generalized semi-infinite programs via restriction of the right hand side. J. Glob. Optim. 61 (1), 1–17 (2015)
Pang, L.P., Lv, J., Wang, J.H.: Constrained incremental bundle method with partial inexact oracle for nonsmooth convex semi-infinite programming problems. Comput. Optim. Appl. 64 (2), 433–465 (2016)
Pang, L., Wu, Q., Wang, J., et al.: A discretization algorithm for nonsmooth convex semi-infinite programming problems based on bundle methods. Comput Optim Appl. 76, 125–153 (2020)
Polyak, B.T.: Introduction to Optimization. Optimization Software, Inc., Publications Division, New York (1987)
Rustem, B., Nguyen, Q.: An algorithm for the inequality-constrained discrete min-max problem. SIAM J. Optim. 8, 265–283 (1998)
Sagastizábal, C., Solodov, M.: An infeasible bundle method for nonsmooth convex constrained optimization without a penalty function or a filter. SIAM J. Optim. 16 (1), 146–169 (2005)
Shiu, T.J., Wu, S.Y.: Relaxed cutting plane method with convexification for solving nonlinear semi-infinite programming problems. Comput Optim Appl. 53, 91–113 (2012)
Spingarn, J.E.: Submonotone subdifferentials of Lipschitz functions. Trans. Am. Math. Soc. 264, 77–89 (1981)
Stein, O., Steuermann, P.: The adaptive convexification algorithm for semi-infinite programming with arbitrary index sets. Math. Program. 136, 183–207 (2012)
Still, G.: Discretization in semi-infinite programming: the rate of convergence. Math. Programm. 91 (1), 53–69 (2001)
Tang, C., Liu, S., Jian, J., et al.: A feasible SQP-GS algorithm for nonconvex, nonsmooth constrained optimization. Numer Algor. 65, 1–22 (2014)
Wang, S.X., Yuan, Y.X.: Feasible method for semi-infinite programs. SIAM J. Optim. 25 (4), 2537–2560 (2015)
Xu, M.W., Wu, S.Y., Ye, J.J.: Solving semi-infinite programs by smoothing projected gradient method. Comput. Optim. Appl. 59, 591–616 (2014)
Yang, Y., Pang, L., Ma, X., et al.: Constrained Nonconvex Nonsmooth Optimization via Proximal Bundle Method. J Optim Theory Appl. 163, 900–925 (2014)
Zhang, L.P., Wu, S.Y., López, M.A.: A new exchange method for convex semi-infinite programming. SIAM J. Optim. 20 (6), 2959–2977 (2010)
Funding
This work was supported by the Key Research and Development Projects of Shandong Province (NO. 2019GGX104089), and the Natural Science Foundation of Shandong Province (NO. ZR2019BA014).
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix: 1: Test problems
Appendix: 1: Test problems
Problem 1
[23, 40] Dimension: n= 2, \(f(x) = 8|{x_{1}^{2}}-x_{2}|+(1-x_{1})^{2}\), \(g(x) = \max \limits \left \{\sqrt {2}x_{1},\ 2 x_{2}\right \}-1\), X = [− 2, 2]2, x0 = (1, 1)T.
Problem 2
[34] Dimension: n= 2, \(f(x) = \max \limits \left \{{{x_{1}^{2}}}+{{x_{2}^{4}}},\ (2-x_{1})^{2}+(2-x_{2})^{2},\ 2\exp (x_{2}-x_{1})\right \}\), \(g(x) = \max \limits \left \{-{{x_{1}^{4}}}-2{{x_{2}^{2}}}-1,\ 2 {{x_{1}^{3}}}-{{x_{2}^{2}}}-2.5\right \}\), X = [− 4, 4]2, x0 = (2, 2)T.
Problem 3
[34] Dimension: n= 2, \(f(x) = \max \limits \left \{{{x_{1}^{2}}}+{{x_{2}^{2}}}+x_{1} x_{2},\ -{{x_{1}^{2}}}-{{x_{2}^{2}}}-x_{1} x_{2},\ \sin \limits {x_{1}},\! -\sin \limits {x_{1}},\\ \cos \limits {x_{2}},\ -\cos \limits {x_{2}}\right \}\), \(g(x) = \max \limits \left \{-{{x_{1}^{4}}}-2{{x_{2}^{2}}}-1,\ 2 {{x_{1}^{3}}}-{{x_{2}^{2}}}-2.5\right \}\), X = [− 4, 4]2, x0 = (3, 1)T.
Problem 4
[34] Dimension: n= 2, \(f(x) = \max \limits \left \{{{x_{1}^{4}}}+{{x_{2}^{2}}},\ (2-x_{1})^{2}+(2-x_{2})^{2},\ 2\exp (x_{2}-x_{1})\right \}\), \(g(x) = \max \limits \left \{{{x_{1}^{2}}}-{{x_{2}^{2}}},\ -2{{x_{1}^{3}}}-{{x_{2}^{2}}}\right \}\), X = [− 4, 4]2, x0 = (0, 1)T.
Problem 5
[34] Dimension: n= 2, \(f(x) = \max \limits \left \{{{x_{1}^{2}}}+{{x_{2}^{2}}},\ (2-x_{1})^{2}+(2-x_{2})^{2},\ 2\exp (x_{2}-x_{1})\right \}\), \(g(x) = \max \limits \left \{{x_{1}}+{x_{2}}-2,\ -{{x_{1}^{2}}}-{{x_{2}^{2}}}+2.25\right \}\), X = [− 4, 4]2, x0 = (2.1, 1.9)T.
Problem 6
[34] Dimension: n= 2, \(f(x) = \max \limits \left \{10(x_{2}-{x_{1}^{2}}),\ 10({x_{1}^{2}}-x_{2}),\ 1-x_{1},\ x_{1}-1\right \}\), \(g(x) = \max \limits \left \{100{x_{1}^{2}}+{x_{2}^{2}}-101,\ 80{x_{1}^{2}}-{x_{2}^{2}}-79\right \}\), X = [− 4, 4]2, x0 = (− 1.2, 1)T.
Problem 7
[19] Dimension: n= 20, 50, 100, 200, \(f(x) =\max \limits \left \{{\sum }_{i=1}^{n-1}({x_{i}^{2}}+(x_{i+1}-1)^{2}+x_{i+1}-1)\right .\), \(~~~~~~~~~~~~~\left .{\sum }_{i=1}^{n-1}(-{x_{i}^{2}}-(x_{i+1}-1)^{2}+x_{i+1}+1)\right \}\), \(g(x) = {\sum }_{i=1}^{n-1}({x_{i}^{2}}+x_{i+1}^{2}+x_{i} x_{i+1}-2x_{i}-2x_{i+1}+1.0)\), X = [− 10, 10]n, x0 = ones(n, 1).
Problem 8
[19, 40] Dimension: n= 10, 50, 100, 200, 500, 1000, \(f(x) = {\sum }_{i=1}^{n-1}(-x_{i}+2({x_{i}^{2}}+x_{i+1}^{2}-1)+1.75|{x_{i}^{2}}+x_{i+1}^{2}-1|)\), \(g(x) = {\sum }_{i=1}^{n-2}((3-2 x_{i+1})x_{i+1}-x_{i}-2x_{i+2}+2.5)\), X = [− 10, 10]n, x0 = ones(n, 1).
Problem 9
[41] Dimension: n = 2, p = 1, \(f(x) = \frac {1}{3}{x_{1}^{2}}+{x_{2}^{2}}+\frac {1}{2}x_{1}\), \(g(x,t) = (1-{x_{1}^{2}} t^{2})^{2}-x_{1} t^{2}-{x_{2}^{2}}+x_{2}\), X = [− 2, 2]2, T = [0, 1], x0 = (− 1,− 1)T.
Problem 10
[28] Dimension: n = 3, p = 1, \(f(x) = \exp (x_{1})+\exp (x_{2})+\exp (x_{3})\), g(x,t) = 1/(1 + t2) − x1 − x2t − x3t2, X = [− 2, 2]3, T = [0, 1], x0 = (1, 1, 1)T.
Problem 11
[28] Dimension: n = 3, p = 1, f(x) = x1 + x2/2 + x3/3, \(g(x,t) = \exp (t-1)-x_{1} -x_{2} t-x_{3} t^{2}\), X = [− 2, 2]3, T = [0, 1], x0 = (1, 1, 1)T.
Problem 12
[41] Dimension: n = 3, p = 1, \(f(x) = {x_{1}^{2}}+{x_{2}^{2}}+{x_{3}^{2}}\), \(g(x,t) = x_{1} +x_{2}\exp (x_{3} t)+\exp (2t)-2\sin \limits (4t)\), X = [− 4, 2]3, T = [0, 1], x0 = (1, 1, 1)T.
Problem 13
[31, 32] Dimension: n = 3, p = 2, f(x) = |x1| + |x2| + |x3|, \(g(x,t) = x_{1}+x_{2}\exp (x_{3} t)-\exp (2x_{1} t)+\sin \limits (4 t)\), X = [− 1, 1]3, T = [0, 1], x0 = ones(3, 1).
Problem 14
[31, 32] Dimension: n = 3, p = 2, f(x) = |x1| + |x2| + |x3|, \(g(x,t) = x_{1}+x_{2}\exp (x_{3} t_{1})-\exp (2 t_{2})+\sin \limits (4 t_{1})\), X = [− 1, 1]3, T = [0, 1]2, x0 = ones(3, 1).
Problem 15
[31, 32] Dimension: n = 4, p = 2, f(x) = 1/2(|x1| + |x2| + |x3| + |x4|), \(g(x,t) = \sin \limits (t_{1} t_{2})-x_{1}-x_{2} t_{1}-x_{3} t_{2}-x_{4} t_{1} t_{2}\), X = [− 4, 4]6, T = [0, 1]2, x0 = ones(4, 1).
Problem 16
Dimension: n = 6, p = 2, \(f(x) = {\sum }_{i=1}^{n-1}(-x_{i}+2({x_{i}^{2}}+x_{i+1}^{2}-1)+1.75|{x_{i}^{2}}+x_{i+1}^{2}-1|)\), \(g(x,t) = \sin \limits (t_{1} t_{2})-x_{1}-x_{2} t_{1}-x_{3} t_{2}-x_{4} {t_{1}^{2}}-x_{5} t_{1} t_{2}-x_{6} {t_{2}^{2}}\), X = [− 4, 4]6, T = [0, 1]2, x0 = ones(6, 1).
Problem 17
Dimension: n = 6, p = 2, \(f(x) = {\sum }_{i=1}^{n-1}(-x_{i}+2({x_{i}^{2}}+x_{i+1}^{2}-1)+1.75|{x_{i}^{2}}+x_{i+1}^{2}-1|)\), \(g(x,t) = (1+{t_{1}^{2}}+{t_{2}^{2}})^{2}-x_{1}-x_{2} t_{1}-x_{3} t_{2}-x_{4} {t_{1}^{2}}-x_{5} t_{1} t_{2}-x_{6} {t_{2}^{2}}\), X = [− 4, 4]6, T = [0, 1]2, x0 = ones(6, 1).
Problem 18
Dimension: n = 6, p = 2, \(f(x) = {\sum }_{i=1}^{n-1}\max \limits \left \{{x_{i}^{2}}+(x_{i+1}-1)^{2}+x_{i+1}-1,\ -{x_{i}^{2}}-(x_{i+1}-1)^{2}+x_{i+1}+1\right \}\), \(g(x,t) = \exp ({t_{1}^{2}}+{t_{2}^{2}})-x_{1}-x_{2} t_{1}-x_{3} t_{2}-x_{4} {t_{1}^{2}}-x_{5} t_{1} t_{2}-x_{6} {t_{2}^{2}}\), X = [− 4, 4]6, T = [0, 1]2, x0 = ones(6, 1).
Problem 19
Dimension: n = 6, p = 2, \(f(x) = {\sum }_{i=1}^{n-1}\max \limits \left \{{x_{i}^{2}}+(x_{i+1}-1)^{2}+x_{i+1}-1,\ -{x_{i}^{2}}-(x_{i+1}-1)^{2}+x_{i+1}+1\right \}\), \(g(x,t) = \sin \limits (t_{1} t_{2})-x_{1}-x_{2} t_{1}-x_{3} t_{2}-x_{4} {t_{1}^{2}}-x_{5} t_{1} t_{2}-x_{6} {t_{2}^{2}}\), X = [− 4, 4]6, T = [0, 1]2, x0 = ones(6, 1).
Problem 20
Dimension: n = 6, p = 2, \(f(x) = {\sum }_{i=1}^{n-1}\max \limits \left \{{x_{i}^{2}}+(x_{i+1}-1)^{2}+x_{i+1}-1,\ -{x_{i}^{2}}-(x_{i+1}-1)^{2}+x_{i+1}+1\right \}\), \(g(x,t) = (1+{t_{1}^{2}}+{t_{2}^{2}})^{2}-x_{1}-x_{2} t_{1}-x_{3} t_{2}-x_{4} {t_{1}^{2}}-x_{5} t_{1} t_{2}-x_{6} {t_{2}^{2}}\), X = [− 4, 4]6, T = [0, 1]2, x0 = ones(6, 1).
Problem 21
Dimension: n = 6, p = 2,
\(f(x)\! =\! \max \limits \left \{{\sum }_{i=1}^{n-1}({x_{i}^{2}}+(x_{i+1}-1)^{2}+x_{i+1}-1)\right .\), \(~~~~~~~~~~~~~~~~~\left .{\sum }_{i=1}^{n-1}(-{x_{i}^{2}}-(x_{i+1}-1)^{2}+x_{i+1}+1)\right \}\),
\(g(x,t) = \sin \limits (t_{1} t_{2})-x_{1}-x_{2} t_{1}-x_{3} t_{2}-x_{4} {t_{1}^{2}}-x_{5} t_{1} t_{2}-x_{6} {t_{2}^{2}}\),
X = [− 4, 4]6, T = [0, 1]2, x0 = ones(6, 1).
Problem 22
Dimension: n = 6, p = 2, \(f(x) =\max \limits \left \{\sum \limits _{i=1}^{n-1}({x_{i}^{2}}+(x_{i+1}-1)^{2}+x_{i+1}-1)\right .\), \(~~~~~~~~~~~~~\left .\sum \limits _{i=1}^{n-1}(-{x_{i}^{2}}-(x_{i+1}-1)^{2}+x_{i+1}+1)\right \}\),
\(g(x,t) = (1+{t_{1}^{2}}+{t_{2}^{2}})^{2}-x_{1}-x_{2} t_{1}-x_{3} t_{2}-x_{4} {t_{1}^{2}}-x_{5} t_{1} t_{2}-x_{6} {t_{2}^{2}}\),
X = [− 4, 4]6, T = [0, 1]2, x0 = ones(6, 1).
Problem 23
Dimension: n = 6, p = 2,
\(f(x) = \max \limits _{1\leq i\leq n }\left \{h(-{\sum }_{i=1}^{n}x_{i}),\ h(x_{i})\right \}\), where \(h(y)=\ln (|y|+1),\ \forall y\in \mathbb {R}\),
\(g(x,t) = \sin \limits (t_{1} t_{2})-x_{1}-x_{2} t_{1}-x_{3} t_{2}-x_{4} {t_{1}^{2}}-x_{5} t_{1} t_{2}-x_{6} {t_{2}^{2}}\),
X = [− 4, 4]6, T = [0, 1]2, x0 = ones(6, 1).
Rights and permissions
About this article
Cite this article
Pang, LP., Wu, Q. A feasible proximal bundle algorithm with convexification for nonsmooth, nonconvex semi-infinite programming. Numer Algor 90, 387–422 (2022). https://doi.org/10.1007/s11075-021-01192-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11075-021-01192-9