1 Introduction and general aim

We are interested in the problem

$$\begin{aligned} \min f(u),\quad u\in \mathbb {R}^n, \end{aligned}$$
(1.1)

where \(f(\cdot ):\mathbb {R}^n\rightarrow \mathbb {R}\) is a (finite-valued) convex function. For each given \(u\in \mathbb {R}^n\), an oracle—a noisy one-delivers inexact information, namely

$$\begin{aligned} \left\{ \begin{array}{l} f_u=f(u)-\eta _u\quad \text{ and }\\ g_u\in \mathbb {R}^n\,\text{ such } \text{ that }\,f(\cdot )\geqslant f_u+\left\langle g_u,\cdot -u\right\rangle -\eta ^g_u \\ \text{ with } \eta _u\leqslant \eta \, \text{ and }\, \eta ^g_u \leqslant \eta \quad \text{ for } \text{ all }\; u \in \mathbb {R}^n \end{array}\right. \end{aligned}$$
(1.2)

where the error bound \(\eta \) is possibly unknown. For a given \(u\), function values will be denoted by a letter such as \(f_u\); reserving the notation \(f(\cdot )\) for the function itself.

To solve (1.1) we put in place a framework for bundle methods dealing with oracles of the type (1.2). The considered setting is versatile and general, in the sense that it covers and extends previous literature, such as the inexact bundle methods [12, 15, 20], the incremental bundle method [6], and the partly inexact method [16]. We also consider new methods, the Controllable Bundle Algorithm 5.4 and the Asymptotically Exact Algorithm in Sect. 7.1.4. The latter is a proximal variant of the level bundle method for oracles with on-demand accuracy considered in [4].

Similarly to the spirit of [3], our development highlights the main arguments and assumptions used to establish each intermediate result. The analysis is presented in a way that reveals how different procedures controlling oracle noise result in algorithms solving (1.1) with different degrees of accuracy.

To give a flavor of different situations fitting (1.2), Sect. 2 starts with a broad set of examples. Section 3 organizes the essential features of bundle methods in two sets of parameters whose particularization gives rise to specific algorithms. The parametric setting is useful to state a general algorithmic pattern in Sect. 4. The important mechanism of noise attenuation, to be put in place when the oracle error cannot be controlled, is addressed in Sect. 5. This section also illustrates a parameter specification for the algorithmic pattern, the Controllable Bundle Algorithm 5.4, used along Sect. 6 to guide the reader through the various convergence results therein. The final Sect. 7 considers many algorithms covered by our synthetic theory, including constrained bundle methods.

2 Oracle examples

By the second line in (1.2),

$$\begin{aligned} f(\cdot )\geqslant f(u)+\left\langle g_u ,\cdot -u\right\rangle -(\eta _u+\eta ^g_u) \end{aligned}$$
(2.1)

from which, evaluating at \(u\) we deduce that, independently of the errors sign, \(\eta _u+\eta ^g_u \geqslant 0\). As a result, \(g_u\) is a Convex Analysis \(\varepsilon \)-subgradient:

$$\begin{aligned} g_u\in \partial _{\eta _u+\eta ^g_u} f(u)\quad \text{ with }\, \eta _u +\eta ^g_u \geqslant 0 \quad \text{ for } \text{ all }\; u \in \mathbb {R}^n. \end{aligned}$$
(2.2)

Even if in (1.2) the value of the upper error bound \(\eta \) is unknown, the inequality above implies that \(\eta \geqslant \eta _u\geqslant -\eta ^g_u\geqslant -\eta \): both oracle errors are bounded from below by \(-\eta \).

An exact oracle has \(\eta _u\equiv \eta ^g_u\equiv 0\), the output is \(f_u=f(u)\) and a true subgradient. In an important subclass of inexact oracles illustrated by Examples 2.1 and 2.2, \(\eta ^g_u\equiv 0\) and \(\eta _u\geqslant 0\), by (2.2). Following [21], we shall call this subclass of lower oracles, because a lower linearization of \(f(\cdot )\) is available:

$$\begin{aligned} f(u)-\eta _u= f_u\leqslant f(u)\quad \text{ and }\quad f(\cdot )\geqslant f_u+\left\langle g_u ,\cdot -u\right\rangle . \end{aligned}$$
(2.3)

Upper oracles, by contrast, can over-estimate function values: in (2.1) the error \(\eta ^g\) is positive.

Example 2.1

(Minimax: Lagrangian) For given functions \(h(\cdot )\) and \(c(\cdot )\) and \(X\) a nonempty compact set, (1.1) is dual to the primal problem

$$\begin{aligned} \max _{x\in X}\, h(x),\quad c(x)=0\in \mathbb {R}^n. \end{aligned}$$
(2.4)

Specifically, for a multiplier \(u\in \mathbb {R}^n\) the dual function is given by

$$\begin{aligned} f(u) :=\max _{x\in X}\,L(x,u),\quad \text{ with } \, L(x,u) :=h(x)+\left\langle u, c(x)\right\rangle . \end{aligned}$$

In a more general setting \(L(x,\cdot )\) is convex. Suppose that given \(u\), the oracle outputs \(f_{u} :=L(x,u)\) for some \(x\in X\), together with some \(g_{u}\in \partial _u L(x,u)\). By convexity of \(L(x,\cdot )\) and definition of \(f(\cdot )\),

$$\begin{aligned} f_{u} +\left\langle g ,\cdot -u_0\right\rangle = L(x,u)+\left\langle g ,\cdot -u_0\right\rangle \leqslant L(x,\cdot )\leqslant f(\cdot ). \end{aligned}$$

For this lower oracle (2.3) holds. Such is the case in Lagrangian relaxation or column generation when the operation \(\max _{x\in X}L(x,u)\) is not performed exactly.\(\square \)

Example 2.2

(Minimax: Two-Stage Stochastic Linear Programs) Consider a stochastic linear program with decision variables organized in two levels, denoted by \(u\) and \(y\) for the first and second stage, respectively. If \(\xi \in \Xi \) represents uncertainty, for vectors \(e\) and \(q(\xi )\) and matrices \(T(\xi )\) and \(W\), the corresponding two-stage linear program with fixed recourse is

$$\begin{aligned} \left\{ \begin{array}{l@{\quad }l} \min _{u,y}&{} \left\langle e, u\right\rangle +\mathbb {E}[\left\langle q(\xi ),y\right\rangle ] \\ \text{ s.t. } &{} T(\xi )u+Wy= d(\xi )\quad \text{ for } \text{ almost } \text{ every }\, \xi \in \Xi ,\\ &{}y\geqslant 0, \end{array}\right. \end{aligned}$$

where we use the symbol \(\mathbb {E}(\cdot )\) for the expected value. For fixed \(u\) and \(\xi \) the recourse function

$$\begin{aligned} Q(u;\xi ) :=\inf _{y\geqslant 0}\Bigl \{ \left\langle q(\xi ),y\right\rangle \; \text{ s.t. }\;Wy= d(\xi )-T(\xi )u\Bigr \} \end{aligned}$$

gives in (1.1) an objective of the form \( f(u) :=\left\langle e, u\right\rangle + \mathbb {E}[Q(u;\xi )]\), which is finite-valued when recourse is relatively complete. We now explain how to build different oracles for this type of problems.

A dumb lower oracle For each fixed \(u\) and a given realization \(\xi \), the evaluation of the recourse function can be done by solving the dual linear program

$$\begin{aligned} Q(u;\xi )=\sup _{x} \Bigl \{ \left\langle d(\xi )-T(\xi )u, x\right\rangle \; \text{ s.t. } \; W^\top x\,\leqslant \, q(\xi )\Bigr \}. \end{aligned}$$

If, to speed up calculations, instead of performing the max-operation for the considered \(\xi \) we just take a feasible point \(x_{u,\xi }\) (satisfying \(W^\top x_{u,\xi }\leqslant q(\xi )\)), then an oracle taking \(f_{u} :=\left\langle e,u\right\rangle +\mathbb {E}[\left\langle d(\xi ) \right. \,\left. -T(\xi )u, x_{u,\xi }\right\rangle ]\), and \(g_{u} :=e-\mathbb {E}[T(\xi )^\top x_{u,\xi }]\) is of lower type and fits (2.3).

A controllable lower oracle A better estimate can be computed by making some iterations of a primal-dual linear programming solver. The oracle receives as additional input an error bound \(\bar{\eta }_u\geqslant 0\) and stops the primal-dual solver as soon as it finds a feasible point \(x_{u,\xi }\) for which \(\left\langle d(\xi )-T(\xi )u, x_{u,\xi }\right\rangle - Q(u;\xi )\leqslant \bar{\eta }_u\). For this oracle the subgradient error \(\eta ^g_u\) is null and \(f_u \in [f(u)-\bar{\eta }_u, f(u)]\) for any error bound \(\bar{\eta }_u\) chosen by the user [4].

Asymptotically exact oracles To build an oracle that is eventually exact everywhere from the controllable oracle, just take \(\bar{\eta }_{u}\rightarrow 0 \), to force the error bound to vanish along iterations.

A smarter oracle, called partly asymptotically exact, requires eventual exactness only for some input points \(u_k\). This is done by combining the three preceding lower oracles, as follows. Together with \((u,\bar{\eta }_u )\), the oracle receives as additional input a target \(\gamma _u\) and must compute \(f_u\) within the “on-demand” accuracy \(\bar{\eta }_u\) only when the target is reached:

$$\begin{aligned}&f_u \text{ is } \text{ computed }\\&\left\{ \begin{array}{l} \text{ as } \text{ in } \text{ the } \text{ dumb } \text{ lower } \text{ oracle }\, (f_u\in [f(u)-\eta , f(u)])\, \hbox {if}\, f_u > \gamma _u,\\ \text{ as } \text{ in } \text{ the } \text{ controllable } \text{ lower } \text{ oracle }\, (f_u\in [f(u)-\bar{\eta }_u, f(u)])\, \hbox {if} \,f_u \leqslant \gamma _u. \end{array}\right. \end{aligned}$$

The bound \(\eta \geqslant 0\) may be unknown while \(\bar{\eta }_u\geqslant 0\) is known and controllable. In addition to the dumb and controllable oracles, a third one, asymptotically exact, comes into play when the user sets the target as a goal of decrease for \(f\) at \(u_k\) and drives \(\bar{\eta }_{u_k}\) to zero; see [4].

An upper oracle Instead of considering all the random events (\(\Xi \) may be infinite), a small finite subset can be drawn from the sampling space for each given \(u\). The recourse function is computed exactly only for \(\xi \) in that subset; the corresponding minimizers \(x_{u,\xi }\) give approximate functional and gradient values in (2.3) whose errors have unknown sign. The bound \(\eta \) may be unknown but exists and depends on the probability distribution.\(\square \)

Example 2.3

(Chance-Constrained Programs) For a probability level \(p\in (0,1]\), a simple convex and compact polyhedron \(U\), and a log-concave probability distribution for \(\xi \), in

$$\begin{aligned} \left\{ \begin{array}{l@{\quad }l} \min _{u\in U}&{} \left\langle h, u\right\rangle \\ \text{ s.t. } &{} \mathbb {P}(Tu\leqslant \xi )\geqslant p \end{array}\right. \end{aligned}$$

the constraint is convex and can be smooth, but its oracle requires computing a costly gradient. To allow for approximate calculations, [21] minimizes over \(U\) the improvement function

$$\begin{aligned} f(u)=\max \{\left\langle h,u\right\rangle - \tau _1,\, -\ln (p)-\ln [\mathbb {P}(Tu\leqslant \xi )] -\tau _2\}\, \text{ for } \text{ a } \text{ parameter }\, \tau \in \mathbb {R}^2; \end{aligned}$$

see Sect. 7.4. The upper oracle in [21] delivers an unknown error that is bounded if so is \(U\). The error can be driven to zero at the expense of heavy computations. \(\square \)

Example 2.4

(Convex Composite Functions) All the functions above involve some maximization operation. In a more general setting, including eigenvalue optimization [11], given a convex function \(h(\cdot )\) that is positively homogeneous (like the \(\max \)-function) and a convex smooth operator \(c(\cdot )\), the objective in (1.1) can have the form \(f(\cdot )=(h\circ c)(\cdot )\). Suppose \(c:\mathbb {R}^n\rightarrow \mathbb {R}^m\) and let \(Dc(\cdot )\) denote its Jacobian. Given \(\hat{u}\in \mathbb {R}^n\), the function \(F(\cdot ;{\hat{u}}):\mathbb {R}^m\rightarrow \mathbb {R}\) defined by \(F(\cdot ;{\hat{u}}) :=h(c(\hat{u})+Dc(\hat{u})(\cdot -\hat{u}))\) is used in the composite bundle method [19] to solve (1.1). Since computing the Jacobian matrix is expensive, using oracle information for the convex function \(F(\cdot ;{\hat{u}})\) eases calculations. With respect to the true \(f\)-information, nothing can be said about the error sign: the oracle can be of upper type; see Sect. 7.5.\(\square \)

We will retain from these oracles that several situations can occur:

  • Dumb oracle not much is known about output \(f_u\) and \(g_u\). Such is the case in Example 2.2. of the dumb lower and the (totally dumb) upper oracles.

  • Informative oracle in addition to \(f_u\) and \(g_u\), the oracle returns some reliable upper bound \(\bar{\eta }_u\) for \(\eta _u\). For Example 2.1, when the \(L(\cdot ,u)\)-maximization is a constrained problem solved by a primal-dual method, the oracle outputs some feasible \(x_u\), giving \(f_u=L(x_u,u)\) and computes with dual arguments an upper bound \(M_u\) (used to define \(\bar{\eta }_u :=M_u-f_u\geqslant 0\)).

  • Controllable oracle in addition to \(u\), the oracle receives \(\bar{\eta }_u\) and must compute \(f_u\) within \(\bar{\eta }_u\)-accuracy. This situation has itself several sub-cases:

    • Achievable high accuracy. A smaller \(\bar{\eta }_u\) implies an acceptable increase in computing times, as with the controllable lower oracle in Example 2.2.

    • Partly achievable high accuracy. It is acceptable to make exact calculations at “good” points, as in the partly asymptotically exact oracle in Example 2.2. By this token, \(\bar{\eta }_u\) can be managed more aggressively. We shall see in Sect. 7.5 that the convex composite oracle in Example 2.4 fits this category.

    • Exorbitantly high accuracy. For instance if in Example 2.1 the \(L(\cdot ,u)\)-maximization is a difficult, combinatorial and/or large scale problem, or when in Example 2.2 the sampling space is too large and/or the probability distribution too involved.

3 Parametric characterization of bundle methods

An algorithm to solve (1.1) should construct a sequence \(\{(\hat{u}_k,\hat{g}_k)\}\) whose respective primal and dual terms aim at minimizing \(f(\cdot )\) and approaching \(0\in \mathbb {R}^n\) (thus providing a certificate of optimality). The centers \(\hat{u}_k\) defined below are extracted from the algorithm iterate sequence \(\{{u}_k\}\) by collecting points that provide sufficient progress towards the goal of solving (1.1); for instance, by reducing the functional value. We describe this construction for proximal bundle methods with the least possible references to the oracle errors. Noise comes into play only a posteriori, to determine to which extent the algorithm really solves (1.1).

3.1 Defining the set P

To make the parametric notation clear, we start with the pure cutting-plane methods [2, 14]. Having called the oracle at a number of points \(u_j\), these algorithms accumulate linearizations

$$\begin{aligned} f^{\scriptscriptstyle L}_j(u) :=f_{u_j}+\left\langle g_{u_j} ,u-u_j\right\rangle . \end{aligned}$$
(3.1)

To compute the next iterate, \(u_{k+1}\), the master-program minimizes the cutting-plane model

$$\begin{aligned} \check{f}_k(\cdot ):= \max {\bigl \{f^{\scriptscriptstyle L}_j(\cdot )\,:\, j \in {J}_k :=\{1,\ldots , k\}\bigr \}}. \end{aligned}$$

As a result, the following set of parameters fully characterizes a cutting-plane method:

$$\begin{aligned} \mathbf{P} \,=\left\{ \begin{array}{l} \text{ the } \text{ convex }\, model\,\check{f}_k(\cdot ) \text{ and } \\ \text{ a } \text{ measure } \text{ of } \text{ progress, } \text{ such } \text{ as }\,f_{u_{k+1}}-\check{f}_k(u_{k+1}) \end{array} \right\} , \end{aligned}$$

that is, the optimality certificate used to stop the algorithm. Taking the maximum over \(J_k\) in the inequalities \(f^{\scriptscriptstyle L}_j(\cdot )\leqslant f(\cdot )+\eta ^g_{u_j}\), obtained from (1.2), gives:

$$\begin{aligned} \check{f}_k(\cdot )\leqslant f(\cdot )+\max _{j\in J_k}\eta ^g_{u_j} \quad \text{ for } \text{ all }\, k. \end{aligned}$$
(3.2)

For stabilized cutting-plane variants, such as [2, 4, 8, 11], the set P   includes some devices guaranteeing descent for special iterates, called centers:

$$\begin{aligned} \mathbf{P} \,=\left\{ \begin{array}{l} \text{ a } \text{ convex } \text{ model }\, f^{\scriptscriptstyle M}_k,\, \hbox {possibly different from the cutting-plane one,}\\ \text{ a } \text{ measure } \text{ of } \text{ progress } \text{ to } \text{ stop } \text{ the } \text{ algorithm, }\\ \text{ a }\, stability center\, \hat{u}_k,\, \hbox {a past iterate deemed ``good enough''},\\ \text{ a } \text{ proximal } \text{ stabilization }\, \frac{1}{2t_k}|\cdot -\hat{u}_k|^2,\\ \text{ other } \text{ parameters } \text{ and } \text{ updating } \text{ rules, } \text{ including }\, t_{k+1}, f^{\scriptscriptstyle M}_{k+1}(\cdot ),\, \hbox {etc.} \end{array} \right\} . \end{aligned}$$

The quadratic norm, or proximal stabilization, is replaced by more general terms in [7]. In all these methods, \(u_{k+1}\) is the unique minimum of a stabilized model function:

$$\begin{aligned} \min _{u\in \mathbb {R}^n} f^{\scriptscriptstyle S}_k(u),\quad \text{ for } \,f^{\scriptscriptstyle S}_k(u) :=f^{\scriptscriptstyle M}_k(u)+\frac{1}{2t_k}|u-\hat{u}_k|^2. \end{aligned}$$
(3.3)

When the problem (1.1) is constrained by a simple polyhedron as in Example 2.3, the minimization in (3.3) incorporates this feasible set; see Sect. 7.3.

The model in P  does not need to be one based on cutting planes, although the relation

$$\begin{aligned} f^{\scriptscriptstyle M}_k(\cdot )\leqslant \check{f}_k(\cdot ) \quad \text{ for } \text{ all }\, k, \end{aligned}$$
(3.4)

generally holds, for example when compressing the bundle \(J_k\); see Sects. 5.2 and 7.1. Traditional bundle methods work with polyhedral approximations; in eigenvalue optimization, the spectral bundle methods [11, 18] use non-polyhedral models. The main requirement on the choice of \(f^{\scriptscriptstyle M}_k(\cdot )\) is pragmatic: problem (3.3) should be easily solvable. Relevant assumptions will be given later, as the need arises; for now we just mention that for lower oracles like in (2.3), taking \(f^{\scriptscriptstyle M}_k(\cdot ) :=\check{f}_k(\cdot )\) gives

$$\begin{aligned} f^{\scriptscriptstyle M}_k(\cdot )\leqslant f(\cdot )\quad \text{ for } \text{ all }\, k. \end{aligned}$$
(3.5)

When this inequality holds we shall say that a lower model is available.

3.2 Ensuring descent for the center subsequence: the set D

We single out from the set P the criterion to decide when an iterate becomes the next center, and write it as a rule depending on objects specified by a second set, checking if the iterate provides sufficient descent. The resulting set D is defined first for exact oracles.

3.2.1 Exact oracles

When the oracle calculations are exact, descent is determined by observing progress towards the goal of minimizing the objective function. Progress can be measured relative to either the model, or some nominal reference value, or the objective function itself. The three corresponding measures are respectively called and denoted by

$$\begin{aligned} model\,\, decrease\,\, \delta ^{\scriptscriptstyle M}_k, nominal\,\, decrease \,\, \delta ^{\scriptscriptstyle N}_k,\,\, \hbox {and}\,\, effective\,\, decrease\,\, \delta ^{\scriptscriptstyle E}_k. \end{aligned}$$

With exact oracles, the model decrease is \(\delta ^{\scriptscriptstyle M}_k= f(\hat{u}_k)- f^{\scriptscriptstyle M}_k(u_{k+1}).\) The nominal decrease, a non-negative measure because (3.3) minimizes the stabilized model \(f^{\scriptscriptstyle S}_k(\cdot )\), is given by

$$\begin{aligned} \delta ^{\scriptscriptstyle N}_k := \delta ^{\scriptscriptstyle M}_k -\frac{\alpha _k}{t_k}|u_{k+1}-\hat{u}_k|^2 \text{, } \text{ for } \text{ some }\, \alpha _k\in [0,1]. \end{aligned}$$
(3.6)

Finally, the effective decrease has the expression \(\delta ^{\scriptscriptstyle E}_k=f(\hat{u}_k)-f(u_{k+1}).\)

We shall see in Proposition 6.1 that driving the model decrease to zero is an important ingredient for the convergence analysis. The other two measures are involved in the rule deciding when the iterate is “good enough”. For \(u_{k+1}\) to become the next center, the difference \( f(u_{k+1})- f(\hat{u}_k)= -\delta ^{\scriptscriptstyle E}_k\) should be sufficiently negative. The descent test

$$\begin{aligned} m\,\delta ^{\scriptscriptstyle N}_k\leqslant \delta ^{\scriptscriptstyle E}_k \quad \text{ for } \text{ a } \text{ given } \text{ parameter } \, m\,\in (0,1), \end{aligned}$$
(3.7)

is used by the algorithm to decide between making

figure a

3.2.2 Inexact oracles

To make the rule (3.7) precise, the set of descent parameters should be

$$\begin{aligned} \mathbf{D} \,=\{m,\delta ^{\scriptscriptstyle N}_k\text{- } \text{ by } \text{ choosing }\, \alpha _k\, \hbox {and}\, \delta ^{\scriptscriptstyle M}_k\, \hbox {in}\, (3.6)\, \hbox {and}\,\delta ^{\scriptscriptstyle E}_k\}. \end{aligned}$$

When the oracle output has some error neither \(f(u_{k+1})\) nor \(f(\hat{u}_k)\) are available: only estimates \(f_{u_{k+1}}\) or \(f_{\hat{u}_k}\) are at hand and some representatives for the decrease measures need to be defined. Regarding the model decrease, we let

$$\begin{aligned} \delta ^{\scriptscriptstyle M}_k :=\ell _k-f^{\scriptscriptstyle M}_k(u_{k+1}), \end{aligned}$$
(3.8)

where the substitute for \(f(\hat{u}_k)\) is a “level” \(\ell _k\) chosen so that the inequality \(\ell _k\geqslant f_{\hat{u}_k}\) holds:

$$\begin{aligned} \left\{ \begin{array}{l@{\quad }l} \ell _k\in [f_{\hat{u}_k},f(\hat{u}_k)]&{} \text{ if }\,\, \eta ^g_u\equiv 0,\, \hbox {i.e., the oracle is of lower type, and}\\ \ell _k= f_{\hat{u}_k}&{} \text{ otherwise. } \end{array}\right. \end{aligned}$$
(3.9)

Theorem 4.5 below shows that the level, to be specified in the set P , is a natural estimate for the optimal value of (1.1).

Regarding the nominal decrease (3.6), only non-negative \(\delta ^{\scriptscriptstyle N}_k\) in (3.7) give centers with strictly decreasing function values. We shall see in Sect. 5 that in some situations, for example with upper oracles, to ensure \(\delta ^{\scriptscriptstyle N}_k\geqslant 0\) some corrective action, called of noise attenuation, needs to be introduced.

Finally, for the effective decrease only two choices have been proposed in the literature:

  • Observed decrease \(\delta ^{\scriptscriptstyle E}_k=f_{\hat{u}_k}-f_{u_{k+1}}\) as in [12, 15], regardless of the oracle noise.

  • Realistic decrease \(\delta ^{\scriptscriptstyle E}_k=\hat{f}_k-f_{u_{k+1}}\) as in [6, 8], for the “threshold” between centers

    $$\begin{aligned} \hat{f}_k :=\max \left\{ \begin{array}{ll}f_{\hat{u}_k}&{}\\ \max _{j\leqslant k} f^{\scriptscriptstyle M}_j(\hat{u}_k)&{}\text{ for } \text{ iterations }\, j\, \hbox {following the one generating}\, \hat{u}_k \end{array}\right\} . \end{aligned}$$
    (3.10)

    When (3.5) holds, the threshold takes a better account of reality than \(f_{\hat{u}_k}\). To compute it, one sets \(\hat{f}_k :=f^{\scriptscriptstyle M}_k({\hat{u}_k})\) after a descent step and \(\hat{f}_k=\max \{f^{\scriptscriptstyle M}_k({\hat{u}_k}), \hat{f}_{k-1}\}\) after a null step.

Of these two choices, the observed one is the only possible proposal when the oracle accuracy is not controllable and the error bound is unknown. Convergence (up to the model and oracle precision) can be shown in this case using our synthetic theory; see Sect. 7.2. On the other hand, the realistic decrease is appealing but, as shown in Sect. 7.1.3, to ensure convergence the oracle must be partly asymptotically exact with errors vanishing fast enough.

For a lower oracle with a known error bound (such as the controllable oracle in Example 2.2) and with a lower model, there is a third type of decrease:

  • Conservative decrease  \(\delta ^{\scriptscriptstyle E}_k= (f_{\hat{u}_k}+\bar{\eta }_{\hat{u}_k})- (f_{u_{k+1}}+\bar{\eta }_{u_{k+1}})\), involving the knowledge of the oracle error upper bounds \(\bar{\eta }_{\hat{u}_k}\) and \(\bar{\eta }_{u_{k+1}}\). Thanks to this additional information, the corresponding new method, the Controllable Bundle Algorithm 5.4, eventually solves (1.1) up to the oracle accuracy at descent steps; see Corollary 6.12.

4 Main ingredients in the algorithm

Coping with inexact oracles increases substantially the technicalities. To clarify notation, we adopt the convention that a superscript \((\cdot )^{\scriptscriptstyle M}\) [resp. \((\cdot )^{\scriptscriptstyle S}\), resp. a hat \(\widehat{(\cdot )}\)] connotes an item attached to the original model [resp. the objective function in (3.3), resp. the stability center].

4.1 Aggregate objects and algorithmic pattern

Once (3.3) is solved to produce the next iterate, two key objects are the aggregate linearization \(f^{\scriptscriptstyle L}_{-k}(\cdot )\) and aggregate subgradient \(\hat{g}_k\) introduced below.

Lemma 4.1

(Aggregate objects) If the model \(f^{\scriptscriptstyle M}_k(\cdot )\) is convex,

$$\begin{aligned} u_{k+1}=\hat{u}_k-t_k\hat{g}_k,\quad \text{ for } \text{ some }\, \hat{g}_k\in \partial f^{\scriptscriptstyle M}_k(u_{k+1}) \end{aligned}$$
(4.1)

is the unique solution of the master-program (3.3) and the affine function

$$\begin{aligned} f^{\scriptscriptstyle L}_{-k}(u) :=f^{\scriptscriptstyle M}_k(u_{k+1})+\left\langle \hat{g}_k ,u-u_{k+1}\right\rangle \end{aligned}$$
(4.2)

is an underestimate of the model: \(f^{\scriptscriptstyle L}_{-k}(\cdot )\leqslant f^{\scriptscriptstyle M}_{k}(\cdot )\).

Proof

Since its objective function is finite-valued and strongly convex, (3.3) has a unique solution characterized by the optimality condition \(0\in \partial f^{\scriptscriptstyle S}_k(u)= \partial f^{\scriptscriptstyle M}_k(u)+(u-\hat{u}_k)/t_k\), which gives (4.1). The inequality \(f^{\scriptscriptstyle L}_{-k}(\cdot )\leqslant f^{\scriptscriptstyle M}_k(\cdot )\) is just the subgradient relation.\(\square \)

As a consequence of (4.1) the nominal decrease in (3.8) has the equivalent expressions

$$\begin{aligned} \delta ^{\scriptscriptstyle N}_k=\delta ^{\scriptscriptstyle M}_k-\alpha _k t_k|\hat{g}_k|^2 =[\ell _k-f^{\scriptscriptstyle M}_k(u_{k+1})]-\alpha _k t_k |\hat{g}_k|^2\, \text{ for } \text{ some }\, \alpha _k\geqslant 0. \end{aligned}$$
(4.3)

Without further specification of the sets P  and D , an abstract algorithmic pattern, depending on these objects, can now be outlined.

Algorithm Pattern 4.2 (P, D)

Having P  and D , a starting point \(u_1\) is chosen. The oracle output \((f_{u_1},g_{u_1})\) is available. Set \(k=1\) and initialize \(\hat{u}_1=u_1\).

Step 1. :

Having the model and \(t_k>0\) defined by P , solve (3.3) to obtain \(\hat{g}_k\), \(u_{k+1}\) and \(f^{\scriptscriptstyle L}_{-k}(\cdot )\) as in Lemma 4.1. Based on the definitions in D , compute the nominal decrease (4.3) and determine the need of noise attenuation.Loop in Step 1 until noise needs no more attenuation.

Step 2. :

Call the oracle at \(u_{k+1}\) to obtain the output \((f_{u_{k+1}},g_{u_{k+1}})\).

Step 3. :

With the objects in D , the descent test (3.7) decides between making

figure b
Step 4. :

Increase \(k\) by 1 and loop to Step 1. \(\square \)

As our aim is to state general principles for convergence, this is an algorithmic pattern, independent of the bundle variant and its particular choices for P and D . The convergence analysis in Sect. 6 is done without specifying these sets. Section 7 reviews several instances,

$$\begin{aligned} \hbox {Specific Algorithm} = \hbox {Alg. Pattern} 4.2 (\mathbf{P} ,\mathbf{D} )\,\text{ for } \text{ specific }\,\, \mathbf{P} ,\mathbf{D} . \end{aligned}$$

For the new Algorithm 5.4, concrete sets P and D are given in Sect. 5.2.

Step 1 will be endowed with a stopping test based on the approximate optimality inequality (4.12) below. Also, Step 1 may also need to incorporate a loop of noise attenuation. This mechanism ensures eventual non-negativity of \(\delta ^{\scriptscriptstyle N}_k\), so that the rule (3.7) resembles a descent test. We shall see in Sect. 5 how to deal with this issue by modifying Step 1 as in (5.5); if the noise attenuation loop in the reformulated Step 1 is infinite, Corollary 5.3 shows optimality of the current center (up to the model and oracle precision).

Naturally, the model cannot be totally arbitrary: Step 3 (right branch) imposes satisfaction of lower bounds. Upper bounds are also required; they are somewhat linked with the choice of \(t_k\) and will arise in Sect. 6.2, specifically in relation (6.5) therein.

4.2 Measuring optimality

To establish approximate optimality conditions it is best to refer information to the center \(\hat{u}_k\).

4.2.1 Shifting the bundle information

With exact oracles the origin is translated to the stability elements \((\hat{u}_k,f(\hat{u}_k))\). For inexact oracles, the level \(\ell _k\) replaces \(f(\hat{u}_k)\) and defines the aggregate linearization gap, illustrated in Fig. 1 for a lower model and oracle:

$$\begin{aligned} \hat{e}_k :=\ell _k-f^{\scriptscriptstyle L}_{-k}(\hat{u}_k). \end{aligned}$$
(4.4)
Fig. 1
figure 1

Translation of the origin to the stability elements and aggregate gap \(\hat{e}_k\)

Combining (4.4), (4.2) and (4.1) gives for the model decrease (3.8) the expression

$$\begin{aligned} \delta ^{\scriptscriptstyle M}_k=\hat{e}_k+t_k|\hat{g}_k|^2, \end{aligned}$$
(4.5)

traditionally used as an optimality criterion with exact oracles. As explained next, small values for \(\delta ^{\scriptscriptstyle M}_k\) can be deceiving when the oracle delivers inexact information.

Remark 4.3

(Noise Attenuation) Recall from (3.9) that for lower oracles the level can be any value between \( f_{\hat{u}_k}\) and \(f({\hat{u}_k})\). For the function in Fig. 1, if the level is \(\ell _k=f_{\hat{u}_k}\) the aggregate gap becomes negative and in (4.5) the model decrease may get close to zero. However, the figure shows that the center \(\hat{u}_k\) is far from near optimality: rather than stopping, the algorithm should look for iterates further from \(\hat{u}_k\). This is the basis of noise attenuation to cope with negative gaps, a track initiated by [12], deeply developed in [15] and described in (5.5) below: increasing \(t_k\) diminishes the attraction toward the suspect center; and (4.5) shows that \(\delta ^{\scriptscriptstyle M}_k\) increases (unless \(\hat{g}_k\) vanishes).\(\square \)

The identity \(f^{\scriptscriptstyle M}_k(u_{k+1})=f^{\scriptscriptstyle L}_{-k}(\hat{u})-\left\langle \hat{g}_k ,\hat{u}_k-u_{k+1}\right\rangle \) from (4.2), gives for the aggregate linearization the equivalent expression

$$\begin{aligned} f^{\scriptscriptstyle L}_{-k}(u)=\ell _k-\hat{e}_k+\left\langle \hat{g}_k ,u-\hat{u}_k\right\rangle , \end{aligned}$$
(4.6)

from which follow the useful relations below, derived from (4.1):

$$\begin{aligned} \hat{e}_k= \ell _k-f^{\scriptscriptstyle M}_k(u_{k+1})-\left\langle \hat{g}_k ,\hat{u}_k-u_{k+1}\right\rangle =\ell _k-f^{\scriptscriptstyle M}_k(u_{k+1})-t_k|\hat{g}_k|^2. \end{aligned}$$
(4.7)

At this point we introduce an important convergence parameter:

$$\begin{aligned} \phi _k :=\hat{e}_k+\left\langle \hat{g}_k,\hat{u}_k\right\rangle = \ell _k-f^{\scriptscriptstyle L}_{-k}(\hat{u}_k)+\left\langle \hat{g}_k,\hat{u}_k\right\rangle . \end{aligned}$$
(4.8)

As shown in Theorem 4.5 below, proving convergence for an algorithm amounts to

$$\begin{aligned} \text{ finding } \text{ a }\, {K^\infty }-\hbox {subsequence}\,\,\{(\phi _k,\hat{g}_k)\}\, \hbox {converging to}\, (\phi ,0)\, \hbox {with}\,\, \phi \leqslant 0, \end{aligned}$$
(4.9)

for certain infinite iteration sets \({K^\infty }\) generated by the considered algorithm, so the stopping criterion in the set P checks that both \(\phi _k\) and \(\hat{g}_k\) are sufficiently small.

Remark 4.4

(Duality gap interpretation of \(\phi _k\)) For the primal problem (2.4) in Example 2.1, (1.1) comes from Lagrangian relaxation or column generation and \(\phi _k\) has a nice interpretation as a duality gap. Recall that the dual function is given by

$$\begin{aligned} f(u) :=\max _{x\in X}\,L(x,u),\quad \text{ with } \, L(x,u) :=h(x)+\left\langle u, c(x)\right\rangle . \end{aligned}$$

Calling \(x_j\) the primal point computed at \(u_j\), the oracle output is

$$\begin{aligned} f_{u_j}=L(x_j,u_j)=h(x_j)+\left\langle u_j,c(x_j)\right\rangle \quad \text{ and }\quad g_{u_j}=c(x_j). \end{aligned}$$

The Lagrangian is affine with respect to \(u\), so \(f^{\scriptscriptstyle L}_j(\cdot )=L(x_j,\cdot )\). Moreover, when the model is the cutting-plane function, a subgradient \(\hat{g}_k\) of the max-function \(f^{\scriptscriptstyle M}_k(\cdot )\) is a convex combination of active subgradients. Hence, for \(\lambda _j\geqslant 0\) such that \(\sum _j\lambda _j=1\), \(\hat{g}_k=\sum _j\lambda _jg_{u_j}=\sum _j\lambda _jc(x_j),\) and for each \(j\) such that \(\lambda _j>0\), \(f_k^{\scriptscriptstyle M}(u_{k+1})=f^{\scriptscriptstyle L}_j(u_{k+1})=L(x_j,u_{k+1})= h(x_j)+\left\langle u_{k+1}, c(x_j)\right\rangle .\) Thus,

$$\begin{aligned} \begin{array}{llll} f_k^{\scriptscriptstyle M}(u_{k+1}) &{}=&{} \sum _j\lambda _jh(x_j)+\left\langle u_{k+1},\sum _j\lambda _jc(x_j)\right\rangle &{} [\hbox {by convex combination}] \\ &{}=&{} \sum _j\lambda _jh(x_j)+\left\langle u_{k+1},\hat{g}_k \right\rangle &{} [\hbox {by the expression above for}\, \hat{g}_k] \\ &{}=&{} \ell _k-\hat{e}_k-\left\langle \hat{g}_k ,\hat{u}_k-u_{k+1}\right\rangle \, &{} [\hbox {by}\, (4.7)\, \hbox {and}\, (4.1)] \end{array} \end{aligned}$$

Using the first identity in (4.8), we have therefore proved that

$$\begin{aligned} \sum _j\lambda _jh(x_j)=\ell _k-\hat{e}_k-\left\langle \hat{g}_k,\hat{u} _k\right\rangle = \ell _k-\phi _k. \end{aligned}$$
(4.10)

Assume for simplicity that (2.4) is a linear program: \(h(x)=\left\langle h, x\right\rangle \), \(c(x)=Ax-a\), and introduce the primal point \(\hat{x}_k :=\sum _j\lambda _jx_j\). By (4.10), \(\left\langle h,\hat{x}_k\right\rangle =\ell _k-\phi _k\) and, as \(\ell _k\) from (3.9) is essentially a functional dual value,

$$\begin{aligned} \phi _k&= \ell _k-\left\langle h,\hat{x}_k\right\rangle \, \text{ estimates } \text{ the } \text{ duality } \text{ gap } \text{ and }\\&\quad \hat{g}_k=A(\hat{x}_k-a)\, \text{ is } \text{ a } \text{ constraint } \text{ value. } \end{aligned}$$

So both \(\phi _k\) and \(\hat{g}_k\) should be driven to zero (at least with exact oracles, having \(\ell _k=f(\hat{u}_k)\)).\(\square \)

4.2.2 Convergence: what it means

The following weakened form of (3.5), extends (3.2) to a general model:

$$\begin{aligned} \text{ for } \text{ some }\,\eta ^{\scriptscriptstyle M}\geqslant 0\,\text{ the } \text{ inequality }\, f^{\scriptscriptstyle M}_k(\cdot )\leqslant f(\cdot )+ \eta ^{\scriptscriptstyle M}\,\text{ holds } \text{ for } \text{ all }\,k, \end{aligned}$$
(4.11)

The condition is automatic for lower oracles and models (with \(\eta ^{\scriptscriptstyle M}\equiv 0\)) and for upper oracles, it determines the degree of agreement between the model and the true function.

Theorem 4.5

(Conditions for approximate optimality) Suppose the model satisfies (4.11). If the Algorithmic Pattern 4.2 generates a \({K^\infty }\)-subsequence \(\{(\phi _k,\hat{g}_k)\}\) satisfying (4.9), then the level defined in (3.9) eventually estimates the infimal value of \(f\). Namely,

$$\begin{aligned} \limsup _{k\in {K^\infty }}\,\ell _k \leqslant \inf f(\cdot )+\phi +\eta ^{\scriptscriptstyle M}\leqslant \inf f(\cdot )+\eta ^{\scriptscriptstyle M}. \end{aligned}$$
(4.12)

Consider an index set \(K'\subset {K^\infty }\) such that \(\{\hat{u}_k\}_{K'}\) has a limit \(\hat{u}\) and define the corresponding asymptotic oracle error at descent steps

$$\begin{aligned} \eta ^\infty := \liminf _{k\in K'} \eta _{\hat{u}_k}. \end{aligned}$$
(4.13)

Then (4.9) implies that \(\hat{u}\) is an \((\eta ^\infty +\eta ^{\scriptscriptstyle M})\)-solution to problem (1.1).

Proof

Use Lemma 4.1 and (4.11) in (4.6): for all \(u\) and all \(k\),

$$\begin{aligned} f(u)+\eta ^{\scriptscriptstyle M}\geqslant f^{\scriptscriptstyle M}_k(u)\geqslant f^{\scriptscriptstyle L}_{-k}(u)= \ell _k-\phi _k+\left\langle \hat{g}_k, u\right\rangle . \end{aligned}$$

Hence \(\left\langle \hat{g}_k, u\right\rangle -f(u) \leqslant \phi _k-\ell _k+\eta ^{\scriptscriptstyle M}\); take the supremum over \(u\) to obtain that

$$\begin{aligned} \sup _u \{\left\langle \hat{g}_k, u\right\rangle -f(u)\}=:f^*(\hat{g}_k)\leqslant \phi _k-\ell _k+\eta ^{\scriptscriptstyle M}. \end{aligned}$$

In view of (4.9), by closedness of the conjugate \(f^*(\cdot )\),

$$\begin{aligned} f^*(0)\leqslant \liminf _{k\in {K^\infty }}f^*(\hat{g}_k)\leqslant \lim _{k\in {K^\infty }}\phi _k+\liminf _{k\in {K^\infty }}{(-\ell _k)}+\eta ^{\scriptscriptstyle M}=\phi -\limsup _{k\in {K^\infty }}\,\ell _k+\eta ^{\scriptscriptstyle M}, \end{aligned}$$

which is just (4.12) since the value of the conjugate function at zero satisfies \(f^*(0)=\inf f(\cdot )\). To see the final statement, given the iterate subsequence defined over \(K'\subset K\), consider \(k\in K'\) and use (3.9) in the first line of (1.2), written at \(u=\hat{u}_k\): \( f(\hat{u}_k)-\eta _{\hat{u}_k}=f_{\hat{u}_k}\leqslant \ell _k\). Passing to the limit yields the desired relation, by lower semicontinuity of \(f(\cdot )\):

$$\begin{aligned} f(\hat{u})-\eta ^\infty \leqslant \limsup _{k\in K'}{[f(\hat{u}_k)-\eta _{\hat{u}_k}]} \leqslant \limsup _{k\in K'}\,\ell _k \leqslant \inf f(\cdot )+\eta ^{\scriptscriptstyle M}. \end{aligned}$$

\(\square \)

Instead of the asymptotic condition (4.9), seemingly introduced for the first time in [15], convergence for bundle methods has always been established by

$$\begin{aligned} \text{ finding } \text{ a }\, {K^\infty }-\hbox {subsequence}\, \{(\hat{e}_k,\hat{g}_k)\}\, \hbox {converging to (0,0)}. \end{aligned}$$
(4.14)

The difference is subtle indeed: both properties are equivalent when \(\{\hat{u}_k\}\) is bounded; precisely, condition (4.9) gives an elegant argument in the unbounded case, which was overlooked before, for example in [3, 13].

Theorem 4.5 gives some insight on the role played by \(\ell _k\), in particular on the reason for its definition (3.9). The aim of the Algorithmic Pattern 4.2 is of course to estimate as accurately as possible the optimal value and a solution of (1.1). The latter is done by means of the stability center \(\hat{u}_k\) while the former can be accomplished in various ways. A straightforward approximation for the optimal value is \(f_{\hat{u}_k}\), but better can be done when both the oracle and the model are of lower type (both (2.3) and (3.5) hold). In this case, the value \(\hat{f}_k\) from (3.10) is more accurate than \(f_{\hat{u}_k}\) and having in (4.12) that \(\ell _k=\hat{f}_k\) ensures by (3.10) that the estimate is the largest available functional value. Nevertheless, notice that such definition for the level is acceptable only if \(\hat{f}_k\) satisfies the relations in (3.9);

$$\begin{aligned} \hat{f}_k\in [f_{u_k},f(u_k)]\,\text{ for } \text{ lower } \text{ oracles } \text{ and }\,\hat{f}_k=f_{u_k}\,\text{ otherwise. } \end{aligned}$$

The inequality \(\hat{f}_k\geqslant f_{\hat{u}_k}\) always holds by the definition (3.10). For lower oracles and lower models, (3.5) ensures in addition that \(\hat{f}_k\leqslant f(\hat{u}_k)\) and, hence, it is possible to set \(\ell _k=\hat{f}_k\).

Theorem 4.5 also clarifies how a “partly asymptotically exact” lower oracle can yield exact optima in the limit, without any error in spite of the oracle inexactness. If the oracle is eventually exact only at descent steps, \(\eta ^\infty =0\) and \(f(\hat{u}_k)-f_{\hat{u}_k}\rightarrow 0\), so (3.9) implies \(\ell _k-f(\hat{u}_k)\rightarrow 0\). If, in addition, the model is of lower type, (3.5) gives \(\eta ^{\scriptscriptstyle M}=0\) in (4.11) and in (4.12) the only possible value for \(\phi \) is zero (like in (4.14)).

5 On noise management and a concrete instance

We now make precise the noise attenuation loop in Step 1. We also provide a particular instance for the sets P and D in the Algorithmic Pattern 4.2, the Controllable Bundle Algorithm 5.4, a new method with the ability of managing the oracle accuracy.

5.1 Properties of the aggregate gap

The rewriting (4.5) suggests that a small \(\delta ^{\scriptscriptstyle M}_k\) might bring closer the traditional convergence property (4.14) for “not too negative” gaps. The relevance of the sign of \(\hat{e}_k\) was also noticed empirically for Fig. 1 in Remark 4.3. The noise attenuation mechanism will be triggered when \(\hat{e}_k\) becomes negative. Below we show how the gap sign relates with objects in the set D , in particular with the sign of the nominal decrease \(\delta ^{\scriptscriptstyle N}_k\), whose non-negativity is fundamental for (3.7) to be a genuine descent test.

Lemma 5.1

(Aggregate gap properties relevant for noise detection) In the Algorithmic Pattern 4.2, consider the level, best function value, model and nominal decreases, and gap defined, respectively, in (3.9), (3.10), (3.8), (4.3), and (4.4). The following holds.

  1. (i)

    \(\hat{e}_k\geqslant \ell _k-\hat{f}_k\) at all iterations.

  2. (ii)

    Satisfaction of the inequality

    $$\begin{aligned} \hat{e}_k\geqslant -\beta _k t_k|\hat{g}_k|^2\,\, \text{ for } \text{ some }\, \beta _k\in \;[b,1-b]\, \hbox {with}\, b\in (0,\frac{1}{2}], \end{aligned}$$
    (5.1)

    is equivalent to any of the relations below

    $$\begin{aligned} \delta ^{\scriptscriptstyle M}_k\geqslant \max \left\{ \hat{e}_k,(1-\beta _k) t_k|\hat{g}_k|^2\right\} \iff \delta ^{\scriptscriptstyle N}_k\geqslant \Bigl (1-(\alpha _k+\beta _k)\Bigr ) t_k|\hat{g}_k|^2. \end{aligned}$$
    (5.2)

    In particular, whenever (5.1) holds the model decrease is non-negative. Furthermore, if we assume in (4.3) and (5.1) that

    $$\begin{aligned} \alpha _k+\beta _k\leqslant 1-b, \end{aligned}$$
    (5.3)

    whenever (5.1) holds the nominal decrease is non-negative too.

  3. (iii)

    If the model satisfies (4.11) then \(\hat{e}_k\geqslant -\bigl (\eta _{\hat{u}_k}+\eta ^{\scriptscriptstyle M}\bigr )\).

Proof

By definition (4.4) and the model subgradient inequality in Lemma 4.1,

$$\begin{aligned} \hat{e}_k = \ell _k-f^{\scriptscriptstyle L}_{-k}(\hat{u}_k) \geqslant \ell _k- f^{\scriptscriptstyle M}_k(\hat{u}_k). \end{aligned}$$
(5.4)

The first item follows from adding \(\pm \hat{f}_k\) to the right hand side, recalling the definition for \(\hat{f}_k\) in (3.10). The second item follows from some simple algebra using (4.5) and (4.3).For the third item, use the level definition (3.9) and the model assumption (4.11) in (5.4) to write \(\hat{e}_k \geqslant f_{\hat{u}_k}- f^{\scriptscriptstyle M}_k(\hat{u}_k) \geqslant f_{\hat{u}_k}- f(\hat{u}_k)-\eta ^{\scriptscriptstyle M}\) and (2.3) ends the proof.\(\square \)

The allowed interval for \(\beta _k\) in (5.1), together with (5.3), implies that \(\alpha _k\in [0,1]\), so the conditions are consistent with those required for \(\alpha _k\) in (3.6).

In view of item (ii) in Lemma 5.1, the inequality (5.1) can be used to detect when the nominal decrease is negative and a corrective action needs to be taken. This test will be incorporated in the Algorithmic Pattern 4.2 as follows. Step 1.  Having the model and \(t_k\) defined by P , solve (3.3) to obtain \(\hat{g}_k\),

\(u_{k+1}\) and \(f^{\scriptscriptstyle L}_{-k}(\cdot )\) as in Lemma 4.1, and compute \(\phi _k\) from (4.8).

Stop if \(\phi _k\) and \(|\hat{g}_k|\) are both small enough. Otherwise, with the elements in D ,

compute the gap (4.4) and determine the need of noise attenuation:

(5.5)

When the test in Step 1 determines that noise became too large the stepsize \(t_k\) is sharply increased, inhibiting any decrease until the next descent step; see (6.14). In Corollary 5.3 we show that this simple mechanism ensures that either the current center is approximately optimal or the noise attenuation loop is finite and the algorithm proceeds to Step 2. First, we make use of the aggregate gap properties in Lemma 5.1 to examine for which models and oracles the noise attenuation test (5.1) can be dismissed.

Corollary 5.2

(Lower models and various oracles) Suppose Algorithmic Pattern 4.2 with Step 1 from (5.5) uses a lower model: in the set P the model satisfies (3.5).

  1. (i)

    There is no need of noise attenuation and all iterations satisfy automatically (5.1) with \(\beta _k\) arbitrary, \(b=0\), so \(\alpha _k\in [0,1]\) in (4.3), whenever one of the conditions below hold.

    1. (ia)

      In the parameter set the level from (3.9) is given by \(\mathbf{P} \,\ni \ell _k :=\hat{f}_k\,\text{ from }\,(3.10)\) (recall that such a definition is possible in particular for lower oracles and models: both (2.3) and (3.5) hold).

    2. (ib)

      The oracle is exact: (2.3) \(\text{ holds } \text{ with }\,\eta _u\equiv 0\).

  2. (ii)

    The noise attenuation loop is finite if the oracle is lower and asymptotically exact at descent steps:

    $$\begin{aligned} (2.3)\, \hbox {holds and in}\, (4.13)\, \eta _{\infty }=0\, \hbox {with}\, K' :=\{k:\text{ the } \text{ descent } \text{ test } \text{(3.7) } \text{ holds } \}. \end{aligned}$$
    (5.6)

Proof

Condition (ia) implies (5.1), by Lemma 5.1(i) and (ii). When the oracle is exact, (ia) gives the result, because \(\ell _k =f(\hat{u}_k)\) in (3.9) and, if the model is lower, \(\hat{f}_k=f(\hat{u}_k)\) in (3.10). Similarly for (ii), reasoning asymptotically for \(k\) satisfying (3.7).\(\square \)

When only finitely many descent steps are generated (\(K'\) is finite), the condition \(\eta _{\infty }=0\) in (5.6) in fact requires an exact evaluation of descent steps.

For general models, the noise attenuation loop can be infinite and the Algorithmic Pattern 4.2 may never reach Step 2. We use the important Theorem 4.5 to show that in this case the current center (the last descent step) is an approximate solution to (1.1).

Corollary 5.3

(Upper oracles and noisy steps) Consider Algorithmic Pattern 4.2 with Step 1 from (5.5) and assume in the set P the model satisfies (4.11).

If for some iteration \(\hat{k}\) the algorithm loops forever in Step 1, then \(t_k\uparrow \infty \) and the set \({K^\infty }:=\{k\geqslant \hat{k}:\, \text{ condition } \text{(5.1) } \text{ does } \text{ not } \text{ hold }\}\) is infinite. As a result, the last descent iterate \(\hat{u} :=\hat{u}_{\hat{k}}\) is an \((\eta _{\hat{u}_{\hat{k}}}+\eta ^{\scriptscriptstyle M})\)-solution to (1.1).

Proof

In the noise attenuation loop, \(t_k\) is increased and the stability center is maintained fixed to \(\hat{u}\). The \({K^\infty }\)-sequence of aggregate gaps \(\{\hat{e}_k\}\) is bounded below by Lemma 5.1(iii), and non-positive because (5.1) does not hold. Since the \({K^\infty }\)-iterates satisfy the negation of (5.1) and \(\beta _k\geqslant b>0\) therein,

$$\begin{aligned} |\hat{g}_k|^2< -\frac{\hat{e}_k}{\beta _k t_k}\leqslant \frac{\eta _{\hat{u}_k}+\eta ^{\scriptscriptstyle M}}{b\, t_k}\,; \end{aligned}$$

hence, \(\hat{g}_k\rightarrow 0\) as \(t_k\) is driven to infinity for \(k\in {K^\infty }\). Therefore, the limit of the \({K^\infty }\)-subsequence \( \{\phi _k=\hat{e}_k+\left\langle \hat{g}_k,\hat{u}\right\rangle \}\) is \(\phi =\lim _{K^\infty }\hat{e}_k\leqslant 0\), because (5.1) does not hold. So the convergence condition (4.9) is satisfied, Theorem 4.5 applies, and the desired result follows.\(\square \)

The analysis above shows that when the Algorithmic Pattern 4.2 performs Step 1 as in (5.5), for any oracle and with a model satisfying (the very reasonable) assumption (4.11),

  • either for all iterations the loop in Step 1 ends with an iterate satisfying (5.1), for which a descent or a null step will be made in Step 3;

  • or for some iteration the loop in Step 1 in infinite and \(\hat{u}\) is optimal up to the oracle and model precision.

For this reason the convergence analysis in Sect. 6 only considers infinite sequences stemming from the alternative in Step 3 in the Algorithmic Pattern 4.2: infinitely many descent steps or an infinite tail of consecutive null steps.

5.2 Controllable bundle method

We now state a concrete algorithm for controllable lower oracles: like in Example 2.2, high accuracy is possible yet the heavy computational burden makes it preferable to avoid exact calculations, The main novelty for this variant is in the specific choice of the level and the accuracy control of the oracle.

We assume there is an informative controllable oracle of lower type: in (1.2) \(\eta ^g_{u}\equiv 0\), so (2.3) holds. Also, the input error bound \(\bar{\eta }_{u_k}\) is sent to the oracle together with the evaluation point \(u_k\) to obtain \(f_{u_k}\in [f(u_k)-\bar{\eta }_{u_k}, f(u_k)]\) and an approximate subgradient \(g_{u_k}\).

The parameter set is given by

$$\begin{aligned} \mathbf{P} =\left\{ \begin{array}{l} \text{ a } \text{ cutting-plane } \text{ model }\,f^{\scriptscriptstyle M}_k(\cdot )\\ \quad = \max \left\{ f^{\scriptscriptstyle L}_j(\cdot ) : j \in {J}_k\subseteq \{-(k-1)\}\cup \{1,\ldots , k\} \right\} \\ \text{ a } \text{ stopping } \text{ test } \text{ checking } \text{ if }\, \phi _k\, \hbox {and}\, \hat{g}_k\, \hbox {are sufficiently small};\\ \text{ the } \text{ proximal } \text{ stabilization }\, \frac{1}{2t_k}|\cdot -\hat{u}_k|^2\, \hbox {and an updating rule for}\, t_k:\\ \quad \text{- } \text{ if } \text{ descent } \text{ step, }\, t_{k+1}\geqslant t_k \\ \quad \text{- } \text{ if } \text{ null } \text{ step, }\, t_{k+1}=\max \{t_{\mathtt{low}},\,\sigma t_k\}\, \hbox {for}\, \sigma \in (0,1]\, \hbox {and}\, t_{\mathtt{low}}>0\\ \text{ the } \text{ current } \text{ stability } \text{ center }\, \hat{u}_k\, \hbox {and its level}\, \ell _k :=\hat{f}_k\, \hbox {from} (3.10)\\ \text{ a } \text{ rule } \text{ to } \text{ update } \text{ the } \text{ oracle } \text{ error } \text{ bound: }\, \bar{\eta }_{u_{k+1}} = \bar{\eta }_{\hat{u}_k} + f_{\hat{u}_k} -\ell _k. \end{array} \right\} . \end{aligned}$$
(5.7)

The model function is not the pure cutting-plane model, whose index set is \(J_k=\{1,\ldots ,k\}\). Here, the aggregate linearization enters the index set, so the polyhedral model allows for bundle compression, and (3.4) is satisfied.

The descent set has the elements:

$$\begin{aligned} \mathbf{D} \,=\left\{ \begin{array}{l} m\in (0,1), \alpha _k\in [ 0,1]\\ \delta ^{\scriptscriptstyle N}_k := \ell _k - f^{\scriptscriptstyle M}_k(u_{k+1})- \alpha _k t_k|\hat{g}_k|^2\\ \delta ^{\scriptscriptstyle E}_k := f_{\hat{u}_k} + \bar{\eta }_{\hat{u}_k} - f_{u_{k+1}} - \bar{\eta }_{u_{k+1}}. \end{array}\right\} . \end{aligned}$$
(5.8)

Algorithm 5.4

Controllable Bundle Method (=Alg.Pattern 4.2 with P ,D from (5.7), (5.8))

The user chooses the starting point \(u_1\) and \(t_1\geqslant t_{\mathtt{low}}\). The oracle output \((f_{u_1},g_{u_1})\) is available. Set \(k=1\) and initialize \(\hat{u}_1=u_1\), \(J_1=\{1\}\), and \(\ell _1=f_{u_1}\).

Step 1.:

Obtain \(\hat{g}_k\), \(u_{k+1}\) and \(f^{\scriptscriptstyle L}_{-k}(\cdot )\) from Lemma 4.1. by solving the quadratic program

$$\begin{aligned} \min _{r,u}\,r+\frac{1}{2t_k}|u-\hat{u}_k|^2\; \text{ s.t. } \; r\geqslant f^{\scriptscriptstyle L}_j(u), j \in {J}_k. \end{aligned}$$

Compute \(\phi _k\) as in (4.8); if \(\phi _k\) and \(|\hat{g}_k|\) are small enough, stop.

Step 2.:

Update the oracle error bound \(\bar{\eta }_{u_{k+1}} = \bar{\eta }_{\hat{u}_k} + f_{\hat{u}_k} -\ell _k\) and call the oracle with input \((u_{k+1},\bar{\eta }_{u_{k+1}})\) to obtain the output \((f_{u_{k+1}},g_{u_{k+1}})\).

Step 3.:

Check the descent test (3.7), or the equivalent relation:

$$\begin{aligned} f_{u_{k+1}} + \bar{\eta }_{u_{k+1}}\leqslant f_{\hat{u}_k} + \bar{\eta }_{\hat{u}_k} - m(\ell _k - f^{\scriptscriptstyle M}_k(u_{k+1}){- \alpha _k t_k|\hat{g}_k|^2}) \end{aligned}$$

and perform one of the steps below:

figure c
Step 4.:

Increase \(k\) by 1 and loop to Step 1. \(\square \)

The level choice satisfies \(\ell _k=\hat{f}_k\geqslant f_{\hat{u}_k}\) by (3.10). Also, note that the oracle accuracy is automatically adjusted, a useful feature if the initial bound \(\bar{\eta }_{u_1}\) was taken too large. Nevertheless, the update in Step 2 of Algorithm 5.4, forces the error bound sequence \(\{\bar{\eta }_{u_k}\}\) to be nonincreasing, so if the user chooses to start with exact calculations \((\bar{\eta }_{u_1}=0)\), the algorithm boils down to the classical proximal bundle method for exact oracles. By Remark 6.8 and Corollary 6.12, having nonincreasing accuracy at all iterations is crucial for proving convergence of the method. The partly asymptotically exact version in Sect. 7.1.4 drives \(\eta _{\hat{u}_k} \rightarrow 0\), by taking in (5.8) the conservative decrease \(\delta ^{\scriptscriptstyle E}_k := f_{\hat{u}_k} + \bar{\eta }_{\hat{u}_k} - f_{u_{k+1}}\).

6 Convergence analysis

Our main purpose is to state a synthetic convergence theory for the Algorithmic Pattern 4.2, without stating neither the parameter set nor the descent rule. References to Algorithm 5.4 will be given throughout this section to make statements more concrete and guide the reader until the final Theorem 6.11. This general result is suitable for showing convergence of numerous bundle methods, including the inexact variants in [12, 15]; see Sect. 7.

In the Algorithmic Pattern 4.2, Step 1 is given by (5.5) without stopping test. Once finiteness of the noise attenuation loop has been settled down by Corollaries 5.2 and 5.3, the \({K^\infty }\)-subsequences generated by the algorithm satisfy (5.1) and fit one of the following arguments:

  • Bundling: null steps issued from the same center suitably improve the model;

  • Descent: steps satisfying (3.7) will force the convergence property (4.9).

We start with several intermediate stages that are independent of the descent rule (3.7).

6.1 Common results for descent and null steps

The following technical Féjer-like identity is derived from writing (4.1) in the form \(|u_{k+1}-u|^2=|\hat{u}_k-t_k\hat{g}_k-u|^2\) and expanding the square:

$$\begin{aligned} |u_{k+1}-u|^2-|\hat{u}_k-u|^2= t_k^2|\hat{g}_k|^2+2t_k\left\langle \hat{g}_k ,u-\hat{u}_k\right\rangle \quad \text{ for } \text{ any }\, u\in \mathbb {R}^n. \end{aligned}$$
(6.1)

It is useful to collect descent iterates in the set

$$\begin{aligned} \hat{K} :=\{k\in \mathbb {N}: \text{ the } \text{ descent } \text{ test }\,(3.7)\, \text{ is } \text{ satisfied }\}. \end{aligned}$$

and consider in (4.9) the infinite sets

$$\begin{aligned} {K^\infty }:= \left\{ \begin{array}{l@{\quad }l} \{k\in \mathbb {N}:\, k>\hat{k}\} &{} \text{ if }\, \hat{K}\, \hbox {is finite, with last element}\, \hat{k},\\ {\hat{K}}&{} \text{ if }\, \hat{K}\, \hbox {has infinite cardinality.} \end{array} \right. \end{aligned}$$
(6.2)

Once again, recall that Corollaries 5.2 and 5.3 already ruled out an infinite loop in Step 1 from (5.5), so the infinite sets \({K^\infty }\) in (4.9) can only come from either a descent or a null step.

The general convergence property below does not require the model in P to satisfy any condition. However, we do assume the model satisfies (4.11), so that the corollaries from Sect. 5 apply and the algorithm \({K^\infty }\)-subsequences are well defined.

Proposition 6.1

Suppose that in the set P the Algorithmic Pattern 4.2 has a bounded model and stepsizes bounded away from zero, so that both (4.11) and

$$\begin{aligned} t_k\geqslant t_{\mathtt{low}}>0, \quad \text{ for } \text{ all }\, k \end{aligned}$$
(6.3)

hold. If for any of the two index sets \({K^\infty }\) from (6.2) the model decrease eventually vanishes:

$$\begin{aligned} \lim _{k\in {K^\infty }}\delta ^{\scriptscriptstyle M}_k=0, \end{aligned}$$

the convergence property (4.9) holds for such a set.

Proof

As Corollary 5.3 applies, (5.1) holds for \(k \in {K^\infty }\) and (5.2) yields that

$$\begin{aligned} \lim _{k\in {K^\infty }}t_k|\hat{g}_k|^2=0\quad \text{ and }\quad \left\{ \begin{array}{l@{\quad }l} \lim _{k\in {K^\infty }}\hat{g}_k=0 &{} \text{ by } \text{(6.3) } \\ \lim _{k\in {K^\infty }}\hat{e}_k=0 &{} \text{ by } \text{(4.5) }. \end{array}\right. \end{aligned}$$

Therefore, for (4.9) to hold, we only need to show that \(\phi =\lim _{K^\infty }\phi _k\leqslant 0\). When the index set is \({K^\infty }=\{k>\hat{k}\}\) this is direct from passing to the limit in the identity \(\phi _k=\hat{e}_k+\left\langle \hat{g}_k, \hat{u}\right\rangle \) from (4.8), because \(\hat{u}\) remains fixed to the last descent iterate. When the index set is \({K^\infty }={\hat{K}}\), take two consecutive indices \(k_1\) and \(k_2\) in \(\hat{K} \) and apply (6.1) written with \(k=k_2\) (so that \(\hat{u}_{k_2}=u_{k_1+1}\)) to obtain the identity

$$\begin{aligned} |u_{k_2+1}-u|^2-|u_{k_1+1}-u|^2 = t_{k_2}z_{k_2}(u), \text{ for } z_k :=t_k|\hat{g}_k|^2+2\left\langle \hat{g}_k ,u-\hat{u}_k\right\rangle . \end{aligned}$$

The summation over \(k\) gives that \(-\infty <-|u_0-u|^2\leqslant \sum _{k\in \hat{K} }t_kz_k.\) Existence of some \(\kappa >0\) such that \(z_k\leqslant -\kappa \) for all \(k\in \hat{K} \) would imply \(\sum _{k\in \hat{K} } t_k<+\infty \), which is impossible because of (6.3). Thus we have proved the relations

$$\begin{aligned} 0\leqslant \limsup _{k\in \hat{K} }\,z_k= \limsup _{k\in \hat{K} }{\bigl (t_k|\hat{g}_k|^2+2\left\langle \hat{g}_k,u\right\rangle -2\left\langle \hat{g}_k,\hat{u}_k\right\rangle \bigr )}= -2\liminf _{k\in \hat{K} }\left\langle \hat{g}_k,\hat{u}_k\right\rangle , \end{aligned}$$

because \(t_k|\hat{g}_k|^2\rightarrow 0\). Since \(\hat{e}_k\rightarrow 0\), passing to the limit in (4.8) gives \(\phi =\liminf _{\hat{K} }\phi _k\leqslant 0\).\(\square \)

The relation with previous results in the bundle literature can be seen as follows. By (4.6) and Lemma 4.1, \(f^{\scriptscriptstyle M}_k(u)\geqslant f^{\scriptscriptstyle L}_{-k}(\hat{u}_k)+\left\langle \hat{g}_k ,u-\hat{u}_k\right\rangle = \ell _k-\hat{e}_k+\left\langle \hat{g}_k ,u-\hat{u}_k\right\rangle \) for all \(u\in \mathbb {R}^n\). When the model satisfies (4.11), by the level definition (3.9) and (1.2), \(f(u)+\eta ^{\scriptscriptstyle M}\geqslant f(\hat{u}_k)-\eta _{\hat{u}_k}-\hat{e}_k+ \left\langle \hat{g}_k ,u-\hat{u}_k\right\rangle \) and, hence,

$$\begin{aligned} \text{(4.11) } \text{ implies } \text{ that } \hat{g}_k\in \partial _{\varepsilon _k}f(\hat{u}_k) \text{ for } \varepsilon _k :=\hat{e}_k+\eta _{\hat{u}_k}+\eta ^{\scriptscriptstyle M}\geqslant 0, \end{aligned}$$
(6.4)

noting that \(\varepsilon _k\geqslant 0\) by Lemma 5.1(iii).

For exact oracles, the arguments in Proposition 6.1 do not extend the standard proof of bundle methods. Such a proof is based on the property \(\hat{g}_k\in \partial _{\hat{e}_k}f(\hat{u}_k)\) (not valid for inexact oracles), which allows a refinement of (6.1). If the speed of convergence of \(\delta ^{\scriptscriptstyle M}_k\) to zero can be assessed, better results are obtained: a weaker assumption on the stepsizes is possible and the full sequence \(\{\hat{u}_k\}\) converges when \(f(\cdot )\) has a nonempty set of minimizers. Not unexpectedly, a variant of Proposition 6.1 recovers these two results if the oracle noise is suitably controlled, via the asymptotic error at descent steps \(\eta _{\infty }\) introduced in (5.6).

Theorem 6.2

(Link with the (partly asymptotically) exact case) Consider the Algorithmic Pattern 4.2 applied with an oracle of lower type that is asymptotically exact at descent steps, as in (5.6): \(\eta ^\infty = \liminf _{k\in \hat{K}} \eta _{\hat{u}_k}=0\). Suppose that in the set P the model is lower and the stepsizes series diverges: (3.5) is satisfied and

$$\begin{aligned} \sum _{k\in {K^\infty }} t_k=\infty . \end{aligned}$$

The following holds.

  1. (i)

    If  \(\lim _{k\in {K^\infty }}\delta ^{\scriptscriptstyle M}_k=0,\) then \(\liminf _{K^\infty }f(u_k)=\inf f(\cdot )\).

Suppose in addition that in P the stepsize sequence is bounded from above:

$$\begin{aligned} t_k\leqslant t^{\mathtt{up}}, \quad \text{ for } \text{ all }\, k. \end{aligned}$$
  1. (ii)

    In the null-step-tail case \(({K^\infty }=\{k>\hat{k}\})\) the last descent step \(\hat{u}_k\equiv u_{\hat{k}}=:\hat{u}\) satisfies

    $$\begin{aligned} \hat{u}\,\, minimizes\, f(\cdot ), \lim _{k\in {K^\infty }}u_k=\hat{u}\text{, } \text{ and } \lim _{k\in {K^\infty }}f^{\scriptscriptstyle M}_{k-1}(u_{k})=f(\hat{u}). \end{aligned}$$

Furthermore, suppose both the oracle error and the model decrease series are convergent:

$$\begin{aligned} \sum _{k\in {\hat{K}}}\eta _{\hat{u}_k} < +\infty \text{ in } \text{(5.6) }\quad \text{ and }\quad \displaystyle \sum _{k\in {\hat{K}}}\delta ^{\scriptscriptstyle M}_k < +\infty \,\,\text{ in }\,\, \mathbf{D} . \end{aligned}$$
  1. (iii)

    In the infinite-descent-step case (\({K^\infty }={\hat{K}}\)), for any limit point \(\hat{u}\) of the sequence \(\{\hat{u}_k\}_{k\in {\hat{K}}}\)

    $$\begin{aligned} \hat{u}\,\text{ minimizes }\, f(\cdot ),\, \hbox {and the whole sequence}\, \{\hat{u}_k\}_{k\in {\hat{K}}}\, \hbox {converges to}\, \hat{u}. \end{aligned}$$

Proof

With the oracle and model assumptions Corollary 5.2(ii) applies and the sets \({K^\infty }\) are well defined. To see (i), note that (5.2) yields that \(\delta ^{\scriptscriptstyle M}_k\geqslant (1-\beta _k)t_k| \hat{g}_k|^2 \geqslant (1-b)t_k|\hat{g}_k|^2\geqslant 0\). Since \(\delta ^{\scriptscriptstyle M}_k\rightarrow 0\) by assumption, \(t_k|\hat{g}_k|^2\rightarrow 0\) for the considered subsequence. In (4.11), \(\eta ^{\scriptscriptstyle M}=0\) by (3.5); together with (6.4) the inclusion \(\hat{g}_k\in \partial _{\varepsilon _k}f(\hat{u}_k)\) holds with \(\varepsilon _k :=\hat{e}_k+\eta _{\hat{u}_k}\). Adding \(\eta _{\hat{u}_k}\) to the left hand side inequality in (5.2) gives that \(\varepsilon _k=\hat{e}_k+\eta _{\hat{u}_k}\leqslant \delta ^{\scriptscriptstyle M}_k+\eta _{\hat{u}_k}\) and our assumption on \(\eta _{\hat{u}_k}\) implies that \(\varepsilon _k\rightarrow 0\). Then (i) is [5, Prop. 1.2], where \(\hat{g}\) is denoted \(\gamma \). To see (ii), first note that (4.1) implies that \(|u_{k+1}-\hat{u}|^2 = t_k^2|\hat{g}_k|^2\rightarrow 0\) because the stepsizes are bounded above by assumption. As the model decrease vanishes, (3.8) gives that \(\lim _k \ell _k=\lim _kf^{\scriptscriptstyle M}_k(u_{k+1})\). Together with the level definition (3.9) and the oracle assumption (5.6), which forces \(\eta _{\hat{u}}=0\), we see that \(f(\hat{u})=f_{\hat{u}}\leqslant \lim _k f^{\scriptscriptstyle M}_k(u_{k+1}) \leqslant f(\hat{u}). \) By lower semicontinuity, \(f(\hat{u}) \leqslant \liminf _k f(u_{k+1})\) and (ii) follows. To prove (iii), observe first that \(\hat{u}\) minimizes \(f(\cdot )\) by (i). Then use that \(\hat{g}_k\in \partial _{\varepsilon _k}f(\hat{u}_k)\) and write from (6.1)

$$\begin{aligned} \begin{array}{rcl} |u_{k+1}-\hat{u}|^2-|\hat{u}_{k}-\hat{u}|^2 &{}= &{} t_k^2|\hat{g}_k|^2+ 2t_k\left\langle \hat{g}_k,\hat{u}-\hat{u}_k\right\rangle \\ &{} \leqslant &{} t_k^2|\hat{g}_k|^2+ 2t_k[f(\hat{u})-f(\hat{u}_k)+ \varepsilon _k] \leqslant t_k^2|\hat{g}_k|^2+2t_k\varepsilon _k. \end{array} \end{aligned}$$

The definition of \(\varepsilon _k\) and (4.5) yield that \(t_k^2|\hat{g}_k|^2+2t_k\varepsilon _k= t_k(\delta ^{\scriptscriptstyle M}_k-\hat{e}_k+2\varepsilon _k) \leqslant 2t_k(\delta ^{\scriptscriptstyle M}_k+\eta _{\hat{u}_k}) \). For successive indices \(k_1\) and \(k_2\) in \({\hat{K}}\) summing the inequalities

$$\begin{aligned} |u_{k_2+1}-\hat{u}|^2-|u_{k_1+1}-\hat{u}|^2 \leqslant 2t_k(\delta ^{\scriptscriptstyle M}_k+\eta _{\hat{u}_k}), \end{aligned}$$

together with the assumptions on \(\{\eta _{\hat{u}_k}\}\) and \(\{\delta ^{\scriptscriptstyle M}_k\}\), implies that the rightmost side term forms a convergent series, so (iii) follows from [5, Prop. 1.3]. \(\square \)

In view of Proposition 6.1, obtaining small \(\delta ^{\scriptscriptstyle M}_k\) will be our main concern in Sect. 6.3. We first state conditions ensuring this property when \(\hat{K}\) in (6.2) is finite.

6.2 Null-step tail

As the stability center remains fixed throughout the present subsection, we use the notation \(\hat{u} :=\hat{u}_k\) and assume a weakened form of (4.11), holding only at \(\hat{u}\):

$$\begin{aligned} \text{ for } \text{ some }\, \hat{\eta }^{\scriptscriptstyle M}\geqslant 0\,\text{ the } \text{ inequality }\, f^{\scriptscriptstyle M}_k(\hat{u})\leqslant f(\hat{u})+ \hat{\eta }^{\scriptscriptstyle M}\,\, \text{ holds } \text{ for } \text{ all }\,k. \end{aligned}$$
(6.5)

For the concrete Algorithm 5.4, the model is of lower type and the stronger condition (4.11) always holds with \((\hat{\eta }^{\scriptscriptstyle M}=)\eta ^{\scriptscriptstyle M}=0\).

The null-step situation, in (6.2) \({K^\infty }=\{k>\hat{k}\}\), just relies upon the memory effect implied by \(f^{\scriptscriptstyle M}_{k+1}(\cdot )\geqslant f^{\scriptscriptstyle L}_{-k}(\cdot )\), triggered by the right branch in Step 3 of the Algorithmic Pattern 4.2. We claim that when Step 3 systematically makes a null step, regardless of any descent test,

$$\begin{aligned} \limsup _{k>\hat{k} }{\bigl [f_{u_{k}}-f^{\scriptscriptstyle M}_{k-1}(u_{k})\bigr ]}\leqslant 0. \end{aligned}$$
(6.6)

Two sources of errors make (1.1) difficult: one coming from the oracle and another from the model. Property (6.6) states that for null steps the model inexactness eventually vanishes.

Another important observation on the role of (6.6) refers to the nominal and effective decreases in (3.7). For simplicity, take \(\alpha _k=0\) in (4.3), so that \(\delta ^{\scriptscriptstyle N}_k=\delta ^{\scriptscriptstyle M}_k\) from (3.8). Then

$$\begin{aligned} \delta ^{\scriptscriptstyle E}_k=\ell _k-f_{u_{k+1}}= \delta ^{\scriptscriptstyle M}_k+f^{\scriptscriptstyle M}_k(u_{k+1})-f_{u_{k+1}}= \delta ^{\scriptscriptstyle N}_k-\bigl [f_{u_{k+1}}-f^{\scriptscriptstyle M}_k(u_{k+1})\bigr ]. \end{aligned}$$

When the bracket becomes small, the effective and nominal decreases get close together. This is a little known point in bundle methods: a good (effective) decrease \(\ell _k-f_{u_{k+1}}\) entails a more accurate model approximation \(f_{u_{k+1}}-f^{\scriptscriptstyle M}_k(u_{k+1})\).

To establish (6.6) we state first a technical result linking successive optimal values of the master-program (3.3), based on arguments similar to the exact oracle case.

Lemma 6.3

Consider the Algorithmic Pattern 4.2 and suppose that in the set P the model satisfies (6.5) and the stepsize is not increased at null steps:

$$\begin{aligned} t_{k}\leqslant t_{k-1}\quad \text{ if } \text{ at } \text{ iteration } k-1 \text{ the } \text{ descent } \text{ test } \text{(3.7) } \text{ is } \text{ not } \text{ satisfied. } \end{aligned}$$

For \(u_k\) and \(u_{k+1}\) obtained by a null step issued from the center \(\hat{u}\) the following holds.

  1. (i)

    \(f^{\scriptscriptstyle S}_{k-1}(u_{k})+\frac{1}{2t_{k-1}}|u_{k+1}-u_{k}|^2 \leqslant f^{\scriptscriptstyle S}_{{k}}(u_{k+1})\),

  2. (ii)

    \(f^{\scriptscriptstyle S}_{k-1}(u_{k})+\frac{1}{2t_{k-1}}|\hat{u}-u_{k}|^2 \leqslant f(\hat{u})+\hat{\eta }^{\scriptscriptstyle M}\),

  3. (iii)

    \(f^{\scriptscriptstyle M}_k(u_{k+1})-f^{\scriptscriptstyle M}_{k-1}(u_k)\leqslant f^{\scriptscriptstyle S}_k(u_{k+1})-f^{\scriptscriptstyle S}_{k-1}(u_k)+o_k, \) where we have set

    $$\begin{aligned} o_k :=\frac{\left\langle u_{k+1}-u_{k} ,\hat{u}-u_k\right\rangle }{t_k}= \frac{t_{k-1}}{t_k}\,\left\langle \hat{g}_{k-1} ,u_{k+1}-u_k\right\rangle . \end{aligned}$$
    (6.7)

Proof

For (i) and (ii) we refer to [15, Lemma 3.3]. To see (iii), by the definition in (3.3),

$$\begin{aligned} f^{\scriptscriptstyle S}_k(u_{k+1})-f^{\scriptscriptstyle M}_k(u_{k+1})= \frac{1}{2t_k}|u_{k+1}-u_k+u_k-\hat{u}|^2. \end{aligned}$$

Develop the square and use \(t_k\leqslant t_{k-1}\) in the right hand side to see that

$$\begin{aligned} \frac{1}{2t_k}|u_{k+1}-u_k|^2-o_k+\frac{1}{2t_k}|u_k-\hat{u}|^2 \geqslant -o_k+\frac{1}{2t_{k-1}}|u_{k}-\hat{u}|^2. \end{aligned}$$

The result follows, because \(f^{\scriptscriptstyle S}_k(u_{k+1})-f^{\scriptscriptstyle M}_k(u_{k+1})\geqslant -o_k+f^{\scriptscriptstyle S}_{k-1}(u_k)-f^{\scriptscriptstyle M}_{k-1}(u_k)\).

\(\square \)

With an exact oracle, (6.6) becomes \(f(u_k)-f^{\scriptscriptstyle M}_{k-1}(u_k)\rightarrow 0\), a known result, see for instance [5, Prop. 4.3]. Typically, \(f^{\scriptscriptstyle M}_{k-1}(u_{k-1})=f(u_{k-1})\); so we can write this as

$$\begin{aligned}{}[f(u_k)-f(u_{k-1})]+[f^{\scriptscriptstyle M}_{k-1}(u_{k-1})-f^{\scriptscriptstyle M}_{k-1}(u_k)]\, \rightarrow 0, \end{aligned}$$

easily proved with the Lipschitz property of \(f(\cdot )\) and \(f^{\scriptscriptstyle M}(\cdot )\) (Lemma 6.3 turns out to imply \(u_k-u_{k-1}\rightarrow 0\), see (6.10) below). In the inexact case, the oracle \(f\)-values may behave erratically, as well as the successive models \(f^{\scriptscriptstyle M}(\cdot )\). Since Lemma 6.3(i) implies a better behavior of the stabilized function \(f^{\scriptscriptstyle S}(\cdot )\), item (iii) relates the model to the stabilized model.

Theorem 6.4

(Null steps) Consider the Algorithmic Pattern 4.2 applied with an oracle having locally bounded inaccuracy:

$$\begin{aligned} \forall \,R\geqslant 0,\,\exists \,\eta (R)\geqslant 0\; \text{ such } \text{ that } \; |u|\leqslant R\,\Longrightarrow \,\eta _{u}+\eta ^g_{u}\leqslant \eta (R). \end{aligned}$$
(6.8)

Suppose that in the set P the model satisfies (6.5) and the stepsize is updated so that, whenever at iteration \(k-1\) the descent test (3.7) is not satisfied,

$$\begin{aligned} \text{ there } \text{ exist } \text{ positive }\,\,t_{\mathtt{low}}\,\text{ and }\,\sigma \in (0,1]\,\text{ such } \text{ that }\, t_k\in \left[ \max \Bigl (t_{\mathtt{low}},\sigma t_{k-1}\Bigr ),\,\, t_{k-1}\right] . \end{aligned}$$

Then the asymptotic property (6.6) holds.

Proof

We first establish the preliminary results (6.9) and (6.10) below. The stepsize update satisfies the condition in Lemma 6.3. By item (i) therein, the sequence \(\{f^{\scriptscriptstyle S}_{k-1}(u_{k})\}\) is nondecreasing, hence bounded from below; say \(f^{\scriptscriptstyle S}_{k-1}(u_k)\geqslant -M\) for all \(k\). By Lemma 6.3(ii),

$$\begin{aligned} \frac{1}{2t_{k-1}}|\hat{u}-u_{k}|^2\leqslant f(\hat{u})+\hat{\eta }^{\scriptscriptstyle M}-f^{\scriptscriptstyle S}_{k-1}(u_{k})\leqslant f(\hat{u})+\hat{\eta }^{\scriptscriptstyle M}+M. \end{aligned}$$

Using once more that stepsizes do not increase at null steps, we obtain that the sequence \(\{u_{k}\}\) is bounded. By (2.1) and (2.2), \(g_{u_k}\in \partial _{\eta _{u_k}+\eta ^g_{u_k}} f(u_k)\), and the oracle assumption (6.8) implies that \(\{g_{u_k}\}\) is bounded ([14, Prop. XI.4.1.2]). Our assumption on the stepsize implies in particular that (6.3) holds and, hence,

$$\begin{aligned} \text{ the } \text{ sequences }\, \{u_k\},\, \{\hat{g}_k =(\hat{u}-u_{k+1})/t_k\}\, \hbox {and}\, \{g_{u_k}\}\, \hbox {are bounded}. \end{aligned}$$
(6.9)

By Lemma 6.3(ii), the monotone sequence \(\{f^{\scriptscriptstyle S}_{k-1}(u_{k})\}\) is bounded from above and has a limit. Together with Lemma 6.3(i) and using once again the monotonicity of stepsizes,

$$\begin{aligned} f^{\scriptscriptstyle S}_{k}(u_{k+1})-f^{\scriptscriptstyle S}_{k-1}(u_{k})\rightarrow 0 \quad \text{ and }\quad u_{k+1}-u_{k}\rightarrow 0. \end{aligned}$$
(6.10)

We now use these preliminary results to show (6.6). The right branch in Step 3 of the algorithmic pattern forces \(f^{\scriptscriptstyle M}_k(\cdot )\geqslant f^{\scriptscriptstyle L}_k(\cdot )\) so, by (3.1), \(f_{u_{k}}+\left\langle g_{u_k} ,u-u_{k}\right\rangle =f^{\scriptscriptstyle L}_{k}(u) \leqslant f^{\scriptscriptstyle M}_{{k}}(u).\) In particular, when \(u=u_{k+1}\)

$$\begin{aligned} f_{u_{k}}-f^{\scriptscriptstyle M}_{k-1}(u_{k})&= f^{\scriptscriptstyle L}_{k}(u_{k+1})+\left\langle g_{u_k} ,u_{k}-u_{k+1}\right\rangle -f^{\scriptscriptstyle M}_{k-1}(u_{k}) \\&\leqslant f^{\scriptscriptstyle M}_{k}(u_{k+1})+\left\langle g_{u_k},u_{k}-u_{k+1}\right\rangle -f^{\scriptscriptstyle M}_{k-1}(u_{k}) \\&\leqslant [f^{\scriptscriptstyle S}_{k}(u_{k+1})-f^{\scriptscriptstyle S}_{k-1}(u_{k})]+ [\left\langle g_{u_k},u_{k}-u_{k+1}\right\rangle ]+o_k, \end{aligned}$$

by Lemma 6.3(iii). The results follows: by (6.9) and (6.10), the first two brackets tend to zero; and similarly for the last term, recalling our assumptions for the stepsize and (6.7). \(\square \)

Remark 6.5

(On boundedness) Assumption (6.8) could be refined as follows: the oracle is bounded for any infinite sequence of null steps. We shall make use of this refinement for some concrete instances in Sect. 7 related to Examples 2.3 and 2.4.\(\square \)

Remark 6.6

(On the role of the lower bound \(t_{\mathtt{low}}\)) Assumption (6.3) was only used to establish boundedness of the sequence \(\{\hat{g}_k\}\). This assumption can be dropped if the model is bounded everywhere, i.e., if (4.11) holds instead of (6.5). To see this, notice that in this case (6.4) gives that \(\hat{g}_k\in \partial _{\varepsilon _k}f(\hat{u})\) with \(\varepsilon _k=\hat{e}_k+\eta _{\hat{u}}+\eta ^{\scriptscriptstyle M}\). By local boundedness of the \(\varepsilon _k\)-subdifferential and by (6.8), we only need to show that \(\{\hat{e}_k\}\) is bounded. The latter results from boundedness of \(\{f^{\scriptscriptstyle S}_k(u_{k+1})\}\): plug (4.1) into the expression (4.7) of \(\hat{e}_k\) to obtain

$$\begin{aligned} \hat{e}_k= \ell _k-f^{\scriptscriptstyle M}_k(u_{k+1})-t_k|\hat{g}_k|^2\leqslant \ell _k-f^{\scriptscriptstyle S}_k(u_{k+1})\leqslant \ell _k+M \leqslant f(\hat{u})+M, \end{aligned}$$

where the last inequality follows from (3.9) recalling that \(\hat{u}_k=\hat{u}\).Finally, the assumption (6.3) can also be dropped for oracles of lower type that are partly asymptotically exact, as in Theorem 6.2.\(\square \)

6.3 The role of the descent test and general convergence result

For the bundling argument, property (6.6) allows us to analyze when the model decrease eventually vanishes. Recall that this argument enters the game when in (6.2) we have \({K^\infty }=\{k>\hat{k}\}\)—the descent test (3.7) does not hold. Since such a test depends on the effective decrease, below we give a sufficient condition for \(\delta ^{\scriptscriptstyle M}_k\rightarrow 0\) involving this decrease.

Proposition 6.7

(Effective decrease and bundling mechanism) In the setting of Theorem 6.4, suppose that for the level in the set P and the effective decrease in the set D

$$\begin{aligned} \limsup _{k>\hat{k}}{\bigl [\ell _k-f_{u_{k+1}} - \delta ^{\scriptscriptstyle E}_k\bigr ]}\leqslant 0\,; \end{aligned}$$
(6.11)

then \(\lim _{\hat{k}<k\rightarrow \infty } \delta ^{\scriptscriptstyle M}_k= 0\).

Proof

By Corollary 5.3 and Lemma 5.1 for the null step tail (5.1) holds and the model and nominal decreases satisfy the inequalities in (5.2) for all \(k>\hat{k}\). Subtract the identity \(f^{\scriptscriptstyle M}_k(u_{k+1})=\ell _k-\delta ^{\scriptscriptstyle M}_k\) from both sides of the negation of (3.7) and use (4.3) to obtain

$$\begin{aligned} -\delta ^{\scriptscriptstyle E}_{k}-f_{u_{k+1}}+f_{u_{k+1}}-f^{\scriptscriptstyle M}_k(u_{k+1})> -\ell _k+\delta ^{\scriptscriptstyle M}_k-m\,\delta ^{\scriptscriptstyle N}_k . \end{aligned}$$

Since in (4.3) the parameter \(\alpha _k\geqslant 0\), \(\delta ^{\scriptscriptstyle N}_k\leqslant \delta ^{\scriptscriptstyle M}_k\), and reordering terms we obtain that

$$\begin{aligned} (1-m\,)\delta ^{\scriptscriptstyle M}_k < z_k+f_{u_{k+1}}-f^{\scriptscriptstyle M}_k(u_{k+1}) \text{ for } z_k :=(\ell _k- f_{u_{k+1}}) - \delta ^{\scriptscriptstyle E}_k. \end{aligned}$$

By Theorem 6.4, the property (6.6) holds, together with (6.11) we obtain in the limit that

$$\begin{aligned} (1-m\,)\limsup \delta ^{\scriptscriptstyle M}_k\leqslant \limsup \bigl [z_k+f_{u_{k+1}}-f^{\scriptscriptstyle M}_k(u_{k+1})\bigr ]\leqslant \limsup z_k\leqslant 0. \end{aligned}$$

The result follows, recalling that \(m\,\in \; (0,1)\) and, by (5.2), \(\delta ^{\scriptscriptstyle M}_k\geqslant 0\). \(\square \)

Remark 6.8

(Interpretation for Algorithm 5.4) Condition (6.11) helps analyzing the impact of different definitions for the effective decrease. For the conservative decrease given in Algorithm 5.4, (6.11) is satisfied because \(\ell _k=\hat{f}_k\) and

$$\begin{aligned} \ell _k-f_{u_{k+1}} - \delta ^{\scriptscriptstyle E}_k&= \ell _k - f_{u_{k+1}} -(f_{\hat{u}_k} + \bar{\eta }_{\hat{u}_k} - f_{u_{k+1}} - \bar{\eta }_{u_{k+1}})\\&= \ell _k -f_{\hat{u}_k} - \bar{\eta }_{\hat{u}_k} + \bar{\eta }_{u_{k+1}} \\&= 0, \end{aligned}$$

where the last equality follows from the updating rule \(\bar{\eta }_{u_{k+1}} = \bar{\eta }_{\hat{u}_k} + f_{\hat{u}_k} -\ell _k\) for the error bounds. If we were to take the same level \(\ell _k=\hat{f}_k\), but use instead the realistic decrease \(\delta ^{\scriptscriptstyle E}_k= f_{\hat{u}_k} - f_{u_{k+1}}\), condition (6.11) would not hold. Indeed, reasoning like above,

$$\begin{aligned} \ell _k-f_{u_{k+1}} - \delta ^{\scriptscriptstyle E}_k&= \hat{f}_k - f_{u_{k+1}} -(f_{\hat{u}_k} - f_{u_{k+1}})\\&= \hat{f}_k -f_{\hat{u}_k} \\&\geqslant 0. \end{aligned}$$

The above inequality can be strict due to definition (3.10). In order to ensure (6.11) for this setting, the oracle should be asymptotically exact on descent steps, as in (5.6). Finally, with the observed decrease from [15], \(\ell _k=f_{\hat{u}_k}\) and \(\delta ^{\scriptscriptstyle E}_k= f_{\hat{u}_k} - f_{u_{k+1}}\), condition (6.11) is satisfied for all oracles, but not necessarily (5.1)—this is straightforward from Lemma 5.1 and (6.11). For (5.1) to hold in this setting, the controllable bundle algorithm would need to incorporate the noise attenuation loop, replacing Algorithm 5.4 Step 1 by the one in (5.5). \(\square \)

When the bundling argument applies, satisfaction of (6.11) is easy to accomplish (taking for example \(\delta ^{\scriptscriptstyle E}_k=\ell _k-f_{u_{k+1}}\)). By contrast, when the algorithm generates infinitely many descent iterates (\({K^\infty }=\hat{K}\) in (6.2)), the working horse is the descent test (3.7). For the model decrease to vanish, the effective decrease needs to vanish too; this is (6.12) below, a property that cannot be imposed a priori, but needs to be shown case by case, as in Sect. 7.

Proposition 6.9

(Effective decrease and descent mechanism) Consider the Algorithmic Pattern 4.2 and suppose that for the sets P and D the parameters \(\alpha _k,\beta _k\) satisfy (5.3). If for infinitely many iterations the descent test (3.7) is satisfied and

$$\begin{aligned} \lim _{k\in {\hat{K}}} \delta ^{\scriptscriptstyle E}_k=0 \quad \text{ for } \hat{K} \text{ from } \text{(6.2), } \end{aligned}$$
(6.12)

then \(\lim _{k\in {\hat{K}}}\delta ^{\scriptscriptstyle M}_k=0\).

Proof

By Corollary 5.3, condition (5.1) holds and by Lemma 5.1(ii) the model and nominal decreases satisfy the inequalities in (5.2) for all \(k\in {\hat{K}}\). In view of the assumption (5.3), both \(\delta ^{\scriptscriptstyle M}_k\) and \(\delta ^{\scriptscriptstyle N}_k\geqslant 0\) and satisfaction of (3.7) gives that

$$\begin{aligned} m\,(1-\alpha _k-\beta _k)t_k |\hat{g}_k|^2 \leqslant m\,\delta ^{\scriptscriptstyle N}_k \leqslant \delta ^{\scriptscriptstyle E}_k, \quad \text{ for }\, k\in {\hat{K}}. \end{aligned}$$

The result follows, because \(\alpha _k+\beta _k<1\) by (5.3) and \(\delta ^{\scriptscriptstyle M}_k = \delta ^{\scriptscriptstyle N}_k+\alpha _k t_k|\hat{g}_k|^2\) by (4.3).

\(\square \)

Since with an exact oracle taking \(\delta ^{\scriptscriptstyle E}_k=f(\hat{u}_k)-f(u_{k+1})\) is natural, the condition

$$\begin{aligned} \text{ in } \text{(1.1) } \quad \inf f(\cdot )>-\infty . \end{aligned}$$
(6.13)

implies satisfaction of (6.12). We now show the same holds for Algorithm 5.4.

Remark 6.10

(Interpretation for Algorithm 5.4 (cont.)) Corollary 5.2(ia) Since the oracle satisfies (1.2), the effective decrease is \(\delta ^{\scriptscriptstyle E}_k= f_{\hat{u}_k} + \bar{\eta }_{\hat{u}_k} - f_{u_{k+1}} - \bar{\eta }_{u_{k+1}}\) and, hence,

$$\begin{aligned} \sum _{k \in {\hat{K}}} \delta ^{\scriptscriptstyle E}_k&= f_{\hat{u}_1} + \bar{\eta }_{\hat{u}_1} - \lim _{k \in {\hat{K}}} (f_{\hat{u}_{k}} +\bar{\eta }_{\hat{u}_{k}}) \leqslant f_{\hat{u}_1} + \bar{\eta }_{\hat{u}_1} - \liminf _{k \in {\hat{K}}} (f(\hat{u}_{k})-\eta _{\hat{u}_{k}}+\bar{\eta }_{\hat{u}_{k}})\\&\leqslant f_{\hat{u}_1} + \bar{\eta }_{\hat{u}_1} -[ \liminf _{k \in {\hat{K}}} f(\hat{u}_{k})+ \liminf _{k \in {\hat{K}}}(-\eta _{\hat{u}_{k}}+\bar{\eta }_{\hat{u}_{k}})]. \end{aligned}$$

The sequence \(\{-\eta _{\hat{u}_{k}}+\bar{\eta }_{\hat{u}_{k}}\}\) is contained in \([0,\bar{\eta }_{\hat{u}_{1}}]\), so (6.12) follows from (6.13).

\(\square \)

To prove convergence of a generic proximal bundle method, oracle errors play no major role, but the stepsize needs to be updated in a manner that combines harmoniously all the requirements in the different statements. The following rule, depending on parameters \(t_{\mathtt{low}}>0\) and \(\sigma \in (0,1]\), addresses this issue. The update prevents decreasing the stepsize at null steps if noise was detected via a binary variable \(\mathtt {na}_k\) (equal to 1 if there was noise attenuation after generating the current center \(\hat{u}_k\), and zero otherwise):

$$\begin{aligned} \left\{ \begin{array}{lll} t_{k} &{}\geqslant t_{k-1}+t_{\mathtt{low}}&{}\text{ if }\, k-1\, \hbox {is a noisy step}\\ t_{k} &{}\geqslant t_{\mathtt{low}}&{}\text{ if }\, k-1\, \hbox {is a descent step}\\ t_{k} &{} \in \left[ \max \Bigl (t_{\mathtt{low}}, (1-\mathtt {na}_{k-1})\sigma t_{k-1}+ \mathtt {na}_{k-1}t_{k-1}\Bigr ), t_{k-1}\right] &{}\text{ if }\, k-1\, \hbox {is a null step}.\end{array}\right. \end{aligned}$$
(6.14)

Theorem 6.11

(Convergence) Consider the Algorithmic Pattern 4.2 with Step 1 from (5.5) applied with an oracle (1.2) with locally bounded inaccuracy, as in (6.8). If

$$\begin{aligned}&\text{ in } \text{ the } \text{ set }\, \mathbf{P} \, \left\{ \begin{array}{l} \text{ the } \text{ model } \text{ satisfies } \text{(4.11), }\\ \text{ the } \text{ level } \text{ is } \text{ given } \text{ as } \text{ in } \text{(3.9), } \text{ and }\\ \text{ the } \text{ stepsize } \text{ update } \text{ satisfies } \text{ the } \text{ rule } \text{(6.14) }, \end{array}\right. \end{aligned}$$
(6.15)
$$\begin{aligned}&\text{ in } \text{ the } \text{ set }\, \mathbf{D} \left\{ \begin{array}{l} \text{ the } \text{ effective } \text{ decrease } \text{ from } \text{(4.3) } \text{ satisfies } \text{(6.11) } \text{ and } \text{(6.12) } \text{ and }\\ \text{ the } \text{ parameters }\, \alpha _k\, \hbox {and}\, \beta _k\, \hbox {satisfy (5.3) if (5.1) is not automatic}\\ \text{ or }\, \alpha _k\in [0,1]\, \hbox {otherwise}, \end{array}\right. \end{aligned}$$
(6.16)

then the algorithm always generates some \({K^\infty }\)-subsequence that is optimal, in the sense that (4.9) is satisfied and Theorem 4.5 applies.

Proof

If Step 1 needs to test (5.1) and there is an infinite loop of noise attenuation, since the update in (6.14) drives \(t_k\) to infinity in this case, Corollary 5.3 gives the result. Otherwise, the loop in Step 1 is always finite and the algorithm generates either a last descent step at iteration \(\hat{k}\) followed by a null-step tail, or \(\hat{K}\) in (6.2) is infinite. In the first case (6.14) satisfies the stepsize conditions in Theorem 6.4. As (3.5) implies satisfaction of (6.5) with \(\hat{\eta }=\eta ^{\scriptscriptstyle M}\) the theorem applies. By Proposition 6.7, the model decrease vanishes and the assertion follows from Proposition 6.1, written with \({K^\infty }=\{k>\hat{k}\}\). Finally, if infinitely many descent steps are generated, the assumption that \(\lim _{k\in {\hat{K}}}\delta ^{\scriptscriptstyle M}_k=0\) follows from Proposition 6.9 and the result follows once again from Proposition 6.1, this time written with \({K^\infty }={\hat{K}}\). \(\square \)

Corollary 6.12

(Interpretation for Algorithm 5.4 (end)) If problem (1.1) satisfies (6.13), any limit point \(\hat{u}\) of the Controllable Bundle Algorithm 5.4 is \(\eta ^\infty \)-optimal with \(\eta ^\infty \leqslant \bar{\eta }_{\hat{u}_1}\).

Proof

The sets P and D are given in (5.7) and (5.8), respectively. The sequence of error bounds is nonincreasing and there is no need of noise attenuation, by Corollary 5.2(ia). The update (5.7) fits the rule (6.14) and the oracle inaccuracy is controllable with a nonincreasing sequence of error bounds, so (6.8) is satisfied with \(\eta (R) =\bar{\eta }_{\hat{u}_1}\). Finally, as explained in Remarks 6.8 and 6.10, both (6.11) and (6.12) are satisfied for the choices for \(\delta ^{\scriptscriptstyle N}_k\) and \(\delta ^{\scriptscriptstyle E}_k\) in (5.8). Theorem 6.11 applies with \(\eta ^{\scriptscriptstyle M}=0\) because both the oracle and the model are lower. \(\square \)

The final section reviews a number of methods fitting our convergence framework.

7 Instances

For each bundle variant in this section convergence follows from Theorem 6.11, by analyzing

  • if the oracle is bounded, in the sense of (6.8) or the Remark 6.5; and

  • if conditions (6.15) and (6.16) hold for the elements in the specific sets P and D .

Regarding the set P , in (6.15) the stepsize is given by (6.14), so we only need to check that the level and the model respectively satisfy (3.9) and (4.11).As for the set D and (6.16), we first determine if the considered variant needs to attenuate noise. If such is the case, parameters \(\alpha _k\) and \(\beta _k\) will be given by (5.3); otherwise \(\alpha _k\in [0,1]\). Here we only need to check that the effective decrease satisfies conditions (6.11) and (6.12). For this last property, we assume that problem (1.1) satisfies (6.13).

7.1 A collection of bundle methods for lower oracles and models

We review methods for lower oracles (2.3) and lower models, so that (3.5) holds and in (4.11) the error is null: \(\eta ^{\scriptscriptstyle M}=0\). To keep the master-program size controlled, the cutting-plane model can be endowed with bundle compression, like in Sect. 5.2. This mechanism replaces past linearizations by the aggregate one, \(f_{-k}^L(\cdot )\), so that (3.4) holds and, hence, in (3.9) taking \(\ell _k=\hat{f}_k\) is possible. For all these variants, the set P satisfies (6.15) and Step 1 never needs to attenuate noise because (5.1) always holds.

7.1.1 Exact oracles

Both the classical bundle methods and the spectral algorithms [11, 18] use an exact oracle and, hence, \(\ell _k=\hat{f}_k=f(\hat{u}_k)\). The effective decrease \(\delta ^{\scriptscriptstyle E}_k=f(\hat{u}_k)-f(u_{k+1})\) gives (6.11), because the left hand side therein is null, as for (6.12), it is direct from (6.13). The descent test (3.7)

$$\begin{aligned} f(u_{k+1})\leqslant f(\hat{u}_k)- m\delta ^{\scriptscriptstyle N}_k\quad \text{ for }\quad \delta ^{\scriptscriptstyle N}_k=f(\hat{u}_k)-f^{\scriptscriptstyle M}_k(u_{k+1})-\alpha _kt_k|\hat{g}_k|^2 \end{aligned}$$

usually takes \(\alpha _k\equiv 0\) in (4.3), but any value in \([0,1]\) could be used instead.

7.1.2 Partially inexact oracles

The partially inexact proximal bundle method was introduced in [9] and revisited in [16], for a level variant see [4]. To ensure that the function information is exact at descent steps the oracle should be a particular case of the partly asymptotically exact one given in Example 2.2, with \(\bar{\eta }_u=0\) whenever \(f_u\leqslant \gamma _u\). Then \(\ell _k=\hat{f}_k=f(\hat{u}_k)\) and the observed decrease \(\delta ^{\scriptscriptstyle E}_k :=f(\hat{u}_k)-f_{u_{k+1}}\) yields both a null left hand side in (6.11) and satisfaction of (6.12), by (6.13).

7.1.3 Incremental bundle method

The incremental bundle method [6] was developed for lower oracles with descent errors vanishing fast enough, as in (5.6). The realistic decrease \(\delta ^{\scriptscriptstyle E}_k= \hat{f}_k -f_{u_{k+1}}\) yields in (6.11) a null left hand side. Condition (6.12) is satisfied because \(\hat{f}_k \leqslant f_{\hat{u}_k}+ \eta _{\hat{u}_k}\) and, hence,

$$\begin{aligned} (0\leqslant )\;\sum _{k \in {\hat{K}}} \delta ^{\scriptscriptstyle E}_k \leqslant \sum _{k \in {\hat{K}}} (f_{\hat{u}_k}+ \eta _{\hat{u}_k} - f_{\hat{u}_{k+1}}) = \sum _{k \in {\hat{K}}} (f_{\hat{u}_k}- f_{\hat{u}_{k+1}}) + \sum _{k \in {\hat{K}}} \eta _{\hat{u}_k} <+ \infty . \end{aligned}$$

When \({\hat{K}}\) is finite, the last descent step \(\hat{u}\) is \(\eta ^\infty \)-solution with \(\eta ^\infty =\eta _{\hat{u}}\) not necessarily zero. When \({\hat{K}}\) is infinite, \(\eta ^\infty =0\) and the limit point solves problem (1.1) exactly. The criterion (4.8) makes superfluous the unboundedness detection loop in [6] (see the errata in [5]).

7.1.4 Asymptotically exact bundle method

This is the proximal version of the level bundle method for oracles with on-demand accuracy [4]. The variant is suitable for lower oracles that are eventually exact at descent iterates, like the partly asymptotically exact oracle in Example 2.2: \(\eta ^g_{u}\equiv 0\) and the known error bound \(\bar{\eta }_{\hat{u}_{k}}\) asymptotically vanishes. The conservative decrease \(\delta ^{\scriptscriptstyle E}_k= f_{\hat{u}_k}+\bar{\eta }_{\hat{u}_k}-f_{u_{k+1}}\) yields (6.11):

$$\begin{aligned} \ell _k-f_{u_{k+1}} - \delta ^{\scriptscriptstyle E}_k = \hat{f}_k - (f_{\hat{u}_k}+\bar{\eta }_{\hat{u}_k})\leqslant \hat{f}_k - f(\hat{u}_k) \end{aligned}$$

because the oracle is lower. As for (6.12), by (6.13) combined with (1.2)

$$\begin{aligned} 0\!\leqslant \!\sum _{k \in {\displaystyle {\hat{K}}}} (f_{\hat{u}_k}\!-f_{u_{k+1}})&= \sum _{k \in {\hat{K}}} (f_{\hat{u}_k}-f_{\hat{u}_{k+1}})\nonumber \\&= f_{\hat{u}_1}\!-\lim _{k\in {{\hat{K}}}} f_{\hat{u}_{k}} =f_{\hat{u}_1}-\lim _{k\in {{\hat{K}}}} (f(\hat{u}_{k})\!-\eta _{\hat{u}_k})\!< \!+\!\infty . \end{aligned}$$
(7.1)

So \((f_{\hat{u}_k}-f_{u_{k+1}})\rightarrow 0\) and (6.12) holds because \(\bar{\eta }_{\hat{u}_k}\rightarrow 0\) by the oracle assumption.The descent test (3.7) is \( f_{u_{k+1}}\leqslant f_{\hat{u}_k}+ \bar{\eta }_{\hat{u}_k}- m\delta ^{\scriptscriptstyle N}_k\) for \(\delta ^{\scriptscriptstyle N}_k=\hat{f}_k- f^{\scriptscriptstyle M}_k(u_{k+1})- \alpha _kt_k|\hat{g}_k|^2\).

In this method, even though the oracle never delivers exact evaluations, convergence is still exact, as in Sect. 7.1.2. To see this, it suffices to show that \(\eta _{\hat{u}}=0\) if \({K^\infty }=\{k>\hat{k}\}\) (when \({K^\infty }={\hat{K}},\, 0\leqslant \eta _{u_k}\leqslant \bar{\eta }_{u_k}\rightarrow 0\) by assumption). When there is a last descent step \(\hat{u}\), the definitions of \(\ell _k\) and \(\delta ^{\scriptscriptstyle E}_k\) imply that \(f_{u_{k+1}}> f_{\hat{u}_k}+ \bar{\eta }_{\hat{u}_k} - m\,\delta ^{\scriptscriptstyle N}_k\) if (3.7) does not hold. Then

$$\begin{aligned} \begin{array}{llll} f_{u_{k+1}} - f^{\scriptscriptstyle M}_k(u_{k+1}) &{}>&{} f_{\hat{u}}+ \bar{\eta }_{\hat{u}} - m\,\delta ^{\scriptscriptstyle N}_k - f^{\scriptscriptstyle M}_k(u_{k+1}) &{} [\hbox {by}\, ``\hbox {not}'' \,(3.7)] \\ &{}=&{} f_{\hat{u}} + \bar{\eta }_{\hat{u}} - m\,(\delta ^{\scriptscriptstyle M}_k-\alpha _kt_k|\hat{g}_k|^2)-f^{\scriptscriptstyle M}_k(u_{k+1})&{} [\hbox {by}\, (4.3)]\\ &{}=&{} f_{\hat{u}} + \bar{\eta }_{\hat{u}} - \ell _k - m\,(\delta ^{\scriptscriptstyle M}_k - \alpha _k t_k|\hat{g}_k|^2) +\delta ^{\scriptscriptstyle M}_k &{} [\hbox {by}\, (3.8)] \\ &{}=&{} \bigl ( f_{\hat{u}} + \bar{\eta }_{\hat{u}} - \hat{f}_k\bigr ) +(1-m\,)\delta ^{\scriptscriptstyle M}_k + m\,\alpha _k t_k|\hat{g}_k|^2. &{} [\hbox {because}\, \ell _k=\hat{f}_k] \end{array} \end{aligned}$$

In view of (6.6), \(\limsup _k \bigl (f_{\hat{u}}+\bar{\eta }_{\hat{u}}-\hat{f}_k\bigr ) +(1-m\,)\delta ^{\scriptscriptstyle M}_k+m\,\alpha _kt_k|\hat{g}_k|^2\leqslant 0.\) However, the second and third terms above are non-negative (\(\delta ^{\scriptscriptstyle M}_k \geqslant 0\) by Proposition 6.9). Similarly for the first term because, by the first line in (1.2) together with (3.10) and (3.4) evaluated at \(\hat{u}\),

$$\begin{aligned} f_{\hat{u}} + \bar{\eta }_{\hat{u}} \geqslant f(\hat{u}) \geqslant \hat{f}_k . \end{aligned}$$
(7.2)

Therefore \(f_{u_{k+1}}-f^{\scriptscriptstyle M}_k(u_{k+1})\rightarrow 0\), hence \(\hat{f}_k\rightarrow f_{\hat{u}}+\bar{\eta }_{\hat{u}}\). In the limit, the right hand side in (7.2) yields that \(f_{\hat{u}}+\bar{\eta }_{\hat{u}}=f(\hat{u})\) and, by the first line in (1.2), \( f(\hat{u})\leqslant f_{\hat{u}} + \eta _{\hat{u}} \leqslant f_{\hat{u}} + \bar{\eta }_{\hat{u}} = f(\hat{u}), \) so \(\eta _{\hat{u}} = \bar{\eta }_{\hat{u}}=0\), as stated.

Finally note that for oracles with errors vanishing sufficiently fast, as in (5.6), if the update (6.14) takes stepsizes \(t_k\leqslant t^{\mathtt{up}}\), the stronger convergence result in Theorem 6.2 holds.

7.2 Inexact bundle methods

The methods in [12, 15] also fit our convergence framework. They deal with uncontrollable oracles, bounded in the sense of (6.8) or Remark 6.5, and possibly of upper type: the level (3.9) is \(\ell _k=f_{\hat{u}_k}\). Then with any model of the form (3.4) condition (4.11) holds and the set P satisfies (6.15). The set D satisfies (6.16) with the observed decrease \(\delta ^{\scriptscriptstyle E}_k=f_{\hat{u}_k}-f_{u_{k+1}}\), because (6.11) holds and (6.12) follows directly from (7.1) (the oracle error is bounded).

Theorem 6.11 ensures convergence of the method subsequences satisfying (5.1) (an infinite null tail or infinitely many descent iterates). Corollary 5.3 gives the result for an infinite loop of noise attenuation (after a last descent step (5.1) never holds). In all cases the algorithm determines asymptotically an \((\eta _{\infty }+\eta ^{\scriptscriptstyle M})\)-solution to problem (1.1). The present work generalizes [15] from taking \(\alpha _k\equiv 0\) and \(\beta _k\equiv \frac{1}{2}\) to any pair \(\alpha _k+\beta _k\leqslant 1-b\), and \(\beta _k\in [b,1-b]\) where \(b\in (0,\frac{1}{2}]\).

7.3 Linearly constrained problems

All our results hold when constraining (1.1) to a nonempty polyhedron \(U\):

$$\begin{aligned} \min f(u)\;\text{ s.t. } \; u\in U. \end{aligned}$$
(7.3)

because this simple set can be introduced directly in the master-program (3.3). In Lemma 4.1, the update (4.1) adds a normal element \(\nu _k\in N_U(u_{k+1})\) to the aggregate subgradient, so the aggregate linearization (4.2) is an underestimate of the function over the set \(U\) only, [15, §  2].

The crucial boundedness property (6.9) in Theorem 6.4, which now needs to hold for the sequence \(\{\hat{g}_k+\nu _k=(\hat{u} - u_{k+1})/t_k\}\), still follows from the condition (6.3); we refer to [15, Lemma 3.3] for details.

7.4 Nonlinearly constrained problems

In the following generalization of Example 2.3, the convex problem

$$\begin{aligned} \min \;{h}(u) \; \text{ s.t. } \; u\in U\; \text{ and }\; {c}(u)\leqslant 0 , \end{aligned}$$
(7.4)

has an “easy” objective function \({h}(\cdot )\). a scalar constraint \({c}(\cdot )\) hard to evaluate and a simple polyhedral set \(U\). Supposing a Slater condition holds for this problem, we now consider two solution methods whose convergence can be derived from our theory.

7.4.1 Using improvement functions

Letting \(\bar{h}\) denote the optimal value, solving the (7.4) is equivalent to solve (7.3) with

$$\begin{aligned} f(u) :=\max \{{h}(u)-\bar{h},{c}(u)\} \quad \text{ over } \text{ the } \text{ set } \;U. \end{aligned}$$

When \(\bar{h}\) is unknown, a possible replacement is the objective value at the current center, penalizing infeasibility, to prevent zigzagging. For example, the approximation in [21]:

$$\begin{aligned} f_{u_{k+1}} :=\max \{{h}(u)-{h}(\hat{u}_k)-\rho _k\max ({c}_{\hat{u}_k}, 0),{ c}_u- \sigma _k\max ({c}_{\hat{u } _k } , 0)\}, \end{aligned}$$

for parameters \(\rho _k,\sigma _k\) given. The notation makes explicit that the \(h\)- and \(c\)-oracles are respectively exact and inexact: the subgradient is either the exact \(g^{{h}}(u_{k+1})\) or the inexact \( g^{c}_{u_{k+1}}\), depending on which term realizes the maximum. Since the \({c}\)-oracle is of upper type, so is the \(f\)-oracle and the level (3.9) is \(\ell _k=f_{\hat{u}_k}=(1-\sigma _k)\max ({c}_{\hat{u } _k } , 0)\). The method needs noise attenuation in Step 1, with parameters in (5.3) related to the counterparts in [21] as follows

$$\begin{aligned} 2\alpha _k=\alpha _k^{vAS} \quad \text{ and }\quad 2\beta _k=1-\beta _k^{vAS}\quad \text{ for }\, (\alpha _k^{vAS},\beta _k^{vAS})\, \hbox {from}\, [1]. \end{aligned}$$

The cutting-plane models for the objective and constraint functions give the model:

$$\begin{aligned} f^{\scriptscriptstyle M}_k(u)= \max \{\check{{h}}_k(u)-{h}(\hat{u}_k)-\rho _k\max ({c}_{\hat{u} _k},0), \check{{c}}_k(u)- \sigma _k\max ({c}_{\hat{u } _k } , 0)\}. \end{aligned}$$

Satisfaction of (4.11) follows the condition holding for \(\check{{c}}_k(\cdot )\) because \(U\) is compact, while satisfaction of (6.8) follows from Remark 6.5.

With the (observed) effective decrease \( \delta ^{\scriptscriptstyle E}_ k:=\ell _k-f_{u_{k+1}}\), (6.11) is automatic.

Finally, to show (6.12), we consider one particular case for the penalty (the setting in [21] is more general):

$$\begin{aligned} \rho _{k+1}=\rho _k+1\,\text{ at } \text{ each } \text{ iteration }\, k\, \hbox {satisfying}\, (3.7)\, \hbox {for which}\, {c}_{u_{k+1}}>0. \end{aligned}$$

Descent iterates are either always unfeasible or remain feasible after a first feasible center is found (\(f_{u_{k+1}}\) is a maximum). In the first case, \(c_{\hat{u}_k}>0\) for all \(k\in {\hat{K}}={K^\infty }\). By the penalty update, in the max-operation defining \(f_u\) the second term eventually prevails. Hence,

$$\begin{aligned} \delta ^{\scriptscriptstyle E}_k=\ell _k-f_{u_{k+1}}&= (1-\sigma _k){c}_{\hat{u } _k }- ({c}_{\hat{u } _{k+1} } - \sigma _k{c}_{\hat{u } _k })\\&= {c}_{\hat{u } _k }- {c}_{\hat{u } _{k+1} } \,\text{ for }\, k\in {\hat{K}}\, \hbox {sufficiently large}. \end{aligned}$$

Summing over \(k\in {\hat{K}}\) shows that the effective decrease series is convergent, so (6.12) holds (regardless of the value of \(\sigma _k\)). In the second case, descent steps are eventually feasible and

$$\begin{aligned} \delta ^{\scriptscriptstyle E}_k=\ell _k-f_{u_{k+1}} =0- \max \{ {h}(\hat{u}_{k+1})-{h}(\hat{u}_k),{c}_{\hat{u}_{k+1}} \}\leqslant -{h}(\hat{u}_{k+1})+{h}(\hat{u}_k) \end{aligned}$$

for large \(k\in {\hat{K}}\). Assuming once more (6.13) and taking the sum gives (6.12).

By Theorem 6.11, the algorithm limit points solve (7.3) within the accuracy \(\eta ^\infty +\eta ^{\scriptscriptstyle M}\), and with the Slater assumption, solving this problem is equivalent to solving (7.4), as desired.

7.4.2 Using exact penalties

The Slater condition ensures that in (7.4) the set of Lagrange multipliers is nonempty and bounded. Therefore, for any \(\rho \) greater than the largest Lagrange multiplier, solutions to (7.4) can be found by minimizing over the set \(U\) the exact penalty function

$$\begin{aligned} f(u) := h(u) + \rho \max \{c(u),0\}, \end{aligned}$$

i.e., the setting (7.3). As in Sect. 7.4.1, the transformation of the constrained problem into a linearly constrained one introduces an exogenous inaccuracy that can be easily handled together with the genuine errors in the \(h\)- and \(c\)-oracles.With a model of the form \( f^{\scriptscriptstyle M}_k(u) :=h^{\scriptscriptstyle M}_k(u)+\rho \max {\{c^{\scriptscriptstyle M}_k(u),0\}} \), (4.11) and (6.8) follow from the \(h\)—and \(c\)-models and oracles.The approach bottleneck is estimating the penalty: making successive approximations of this parameter amounts to having a more accurate \(f\)-oracle. A possible penalty update is \(\rho _k = \max \{\rho _{k-1},\lambda _k+1\}\) for a Lagrange multiplier \(\lambda _k \geqslant 0\) of

$$\begin{aligned} \min _{u \in U} \;h^{\scriptscriptstyle M}_k(u) +\frac{1}{2t_k} |u-\hat{u}_k|^2 \; \text{ s.t. } \; c^{\scriptscriptstyle M}_k(u)\leqslant 0, \end{aligned}$$
(7.5)

a Successive Quadratic Programming-like problem that gives the next iterate \(u_{k+1}\). This is a feasible problem (by the Slater assumption) whose solution also solves

$$\begin{aligned} \min _{u \in U} \; h^{\scriptscriptstyle M}_k(u)+\rho _k\max {\{c^{\scriptscriptstyle M}_k(u),0\}}+ \frac{1}{2t_k}|u-\hat{u}_k|^2. \end{aligned}$$

for sufficiently large \(\rho _k\). For the approach to behave as an exact penalty method, the penalty needs to stabilize eventually. This requires the Lagrange multiplier sequence to be bounded, a property that we prove under appropriate assumptions.

Lemma 7.1

Suppose \(U\) is a bounded set and there exists \(u^0 \in U\) such that \(c(u^0)<0\). If the stepsizes satisfy (6.3) and the \(c\)-model satisfies

$$\begin{aligned} c^{\scriptscriptstyle M}_k(\cdot )\leqslant \eta ^{\scriptscriptstyle M}\quad \text{ for } \text{ some } \text{ bound } 0\leqslant \eta ^{\scriptscriptstyle M} < -c(u^0), \end{aligned}$$

the sequence of Lagrange multipliers \(\{\lambda _k\}\) of (7.5) is bounded.

Proof

The optimality condition for (7.5) provides \(p^h_k \in \partial h^{\scriptscriptstyle M}_k(u_{k+1}),\, p^c_k \in \partial c^{\scriptscriptstyle M}_k(u_{k+1})\), \(p^u_k \in N_U(u_{k+1})\), and \(\lambda _k\geqslant 0\) satisfying \(p^h_k + \frac{u_{k+1}-\hat{u}}{t_k}+\lambda _k p^c_k + p^u_k=0\). Without loss of generality, suppose \(\lambda _k>0\), so that \(p_k=p^c_k + p^u_k/\lambda _k\) and \(\lambda _k p_k = -(p^h_k + \frac{u_{k+1}-\hat{u}}{t_k}) \) yield the identity \(\lambda _k |p_k|^2 = -\left\langle p_k,(p^h_k + \frac{u_{k+1}-\hat{u}}{t_k})\right\rangle \). Assume for the moment \(p_k\ne 0\); by Cauchy–Schwarz,

$$\begin{aligned} \lambda _k=-\frac{1}{|p_k|^2}\left( \left\langle p_k,p^h_k\right\rangle +\left\langle p_k,\frac{u_{k+1}-\hat{u}_k}{t_k}\right\rangle \right) \leqslant \frac{|p_k|}{|p_k|^2}\left( |p^h_k| +\frac{|u_{k+1}-\hat{u}_k|}{t_k}\right) . \end{aligned}$$

As \(U\) is bounded and \(h^{\scriptscriptstyle M}_k(\cdot )\) is convex, then \(p^h_k\in \partial h^{\scriptscriptstyle M}_k(u_{k+1})\) is bounded as well. Together with (6.3) we see that there exists a constant \(M>0\) such that \(\lambda _k\leqslant M/|p_k|\) for all \(k\).

It remains to show that the sequence \(\{|p_k|\}\) is bounded away from zero. The definitions of \(p^c_k\) and \(p^u_k\) imply that \(c^{\scriptscriptstyle M}_k(u_{k+1})+\left\langle p^c_k,u^0-u_{k+1}\right\rangle \leqslant c^{\scriptscriptstyle M}_k(u^0)\) and \(\frac{1}{\lambda _k}\left\langle p^u_k, u^0-u_{k+1}\right\rangle \leqslant 0\). By adding these two inequalities and remembering that \(c^{\scriptscriptstyle M}_k(u_{k+1})=0\) because \(\lambda _k>0\), we get \(\left\langle p_k,u^0-u_{k+1}\right\rangle \leqslant c^{\scriptscriptstyle M}_k(u^0)\). Therefore,

$$\begin{aligned} -|p_k||u^0-u_{k+1}|\leqslant \left\langle p_k,u^0-u_{k+1}\right\rangle \leqslant c^{\scriptscriptstyle M}_k(u^0)\leqslant c(u^0) +\eta ^{\scriptscriptstyle M}<0, \end{aligned}$$

where the last inequality follows from the assumption on \(u^0\) and \(\eta ^{\scriptscriptstyle M}\). Since \(U\) is a bounded set, we conclude that \(\liminf _k |p_k| >0\), and hence \(\{\lambda _k\}\) is a bounded sequence. \(\square \)

Once the penalty parameter eventually stabilizes at a value \(\rho \), Theorem 6.11 applies: the limit points of the sequence \(\{\hat{u}_k\}\) solve the constrained problem within an accuracy bound depending also on the value \(\limsup _{k\in {\hat{K}}} \{0,\rho -\rho _{\hat{k}} \}\), the asymptotic error made when estimating the penalty at descent steps.

7.5 Composite functions

The composite bundle method [19] uses the approximation \(F(\cdot ;{\hat{u}})\) in Example 2.4 (with \(\hat{u}=\hat{u}_k\)) as an economic intermediate model for the function \(f(\cdot )=(h\circ c)(\cdot )\). The reason is that evaluating the \(f\)-subgradients is expensive, because they need computing the \(c\)-Jacobian:

$$\begin{aligned} \partial f(u)=Dc(u)^{\scriptscriptstyle \top }\partial h(C)\quad \text{ for }\quad C=c(u). \end{aligned}$$

To interpret this method in our setting, consider as oracle output

$$\begin{aligned} f_{u_{k+1}}&:= F(u_{k+1};\hat{u}_k)= h(C_{k+1})\quad \text{ for }\quad C_{k+1} :=c(\hat{u}_k)+Dc(\hat{u}_k) (u_{k+1}-\hat{u}_k) \\ g_{u_{k+1}}&:= Dc(\hat{u}_k)^{\scriptscriptstyle \top }G_{k+1}\quad \text{ for }\quad G_{k+1}\in \partial h(C_{k+1}). \end{aligned}$$

The \(h\)-oracle is exact everywhere but the \(f\)-oracle is exact at each center \(\hat{u}_k\). By smoothness of the operator \(c\), this oracle satisfies (6.8) for each fixed \(\hat{u}_k\), as in Remark 6.5.

By convexity, \(\check{h}(\cdot )\leqslant h(\cdot )\) and by positive homogeneity, the model

$$\begin{aligned} f^{\scriptscriptstyle M}_k(\cdot ) := \check{h}( c(\hat{u}_k)+Dc(\hat{u}_k) (\cdot -\hat{u}_k)) \end{aligned}$$

stays below \(F(\cdot ;\hat{u}_k)\), but not necessarily below \(f(\cdot )\). An interesting feature of the approach is that, even though the model is not of lower type ((3.5) may not hold), the method does not need to attenuate noise, thanks to the special model structure. More precisely, first note that

$$\begin{aligned} f^{\scriptscriptstyle M}_k(\hat{u}_k)= (\check{h}\circ c)(\hat{u}_k))\leqslant ({h}\circ c)(\hat{u}_k))=f(\hat{u}_k). \end{aligned}$$

Then, because \(\hat{g}_k\in \partial f^{\scriptscriptstyle M}_k(u_{k+1})\) by Lemma 4.1, from (4.6) we see that

$$\begin{aligned} f(\hat{u}_k) = f^{\scriptscriptstyle M}_k(\hat{u}_k) \geqslant f^{\scriptscriptstyle M}_k(u_{k+1})+\left\langle \hat{g}_k,\hat{u}_k-u_{k+1}\right\rangle = f^{\scriptscriptstyle L}_{-k}(\hat{u}_k) . \end{aligned}$$

So taking as level \(\ell _k=f(\hat{u}_k)\) gives a null aggregate gap (4.4) and, by item (ii) in Lemma 5.1, there is no need of noise attenuation. In [19] a null step is declared whenever there is no descent for the approximating function, corresponding to the observed decrease

$$\begin{aligned} \delta ^{\scriptscriptstyle E}_k :=F(\hat{u}_k;\hat{u}_k) - F(u_{k+1};\hat{u}_k)=f(\hat{u}_k)- h(C_{k+1})=\ell _k-f^{\scriptscriptstyle M}_k(u_{k+1}), \end{aligned}$$

which trivially ensures (6.11). Regarding (6.12), the composite bundle method checks a backtracking test before declaring a descent step. More precisely, for \(k\in {\hat{K}}\)

$$\begin{aligned} f(\hat{u}_k)- f(\hat{u}_{k+1})&= [f(\hat{u}_{k})- h(C_{k+1})] + [ h(C_{k+1})- f(\hat{u}_{k+1})]\\&= \delta ^{\scriptscriptstyle E}_k+ [ h(C_{k+1})- f(\hat{u}_{k+1})]. \end{aligned}$$

The usual argument invoking (6.13) would give (6.12) if the second bracket eventually vanished. The backtracking test declares a descent step if, in addition to (3.7), the condition

$$\begin{aligned} \left\langle G_{k+1},C_{k+1}-c(u_{k+1})\right\rangle \geqslant -m_2\delta ^{\scriptscriptstyle N}_k \text{ for } G_{k+1}\in \partial h(c(u_{k+1})) \end{aligned}$$

holds. Otherwise the stepsize is decreased (“backtracking”), with the same model and center. As the number of backtracking steps is finite ([19]), eventually the algorithm generates sets \({K^\infty }\) as in (6.2). The backtracking condition is easy to test, because the additional oracle call does not involve the expensive Jacobian at \(u_{k+1}\). Such a computation is done only if the backtracking test is passed, to define the next cheap oracle, i.e. \(F(\cdot ;\hat{u}_{k+1})\).Since positively homogeneous functions are support functions,

$$\begin{aligned} h(c(u_{k+1}))=\left\langle G_{k+1},c(u_{k+1})\right\rangle \quad \text{ and }\quad h(C_{k+1})\geqslant \left\langle G_{k+1},C_{k+1}\right\rangle , \end{aligned}$$

so checking the need of backtracking is equivalent to testing if \([ h(C_{k+1})- f(\hat{u}_{k+1})] \geqslant -m_2\delta ^{\scriptscriptstyle N}_k\). Together with (3.7) this means that when \({K^\infty }={\hat{K}}\) we have that

$$\begin{aligned} f(\hat{u}_k)- f(\hat{u}_{k+1})= \delta ^{\scriptscriptstyle E}_k+ [ h(C_{k+1})- f(\hat{u}_{k+1})]\geqslant (m_1-m_2)\delta ^{\scriptscriptstyle N}_k. \end{aligned}$$

Since \(m_1-m_2>0\), taking the sum over \(k\in {K^\infty }\) implies that the series of nominal decreases converges, by (6.13). So the effective decrease series converges too and (6.12) follows.

By Theorem 6.11, the limit points of the sequence satisfy (4.12), and in view of the level definition, they solve (1.1) exactly.