1 Introduction

Bilinear programs, for the purpose of this paper, can be written as the following nonconvex nonlinear programming problem:

$$\begin{aligned}&\text{ Min}\,z=f_0 =\sum \limits _{(i,j)\in BL_0} a_{ij0} x_i x_j +h_0 (x) \end{aligned}$$

Subject to

$$\begin{aligned} f_q =\mathop \sum \limits _{(i,j)\in BL_q } a_{ijq} x_i x_j +h_q (x)\le 0\quad q\in Q\backslash \{0\} \end{aligned}$$
(P)
$$\begin{aligned} x\in S\mathop \cap \limits \Omega \subset \mathbb R ^{n} \end{aligned}$$

where \(h_{q}(x)\) is convex and twice differentiable, \(a_{ijq}\) is a scalar with \(i \in I, j \in J\), and \(q \in Q\) represents the set of all functions \(f_{q}\), including the objective function \(f_{0}\) and all constraints. \(BL_{q}\) is an (\(i, j\))-index set which defines the bilinear terms present in the problem. Although \(i \ne j\) for strictly bilinear problems, \(i=j\) can be allowed to accommodate quadratic problems. The set \(S \subset \mathbb R ^{n}\) is convex, and \(\Omega \subset \mathbb R ^{n}\) is an \(n\)-dimensional hyperrectangle defined in terms of the initial variable bounds \(x^{L}\) and \(x^{U}\):

$$\begin{aligned} \Omega =\left\{ {x\in \mathbb R ^{n}:0\le x^{L}\le x\le x^{U}} \right\} \end{aligned}$$

The global optimization of bilinear programs of the form of (P) is important in such areas as water networks and petroleum blending operations [14]. Nonconvex, bilinear constraints are required to model the mixing of various streams in these systems, and are in some cases the only nonlinearities in the models. The pooling problem, stemming from the original Haverly paper [5], contains these bilinear constraints and has received much attention in the literature [1, 2, 69]. Recently, Misener and Floudas have demonstrated a novel logarithmic relaxation for modeling bilinear terms with piecewise McCormick envelopes while addressing various classes of pooling problems [1].

Water network optimization problems containing bilinear terms have also received much attention in the literature [3, 4, 1012]. The same blending constraints present in the pooling problem are present in water network problems, and thus numerous advances in solving bilinear programs have been made addressing these problems.

The global optimization of general nonconvex bilinear programs has received significant attention in the literature [1327]. The convex McCormick envelopes [1] coupled with spatial branch and bound search frameworks have been the basis for many of these global optimization techniques, with piecewise McCormick envelopes being a more recent development. Variations of this approach have been suggested, generalizing the convex envelopes to piecewise over- and under-estimators [1, 17]. Novel ways of representing bilinear terms through reformulation have been another approach reported in the literature [16]. Misener et al. [1], building on the work of Vielma and Nemhauser [28, 29], have shown that a relaxation of the bilinear terms can be achieved with a logarithmic number of binary variables. Teles et al. [30] have introduced a technique to approximate polynomial constraints that exhibits a similar property, where the main element is the discretization of a subset of the variables.

In this paper, we show that the mixed-integer constraints of the multiparametric disaggregation technique (MDT) presented in [30], when applied to the bilinear terms in program (P), can be derived from disjunctive programming and convex hull reformulation. This approximation technique can be used under some conditions to obtain an upper bound. As the main novelty, we propose a rigorous lower bounding formulation derived from the upper bounding constraints. After proving that the solution from the upper and lower bounding formulations converge to that of the original nonlinear formulation in the limit of an infinite number of discretization intervals, two global optimization algorithms are proposed based on such bounds. Finally, we conclude with a comparison of this new relaxation approach and the one based on piecewise McCormick envelopes, and report computational results on small and large problems.

2 Discretization with multiparametric disaggregation

Given a nonconvex bilinear term \(w_{ij} = x_{i} \cdot {x}_{j}\), the MDT described by Teles et al. [30] can be used to obtain an upper bound on problem (P). The complete formulation can be derived from a generalized disjunctive programming (GDP) model [31] followed by a convex hull reformulation [33] and exact linearization [32]. For simplicity in the notation, we first rewrite the bilinear product \(w_{ij} = x_{i} \cdot x_{j}\) as a single bilinear term \(w = u \cdot v\). This product can be represented exactly with the following constraints and disjunction:

$$\begin{aligned}&w=u\cdot v\end{aligned}$$
(1)
$$\begin{aligned}&v=\mathop \sum \limits _{\ell \in \mathbb Z } \lambda _\ell \end{aligned}$$
(2)
$$\begin{aligned}&\mathop {\mathop \vee \limits _{k=0}}\limits ^9 \left[ {\lambda _\ell =10^{\ell }\cdot k} \right]\quad \forall \,\ell \in \mathbb Z \end{aligned}$$
(3)

where \(v\) is discretized through the disjunction in (3) that selects one digit \(k \in K=\{0,1,{\ldots },9\}\) for each power \(\ell \in \mathbb Z \). Here we assume a basis of 10, although other bases can be selected [11]. Note that since (3) is defined over the domain of all the integer numbers, this implies an infinite number of disjunctions. Furthermore, \(v\) can represent any positive real number.

First, we consider the convex hull reformulation [33] of the disjunction in (3) after which we introduce the disaggregated variables,

$$\begin{aligned}&\lambda _\ell =\mathop \sum \limits _{k=0}^9\hat{{\lambda }}_{k,\ell } \quad \forall \ell \in \mathbb Z \end{aligned}$$
(4)
$$\begin{aligned}&\hat{\lambda }_{k,\ell } =10^{\ell }\cdot k\cdot z_{k,\ell } \quad \forall \, \ell \in \mathbb Z ,k\in K\end{aligned}$$
(5)
$$\begin{aligned}&\mathop \sum \limits _{k=0}^9z_{k,\ell } =1\quad \forall \, \ell \in \mathbb Z \end{aligned}$$
(6)
$$\begin{aligned}&z_{k,\ell } \in \{0,1\}\quad \forall \,\ell \in \mathbb Z ,k\in K \end{aligned}$$
(7)

Substituting (5) into (4) and then into (2) leads to the fully discretized (but still exact representation) of \(v\):

$$\begin{aligned} v=\mathop \sum \limits _{\ell \in \mathbb Z } \mathop \sum \limits _{k=0}^9 10^{\ell }\cdot k\cdot z_{k,\ell } \end{aligned}$$
(8)

Considering the product \(w = u \cdot v\) by substituting (8) into (1) leads to (9), which involves nonlinear terms \(u \cdot z _{k,\ell }\).

$$\begin{aligned} w=u\cdot \left[ {\mathop \sum \limits _{\ell \in \mathbb Z }\mathop \sum \limits _{k=0}^910^{\ell }\cdot k\cdot z_{k,\ell } } \right] \end{aligned}$$
(9)

Performing an exact linearization [32], we introduce new continuous variables \(\hat{{u}}_{k,\ell } =u\cdot z_{k,\ell } \) so that:

$$\begin{aligned} w=\mathop \sum \limits _{\ell \in \mathbb{Z }} \mathop \sum \limits _{k=0}^9 10^{\ell }\cdot k\cdot \hat{{u}}_{k,\ell } \end{aligned}$$
(10)

Since \(u\cdot z_{k,\ell } =\left\{ {\begin{array}{ll} 0&\text{ if}\,z_{k,\ell } =0 \\ u&\text{ if}\,z_{k,\ell } =1 \\ \end{array}} \right.\) and \(\hat{{u}}_{k,\ell } \) is non-negative, we introduce the following lower and upper bounding constraints, where \(u^{U}\) and \(u^{L}\) are the non-negative upper and lower bounds on \(u\).

$$\begin{aligned} u^{L}\cdot z_{k,\ell } \le \hat{{u}}_{k,\ell } \le u^{U}\cdot z_{k,\ell } \quad \forall \ell \in \mathbb Z ,k\in K \end{aligned}$$
(11)

Finally, multiplying equation (6) by \(u\) and replacing the bilinear terms by the new continuous variables, results in (12). The full set of mixed integer linear constraints for the exact representation of bilinear product \(w = u \cdot v\) is thus given by Eqs. (68) and (1012).

$$\begin{aligned} u=\mathop \sum \limits _{k=0}^9\hat{{u}}_{k,\ell } \quad \forall \ell \in \mathbb Z \end{aligned}$$
(12)

We should note that the same model formulation can be obtained by the special structured RLT reformulation as described in Sherali et al. [48]. Constraints (68) are directly linked to the choice of using a base-10 representation for variable \(v\). Then, the bounds of variable \(u\) are known: \(u^{L}\le u\le u^{U}\). Multiplying the lower and upper bounding constraints \(u-u^{L}\ge 0\) and \(u^{U}-u\ge 0\) by \(z_{k,\ell } \ge 0\) and replacing the bilinear term \(u\cdot z_{k,\ell } \) by \(\hat{{u}}_{k,\ell } \), leads to (11). Performing the same variable replacement in the constraint defining variable \(w,\) results in (10) while (12) can be obtained as previously described.

2.1 Upper bounding formulation

Since in practice it is infeasible to compute the infinite sums over all integers, we represent \(v\) to a finite level of precision, \(v^{\prime }\), leading to a continuous but approximate representation of the bilinear term, \(w^{\prime }\). The constraints in (8) and (10) are modified in (1314) to allow for a maximum power of 10 (\(P)\) and a minimum power of 10 (\(p)\). For the remaining constraints, (67) and (1112), it suffices to replace \(\ell \in \mathbb Z \) with \(\ell \in L=\{p,p+1,\ldots ,P\}\). These sets of constraints correspond to the equations proposed by Teles et. al. [30]:

$$\begin{aligned} w^{\prime }&= \mathop \sum \limits _{\ell =p}^P \mathop \sum \limits _{k=0}^910^{\ell }\cdot k\cdot \hat{{u}}_{k,\ell }\end{aligned}$$
(13)
$$\begin{aligned} v^{\prime }&= \mathop \sum \limits _{\ell =p}^P\mathop \sum \limits _{k=0}^910^{\ell }\cdot k\cdot z_{k,\ell } \end{aligned}$$
(14)

When we incorporate them into problem (P) by redefining \(w_{ij} = x_{i} \cdot x_{j}\), and selecting \(x_{j}\) as the variable on which discretization is performed, the resulting problem (\((P^{\prime})\)) will represent a mixed-integer approximation to the original problem:

$$\begin{aligned}&\text{ Min}\,z^{\prime }=f_0 =\mathop \sum \limits _{(i,j)\in BL_0} a_{ij0} w_{ij} +h_0 (x)\\&\mathrm{Subject\,to} \end{aligned}$$
$$\begin{aligned} f_q (x)=\mathop \sum \limits _{(i,j)\in BL_q } a_{ijq} w_{ij} +h_q (x)\le 0 \quad q\in Q\backslash \{0\} \end{aligned}$$
(P′)
$$\begin{aligned}&w_{ij}=\mathop \sum \limits _{\ell =p}^P\mathop \sum \limits _{k=0}^910^{\ell }\cdot k\cdot \hat{{x}}_{ijk\ell } \quad \forall (i,j)\in BL_q ,q\in Q\\&x_j=\mathop \sum \limits _{\ell =p}^P\mathop \sum \limits _{k=0}^910^{\ell }\cdot k\cdot z_{jk\ell } \quad \forall j\in \{j|(i,j)\in BL_q ,q\in Q\}\\&x_i=\mathop \sum \limits _{k=0}^9\hat{{x}}_{ijk\ell }\quad \forall (i,j)\in BL_q ,q\in Q,\ell \in L\\&x_i^L \cdot z_{jk\ell } \le \hat{{x}}_{ijk\ell } \le x_i^U \cdot z_{jk\ell } \quad \forall (i,j)\in BL_q ,q\in Q,k\in K,\ell \in L\\&\mathop \sum \limits _{k=0}^9z_{jk\ell } =1\quad \forall j\in \{j|(i,j)\in BL_q ,q\in Q\},\ell \in L\\&z_{jk\ell } \in \{0,1\}\quad \forall j\in \{j|(i,j)\in BL_q ,q\in Q\},k\in K,\ell \in L\\&x\in S\mathop \cap \limits \Omega \subset \mathbb R ^{n} \end{aligned}$$

where \(x_{j}\) and \(w_{ij}\) represent discrete and continuous approximations to the variables, respectively. Note that if the convex functions \(h_{q}(x)\) are linear, problem (\(\mathbf{P}^{\prime }\)) represents a mixed integer linear program (MILP). Otherwise it corresponds to an MINLP.

Further, note that problem (\(\mathbf{P}^{\prime }\)) is a restricted version of problem (P), or equivalently problem (P) is a relaxation of problem (\(\mathbf{P}^{\prime }\)). It then follows that the solution of problem (\(\mathbf{P}^{\prime }\)) either yields an upper bound on problem (P) such that \(z^{\prime } =z^{U} \ge z\), or else problem (\(\mathbf{P}^{\prime }\)) is infeasible. This restricted feasible region can be seen in Figs. 1 and 2.

Fig. 1
figure 1

Feasible region for a bilinear curve \(x_{i}\cdot x_{j} = 0.1, p = -2, P = -1\). The solid curve represents the exact bilinear curve, while the dots represent the approximated (and incomplete) feasible region resulting from the upper bounding formulation

Fig. 2
figure 2

Feasible region for a bilinear surface. The surface is the exact bilinear curve, while the solid lines represent the reduced feasible region resulting from the upper bounding formulation for \(p=P = -1\)

2.2 Infeasibilities in the discretized problem

The mixed-integer approximation problem (\(\mathbf{P}^{\prime }\)) can be infeasible even if the original problem (P) is feasible. For example, if bounds such as \(10^{p-1} \le x_{j} \le 2\times 10^{p-1}\) are enforced, (\(\mathbf{P}^{\prime }\)) will be infeasible, while such a constraint is completely valid and will not necessarily result in an infeasible problem (P). Thus, the parameters \(p\) and \(P\) must be chosen appropriately to avoid such infeasibilities.

Some general guidelines for ensuring precision-based feasibility can be established. For example, the largest power of 10 (\(P\)) must be large enough such that 10\(^{P}\) is of the same order of magnitude as the upper bound on \(x_{j}\). More precisely: \(P= \lfloor \log _{10} x_j^U \rfloor \).

Additionally, \(p\) must be small enough to ensure that at least one (and preferably many) discretization points lie between the lower and upper bounds for \(x_{j}\). Thus, \(p \le P\) is the absolute minimum requirement, but feasibility is more likely as \(p\) is decreased. Note that these guidelines do not guarantee feasibility of (\(\mathbf{P}^{\prime }\)) in all cases, but represent the minimum level of precision needed given reasonable bounds on \(x_{j}\).

3 Lower bounding with multiparametric disaggregation

To obtain a lower bounding problem using multiparametric disaggregation, we first note that in the discretized problem, there always exists a gap between discretization points for a finite \(p\). Thus, we can introduce a slack variable \(\Delta x_{j}\) such that \(x_{j}^{R}=x_{j}^{\prime }+\Delta x_{j}\), where \(x_{j}^{\prime }\) is the discretized representation of \(x_{j}\) from Sect. 2, \(x_{j}^{R}\) is the continuous representation of \(x_{j}\) and the slack variable \(\Delta x_{j}\) is bounded between 0 and \(10^{p}\).

Again switching to the notation \(w = u \cdot v\) for the bilinear term, we have for the continuous representation of \(v\), denoted as \(v^{R}\):

$$\begin{aligned}&v^{R}=\mathop \sum \limits _{\ell =p}^P\mathop \sum \limits _{k=0}^9 10^{\ell }\cdot k\cdot z_{k,\ell } +\Delta v\end{aligned}$$
(15)
$$\begin{aligned}&0\le \Delta v\le 10^{p} \end{aligned}$$
(16)

For the continuous representation of the bilinear term, \(w^{R}\), note that:

$$\begin{aligned} w^{R}=u\cdot v^{R}=u\cdot \left( {v^{\prime }+\Delta v} \right)=w^{\prime }+u\cdot \Delta v=w^{\prime }+\Delta w \end{aligned}$$
(17)

where \(w^{{\prime }}\) and \(v^{\prime }\) are given by (1314). The slack variable \(\Delta w\) replaces the bilinear term \(u\cdot \Delta v\) that can be relaxed using the McCormick envelope, (1819), which in this case coincides with the RLT bound factor products \(u^{L}\le u\le u^{U}\) and \(0\le \Delta v\le 10^{p}\):

$$\begin{aligned}&u^{L}\cdot \Delta v\le \Delta w\le u^{U}\cdot \Delta v\end{aligned}$$
(18)
$$\begin{aligned}&\left( {u-u^{U}} \right)\cdot 10^{p}+u^{U}\cdot \Delta v\le \Delta w\le \left( {u-u^{L}} \right)\cdot 10^{p}+u^{L}\cdot \Delta v \end{aligned}$$
(19)

Introducing these constraints into Problem (P), and expressing the variables in terms of the original variables \(w_{ij}=x_{i} \cdot x _{j}\), we obtain the new optimization problem, (PR):

$$\begin{aligned} \text{ Min}\,\, z^{R}=f_0 =\mathop \sum \limits _{(i,j)\in BL_0 } a_{ij0} w_{ij} +h_0 (x) \end{aligned}$$

subject to

$$\begin{aligned}&f_q (x)=\mathop \sum \limits _{(i,j)\in BL_q } a_{ijq} w_{ij} +h_q (x)\le 0\quad q\in Q\backslash \{0\}\\&w_{ij} =\mathop \sum \limits _{\ell =p}^P\mathop \sum \limits _{k=0}^9 10^{\ell }\cdot k\cdot \hat{{x}}_{ijk\ell } +\Delta w_{ij} \quad \forall (i,j)\in BL_q ,q\in Q \end{aligned}$$
$$\begin{aligned} x_j =\mathop \sum \limits _{\ell =p}^P\mathop \sum \limits _{k=0}^9 10^{\ell }\cdot k\cdot z_{jk\ell } +\Delta x_j \quad \forall j\in \{j|(i,j)\in BL_q ,q\in Q\} \end{aligned}$$
(PR)
$$\begin{aligned}&x_i =\mathop \sum \limits _{k=0}^9\hat{{x}}_{ijk\ell } \quad \forall (i,j)\in BL_Q ,q\in Q,\ell \in L \\&x_i^L \cdot z_{jk\ell } \le \hat{{x}}_{ijk\ell } \le x_i^U \cdot z_{jk\ell } \quad \forall (i,j)\in BL_q ,q\in Q,\ell \in L,k\in K\\&\mathop \sum \limits _{k=0}^9z_{jk\ell } =1 \quad \forall j\in \{j|(i,j)\in BL_q ,q\in Q\},\quad \ell \in L\\&\left. {{\begin{array}{l} {\begin{array}{l} {x_i^L \cdot \Delta x_j \le \Delta w_{ij} \le x_i^U \cdot \Delta x_j } \\ {\Delta w_{ij} \le \left( {x_i -x_i^L } \right)\cdot 10^{p}+x_i^L \cdot \Delta x_j } \\ \end{array}} \\ {\Delta w_{ij} \ge \left(x_i -x_i^U \right)\cdot 10^{p}+x_i^U \cdot \Delta x_j } \\ \end{array} }} \right\} \forall (i,j)\in BL_q ,q\in Q\\&0\le \Delta x_j \le 10^{p}\quad \forall j\in \{j|(i,j)\in BL_q ,q\in Q\}\\&\quad z_{jk\ell } \in \{0,1\}\quad \forall j\in \{j|(i,j)\in BL_q ,q\in Q\},k\in K,\ell \in L\\&\quad x\in S\mathop \cap \Omega \subset \mathbb R ^{n} \end{aligned}$$

While (PR) does not exactly represent the product \(w_{ij} = x_{i} \cdot x_{j}\) and is feasible for values of \(w_{ij}, x_{i}\), and \(x_{j}\) that do not satisfy \(w_{ij} = x_{i} \cdot x _{j}\), the bilinear term is feasible in (PR). Thus, (PR) is a relaxation of (P). The relaxed feasible region resulting from (PR) can be seen in Figs. 3 and 4.

Fig. 3
figure 3

Plot of the lower bounding feasible region for \(w_{ij} = 0.1\) and \(p=P = -1\). The solid line is the true curve, while the dotted lines represent the boundaries of the relaxed feasible region of (PR). Note the similarities to piecewise McCormick envelopes

Fig. 4
figure 4

Feasible region of the relaxed problem (PR). The surface is the exact bilinear curve, while the envelopes represent the relaxed feasible region resulting from the lower bounding formulation for \(p=P = -1\)

The following property can be readily established from the above discussion:

Property 1

The solution of problem (PR) yields a lower bound for problem (P), i.e. \(z^{R} \le z\).

4 Discussion of global optimization algorithms

The upper and lower bounding schemes described can be combined into a global optimization algorithm. First, the following property can be established:

Property 2

As \(p\) approaches\(-\infty , z^{\prime }\) approaches \(z^{R}\).

Proof

As \(p\) approaches –\(\infty \) in (PR)

$$\begin{aligned}&\mathop {\text{ lim}}\limits _{p\rightarrow -\infty } \Delta x_j =\mathop {\text{ lim}}\limits _{p\rightarrow -\infty } 10^{p}=0\\&\mathop {\text{ lim}}\limits _{p\rightarrow -\infty } \Delta w_{ij} =x_i^U \cdot \mathop {\text{ lim}}\limits _{p\rightarrow -\infty } \Delta x_j =\left( {x_i -x_i^L } \right)\cdot \mathop {\text{ lim}}\limits _{p\rightarrow -\infty } 10^{p}+x_i^L \cdot \mathop {\text{ lim}}\limits _{p\rightarrow -\infty } \Delta x_j =0 \end{aligned}$$

Thus, since the variables \(\Delta x_j \) and \(\Delta w_{ij} \) are eliminated, this yields problem (\(\mathbf{P}^{\prime }\)), and hence \(z^{\prime }\) approaches \(z^{R}\). \(\square \)

From Property 2, we can establish that as precision is increased (i.e. \(p\) approaches\(-\infty )\), both (\(\mathbf{P}^{\prime }\)) and (PR) converge to the same value. Assuming \(P\) is large enough such that \(10^{P} \ge x_j^U \), we can further state that (\(\mathbf{P}^{\prime }\)) and (PR) converge such that \(z^{\prime }=z^{R}=z\).

4.1 Algorithm 1

The first global optimization algorithm that can be established from the aforementioned upper and lower bounding is as follows. First, we start at some coarse level of discretization such that \(P \ge p\), and solve both (PR) and (\(\mathbf{P}^{\prime }\)). If the difference in solutions to the upper and lower bounding problems is sufficiently small, then the algorithm terminates; otherwise, precision is increased and the problems are resolved. The algorithm is then as follows:

Algorithm 1

  1. Step 0.

    Choose \(p=P=\lfloor \log _{10} x_j^U \rfloor \)

  2. Step 1.

    Solve (PR) to obtain the lower bound \(z^{R}\).

  3. Step 2.

    Solve (\(\mathbf{P}^{\prime }\)) to obtain the upper bound \(z^{\prime }\). If (\(\mathbf{P}^{\prime }\)) is infeasible, let \(z^{\prime } = +\infty \).

  4. Step 3.

    If (\(z^{\prime } - z^{R})/z^{R} \le \varepsilon \), STOP, the solution is globally optimal. Otherwise, set \(p= p - 1\), and return to step 1.

4.2 Algorithm 2

While Algorithm 1 follows most naturally from the problems (\(\mathbf{P}^{\prime }\)) and (PR), it has several shortcomings. Notably, because (\(\mathbf{P}^{\prime }\)) and (PR) are fairly similar and increase similarly in problem size as precision is added, solving (\(\mathbf{P}^{\prime }\)) and (PR) repeatedly becomes increasingly expensive. Thus, an alternative method for obtaining an upper bound is to use a local NLP algorithm in place of solving the problem (\(\mathbf{P}^{\prime }\)).

Algorithm 2

  1. Step 0.

    Choose \(p=P=\lfloor \log _{10} x_j^U \rfloor \)

  2. Step 1.

    Solve (PR) to obtain the lower bound \(z^{R}\).

  3. Step 2.

    Solve (P) using a local NLP algorithm to obtain some upper bound \(z\) using the solution to (PR) as a starting point.

  4. Step 3.

    If (\(z - z^{R})/z^{R} \le \varepsilon \), STOP, the solution is globally optimal. Otherwise, set \(p= p - 1\) and return to step 1.

Algorithm 2 is generally more computationally efficient than Algorithm 1, as the solution of (P) using a local NLP algorithm is obtained much faster than the increasingly large MILP that (\(\mathbf{P}^{\prime }\)) becomes as \(P-p\) grows.

4.3 Extensions to MINLP

The algorithms in Sects. 4.1 and 4.2 can be readily extended for solving MINLPs with the following general form:

$$\begin{aligned} \text{ Min}\,\,z=f_0 (x,y) \end{aligned}$$

subject to

$$\begin{aligned} f_q (x,y)\le 0\quad q\in Q\backslash \{0\} \end{aligned}$$
$$\begin{aligned} f_q (x,y)=\mathop \sum \limits _{(i,j)\in BL_q } a_{ijq} x_i x_j +h_q (x,y)\quad q\in Q \end{aligned}$$
(P-MINLP)
$$\begin{aligned}&x\in S\mathop \cap \Omega \subset \mathbb R ^{n}\\&y\in \{0,1\} \end{aligned}$$

where \(h_{q}(x,y)\) is jointly convex in \(x\) and \(y\). For most cases in practice the variables \(y\) appear in linear form in these terms [34].

Analogous problems (\(\mathbf{P}^{\prime }\)-\(\mathbf{MINLP}\)) and (PR-MINLP) can be derived, and Algorithm 1 can be used without modification. However, Algorithm 2 requires some modification, as (P-MINLP) cannot be solved using a local NLP algorithm because it is an MINLP. A heuristic can be used to compute the upper bound as in Algorithm 2:

Algorithm 3

  1. Step 0.

    Choose \(p=P=\lfloor \log _{10} x_j^U \rfloor \)

  2. Step 1.

    Solve (PR-MINLP) to obtain \(z^{R}\).

  3. Step 2.

    Fix the binary variables \(y\) in (P-MINLP) to the values found by the solution of (PR-MINLP) in Step 1, reducing it to an NLP. Solve (P-MINLP) with these fixed binary variables using a local NLP algorithm to obtain some \(z\) using the solution to (PR-MINLP) as a starting point.

  4. Step 3.

    If (\(z - z^{R})/z^{R} \le \varepsilon \), STOP, the solution is globally optimal. Otherwise, set \(p= p - 1\) and return to step 1.

A disadvantage of this algorithm is that it could take, in the worst case, an infinite number of iterations to converge. By fixing the binary variables to that of the solution to (PR-MINLP), (P-MINLP) can be rendered infeasible, and algorithm would continue until (PR-MINLP) has enough discretization points to exactly represent (P-MINLP). If this occurs, heuristics or other approaches may be utilized to obtain an upper bound in place of solving (P-MINLP) with fixed binary variables. However, in practice, this is unlikely to occur.

5 Comparison with piecewise McCormick envelopes

A common approach to solving bilinear programs of the form (P) is to bound the bilinear terms using McCormick convex envelopes [35]. This formulation results in a lower bounding LP if the only nonlinearities are bilinear, or a lower bounding convex NLP if there are remaining convex nonlinearities. For each bilinearity \(w_{ij}=x_{i} \cdot x_{j}\), we introduce instead the following constraints:

$$\begin{aligned} {\begin{array}{l} {w_{ij} \ge x_i \cdot x_j^L +x_i^L \cdot x_j -x_i^L \cdot x_j^L } \\ {w_{ij} \ge x_i \cdot x_j^U +x_i^U \cdot x_j -x_i^U \cdot x_j^U } \\ {w_{ij} \le x_i \cdot x_j^L +x_i^U \cdot x_j -x_i^U \cdot x_j^L } \\ {w_{ij} \le x_i \cdot x_j^U +x_i^L \cdot x_j -x_i^L \cdot x_j^U } \\ \end{array}}\quad \forall (i,j)\in BL_q ,q\in Q \end{aligned}$$
(20)

This relaxation yields a lower bound on problem (P). However, this lower bound can be weak depending on the bounds on \(x_{i}\) and \(x_{j}\). To improve the quality of the lower bound, these convex envelopes can be used on discretized portions of the variable range. Thus, piecewise McCormick envelopes can be introduced in (P) to obtain a tighter lower bound at the cost of becoming an MILP or convex MINLP [10, 17, 36]. The envelopes, defined over a set of \(N\) points on the domain of variable \(x_{j}\), can be represented by the following disjunctive constraints:

$$\begin{aligned}&{\vee _{n=1}^{{N}}}\left[ {{\begin{array}{c} {{\begin{array}{c} {y_{jn} } \\ {w_{ij} \ge x_i \cdot x_{jn}^L +x_i^L \cdot x_j -x_i^L \cdot x_{jn}^L \quad \forall \{i|(i,j)\in BL_q ,q\in Q\}} \\ \end{array} }} \\ {w_{ij} \ge x_i \cdot x_{jn}^U +x_i^U \cdot x_j -x_i^U \cdot x_{jn}^U \quad \forall \{i|(i,j)\in BL_q ,q\in Q\}} \\ {w_{ij} \ge x_i \cdot x_{jn}^L +x_i^U \cdot x_j -x_i^U \cdot x_{jn}^L \quad \forall \left\{ {i\text{|}\left( {i,j} \right)\in BL_q ,q\in Q} \right\} } \\ {w_{ij} \ge x_i \cdot x_{jn}^U +x_i^L \cdot x_j -x_i^L \cdot x_{jn}^U \quad \forall \{i|(i,j)\in BL_q ,q\in Q\}} \\ {x_{jn}^L \le x_j \le x_{jn}^U } \nonumber \\ \end{array} }} \right] ~ \forall \{j|(i,j) \in BL_q ,q\in Q\} \nonumber \\&x_i^L \le x_i \le x_i^U \quad \forall \{i|(i,j)\in BL_q ,q\in Q\}\nonumber \\&\left. {{\begin{array}{l} {x_{jn}^L =x_j^L +\left( {x_j^U -x_j^L } \right)\cdot (n-1)/N} \\ {x_{jn}^U =x_j^L +\left( {x_j^U -x_j^L } \right)\cdot n/N} \\ \end{array} }} \right\} \quad \forall \left\{ {j\text{|}\left( {i,j} \right)\in BL_q ,q\in Q} \right\} ,\quad n\in \{1,\ldots ,N\}\nonumber \\ \end{aligned}$$
(21)

Applying the convex hull reformulation [33] for the above disjunctive constraints yields

$$\begin{aligned}&\left. {{\begin{array}{l} {{\begin{array}{l} {{\begin{array}{l} {{\begin{array}{l} {w_{ij} \ge \mathop \sum \nolimits _{n=1}^N \left( {\hat{{x}}_{ijn} \cdot x_{jn}^L +x_i^L \cdot \mathop x_{jn} -x_i^L \cdot x_{jn}^L \cdot y_{jn} } \right)} \\ \end{array} }} \\ {w_{ij} \ge \mathop \sum \nolimits _{n=1}^N \left( {\hat{{x}}_{ijn} \cdot x_{jn}^U +x_i^U \cdot {\hat{x}}_{jn} -x_i^U \cdot x_{jn}^U \cdot y_{jn} } \right)} \\ {w_{ij} \le \mathop \sum \nolimits _{n=1}^N \left( {\hat{{x}}_{ijn} \cdot x_{jn}^L +x_i^U \cdot {\hat{x}}_{jn} -x_i^U \cdot x_{jn}^L \cdot y_{jn} } \right)} \\ \end{array} }} \\ {w_{ij} \le \mathop \sum \nolimits _{n=1}^N \left( {\hat{{x}}_{ijn} \cdot x_{jn}^U +x_i^L \cdot {\hat{x}}_{jn} -x_i^L \cdot x_{jn}^U \cdot y_{jn} } \right)} \\ \end{array} }} \\ {x_i =\mathop \sum \nolimits _{n=1}^N \hat{{x}}_{ijn} } \\ \end{array} }} \right\} \forall \left( {i,j} \right)\in BL_q ,q\in Q\nonumber \\&\left. {{\begin{array}{l} {x_j =\mathop \sum \nolimits _{n=1}^N {\hat{x}}_{jn} } \\ {\mathop \sum \nolimits _{n=1}^N y_{jn} =1} \\ \end{array} }} \right\} \forall \left\{ {j|\left( {i,j} \right)\in BL_q ,q\in Q} \right\} \nonumber \\&x_i^L \cdot y_{jn} \le {\hat{x}}_{ijn} \le x_i^U \cdot y_{jn} \forall \left( {i,j} \right)\in BL_q, \quad q\in Q,n\in \left\{ {1,\ldots ,N} \right\} \nonumber \\&\left. {{\begin{array}{l} {x_{jn}^L \cdot y_{jn} \le \hat{{x}}_{jn} \le x_{jn}^U \cdot y_{jn} } \\ {y_{jn} \in \left\{ {0,1} \right\} } \\ \end{array} }} \right\} \forall \left\{ {j|\left( {i,j} \right)\in BL_q ,q\in Q} \right\} ,\quad n\in \left\{ {1,\ldots ,N} \right\} \end{aligned}$$
(22)

By adding these piecewise McCormick envelope constraints (22) to problem (P), we can define a new relaxed MILP or convex MINLP, (PR-PCM). Furthermore, we can compare the performance of (PR-PCM) to the lower bounding problem derived from multiparametric disaggregation, (PR).

6 Illustrative example (P1)

We first consider as an example the bilinear program by Quesada and Grossmann [37] and originally reported by Al-Khayyal and Falk [18]:

$$\begin{aligned} \text{ Min}\,\,f=-x_1 +x_1 x_2 -x_2 \end{aligned}$$

subject to

$$\begin{aligned} -6x_1 +8x_2 \le 3 \end{aligned}$$
(P1)
$$\begin{aligned} 3x_1 -x_2 \le 3\\ 0\le x_1 ,x_2 \le 1.5 \end{aligned}$$

The global optimum of this bilinear program is \(f = -1.08333\) at (1.165864, 0.497580). Two other local minima correspond to: \(f = -1.0\) at (1, 1), and \(f = -1.005\) at (0.917, 1.062).

6.1 Lower bounding problems

In order to compare the lower bounds predicted by multiparametric disaggregation and piecewise McCormick envelopes, we solve the relaxation problems (PR) and (PR-PCM) resulting from multiparametric disaggregation and piecewise McCormick, respectively. For the specific example Problem (P1), we can derive analogous relaxed problems (P1R) and (P1R-PCM) using the MDT and piecewise McCormick envelopes, respectively. Note that \(x_{1}\) is the variable being discretized.

Problem size and computational results are shown in Table 1 for the lower bounds predicted by (P1R) at \(P = 0\) and \(p = \{0, -1,{\ldots },-6\}\) and (P1R-PCM) at \(N = \{1, 15, 150, 1{,}500, 15{,}000\}\). Solving (P1R-PCM) at various levels of discretization, it becomes clear that problem size increases approximately exponentially with each order of magnitude decrease in the optimality gap. However, this is not the case when solving (P1R) since as precision is added, i.e. \(p\) is decreased, a linear relationship holds. Note that the discretization level of MDT at \(p=-1\) matches that of PCM at \(N=15 (p=-2\) is equivalent to \(N=150\) and so on) and that the lower bounds are exactly the same. Upper bounds are also reported by using as a starting point the solution of the lower bounding relaxation problem as in Algorithm 2 in Sect. 4.2. Thus, the MDT columns are equivalent to the iterations of the algorithm with the reported CPU time being the total time required for the lower and upper bounding problems. These times would need to be accumulated to compare Algorithm 2 directly to BARON, unless the discretization level (\(p\)) is chosen a priori. Results for solving the original NLP (P1) with the global optimization solver BARON are reported in the second column.

Table 1 Comparison between multiparametric disaggregation (MDT) and piecewise McCormick (PCM), problem (P1). For MDT, \(P= 0\) is used

As seen in Table 1, the relaxed problems (P1R) and (P1R-PCM) are considerably larger than the original NLP (P1) since the addition of both continuous and binary variables increases problem size. However, note that the multiparametric disaggregation problem (P1R) requires fewer additional continuous variables and fewer constraints than when utilizing piecewise McCormick envelopes in (P1R-PCM). For a single McCormick envelope, a relatively weak lower bound of \(-1.5\) is obtained leading to an upper bound that is in fact a suboptimal solution. Multiparametric disaggregation with \(p=-4\) approaches an optimality gap of 0.003 %, while piecewise McCormick envelopes, even with 1,500 partitions, only approach an optimality gap of 0.034 %. Further discretization and refinement of the solution is quickly solved using multiparametric disaggregation, as the \(p=-5\) and \(p=-6\) problems are solved in \(<\)1.2 s and fewer than 130 variables. In contrast, increasing the number of partitions to 15,000 in (P1R-PCM), leads to a MILP that cannot be solved in 1-h of computational time.

7 Numerical experiments

The performance of the underestimating problems from multiparametric disaggregation (PR) and piecewise McCormick (PR-PCM) is further evaluated through the solution of three additional small test problems from the literature for different accuracy levels. Since in all problems the functions \(h_{q}(x)\) in (P) are linear, the resulting bounding MILP problems were solved in GAMS 23.8.2 [38] using CPLEX 12.4 (1 thread) [39]. Default options were used except for the relative optimality tolerance, equal to 10\(^{-6}\) and a maximum computational effort equal to 3,600 CPU seconds. Similarly as in problem (P1), the original nonlinear programs were solved by CONOPT 3 [40], following initialization with the values from the MILP, and by BARON 10.2 [41]. We also report results for larger problems in the areas of water networks and multiperiod blending. Such problems were also solved by GloMIQO 1.0.0 [49]. The computational experiments were performed on an Intel i7 950 processor running at 3.07 GHz, with 8 GB of RAM, running Windows 7.

7.1 Small test problems

7.1.1 P2

Problem (P2) is originally from the compilation of test problems by Rijckaert and Martens [42] but has been converted to a bilinear program following simple transformations and addition of new variables. The global optimal solution is \(f = 10122.4932\) at (78, 33, 29.995740, 45, 36.775327, 0.030303, 27.264982, 0.033338). Variables \(x_{1}, x_{2}\) and \(x_{3}\) are selected for parameterization/partitioning.

$$\begin{aligned} \min&f=5.3578x_3^2 +0.8357x_1 x_5 +37.2392x_1 \\ s.t.&0.00002584x_3 x_5 -0.00006663x_2 x_5 -0.0000734x_1 x_4 \le 1 \\&0.000853007x_2 x_5 +0.00009395x_1 x_4 -0.00033085x_3 x_5 \le 1 \\&1330.3294x_6 -0.42x_1 -0.30586x_7 \le x_5 \\&0.00024186x_2 x_5 +0.00010159x_1 x_2 +0.00007379x_3^2 \le 1 \\&2275.1327x_8 -0.2668x_1 -0.40584x_4 \le x_5 \\&0.00029955x_3 x_5 +0.00007992x_1 x_3 +0.00012157x_3 x_4 \le 1 \\&1=x_2 x_6 \\&x_7 x_2 =x_3^2 \\&1=x_8 x_3 \\&78\le x_1 \le 102,\quad 33\le x_2 \le 45,\quad 27\le x_3 \le 45,\quad 27\le x_4 \le 45,\quad 27\le x_5 \le 45\\&\frac{1}{45}\le x_6 \le \frac{1}{33},\quad \frac{27^{2}}{45}\le x_7 \le \frac{45^{2}}{33},\quad \frac{1}{45}\le x_8 \le \frac{1}{27}. \end{aligned}$$

7.1.2 P3

Test problem (P3) corresponds to Problem 106 in Hock and Schittowski [43]. While the objective function and the first three constraints are linear, there are three bilinear inequalities. To study how the choice of parameterized/partitioning variables affects computational performance, three versions are considered: (a) 3 parameterized/partitioned variables, \(x_{1}, x_{2}\) and \(x_{3}\); (b) 4 variables, \(x_{2}, x_{5}, x_{6}\) and \(x_{8}\); (c) 5 variables, \(x_{4}, x_{5}, x_{6}, x_{7}\) and \(x_{8}\). The global optimal solution is \(f = 7049.2479\) at (579.573535, 1359.977872, 5109.696534, 182.039993, 295.612139, 217.960007, 286.427855, 395.612139).

$$\begin{aligned} \min&\;f=x_1 +x_2 +x_3 \\ s.t.&0.0025(x_4 +x_6 )-1\le 0 \\&0.0025(-x_4 +x_5 +x_7 )-1\le 0 \\&0.01(-x_5 +x_8 )-1\le 0 \\&100x_1 -x_1 x_6 +833.33252x_4 -83333.333\le 0 \\&x_2 x_4 -x_2 x_7 -1250x_4 +1250x_5 \le 0 \\&x_3 x_5 -x_3 x_8 -2500x_5 +1250000\le 0 \\&100\le x_1 \le 10000,\;1000\le x_2 ,x_3 \le 10000,\;10\le x_4 ,x_5 ,x_6 ,x_7 ,x_8 \le 100 \end{aligned}$$

7.1.3 P4

Test problem (P4) is taken from Shen and Zhang [44] and after a few transformations the following bilinear problem results. It features a global optimal solution \(f =460212.25\) at (43.165468, 45, 70, 1.228374, 27.749229, 0.814083). Variables \(x_{2}, x_{5}\) and \(x_{6}\) are chosen as the parameterized variables.

$$\begin{aligned} \min&f=168x_1 x_2 +3651.2x_5 +40000x_6 \\ s.t.&1.0425x_1 -x_2 \le 0 \\&0.00035x_1 x_2 -1\le 0 \\&1.25x_4 -x_1 +41.63\le 0 \\&x_1 x_2 -x_3 x_5 =0 \\&x_4 x_6 -1=0 \\&40\le x_1 \le 44,\;40\le x_2 \le 45,\;60\le x_3 \le 70,\;0.1\le x_4 \le 1.4 \\&22.85714\le x_5 \le 33,\;0.714286\le x_6 \le 10 \end{aligned}$$

7.2 Computational statistics

Tables 1, 2, 3 and 4 give the computational results for problems (P1P4) as a function of the discretization level, and the columns correspond to the iterations of Algorithm 2, where the relaxation problem is either generated by multiparametric disaggregation (PR), MDT columns, or by the piecewise McCormick envelopes (PR-PCM), PCM columns. We provide the problem size, lower bound from the relaxation problem, and upper bound from a local NLP solver (the values of the latter remain the same independent of the accuracy level). The optimality gap and total computational effort (CPLEX plus CONOPT, the latter being almost negligible) are also reported. Note that the triplets in Table 2 are related to the number of discrete points for the three variables selected. The n-tuples in Tables 3 and 4 have a similar meaning. In problems (P2P3), the discretization points of PCM match exactly those being used by MDT (e.g. second column of PCM should be compared with second column of MDT).

Table 2 Comparison between multiparametric disaggregation (MDT) and piecewise McCormick (PCM), problem (P2)
Table 3 Comparison between multiparametric disaggregation (MDT) and piecewise McCormick (PCM), problem (P3)
Table 4 Comparison between multiparametric disaggregation (MDT) and piecewise McCormick (PCM), problem (P4)

An interesting observation is that for the same accuracy level, the relaxation from piecewise McCormick (PCM) is tighter than MDT in (P2) but identical in (P1) and (P3). If one looks at the original NLP problems, the latter problems have strictly bilinear terms whereas (P2) features a quadratic term in \(x_{3}\), which is one of the variables being discretized. Applying the piecewise relaxation to a quadratic term \(w_{ii} =x_i \cdot x_i\) makes it possible to use tighter bounds for both parts of \(x_i \) in each partition \(n\), see (23), whereas with MDT just one part is discretized, meaning the use of the overall bounds for the other part. This provides the explanation for multiparametric disaggregation being less tight than piecewise McCormick, which can be observed in Fig. 5.

$$\begin{aligned} \vee ^N_{n=1} {\begin{array}{c} {\left[ {{\begin{array}{c} {{\begin{array}{c} {y_{in} } \\ {w_{ii} \ge x_i \cdot x_{in}^L +x_{i,n}^L \cdot x_i -x_{in}^L \cdot x_{in}^L } \\ \end{array} }} \\ {w_{ii} \ge x_i \cdot x_{in}^U +x_{i,n}^U \cdot x_i -x_{in}^U \cdot x_{in}^U } \\ {w_{ii} \le x_i \cdot x_{in}^L +x_{i,n}^U \cdot x_i -x_{in}^U \cdot x_{in}^L } \\ {x_{in}^L \le x_i \le x_{in}^U } \\ \end{array} }} \right]} \\ {x_{in}^L =x_i^L +\left( {x_i^U -x_i^L } \right)\cdot (n-1)/N} \\ {x_{in}^U =x_i^L +\left( {x_i^U -x_i^L } \right)\cdot n/N} \\ \end{array} } \end{aligned}$$
(23)

For (P4), the lower bounds of \(x_{5}\) and \(x_{6}\) are not integer values and so, while the number of discrete points is roughly the same, their location is not, making it possible for the relaxation from MDT to be better than the one from PCM. For instance, \(p=(-2,-2,-2)\) leads to a slightly better lower bound than \(N=(500,1,000,1,000)\). Due to the better performance of the new method, there are more columns for MDT than for PCM, meaning that higher accuracy levels, i.e. lower optimality gaps, can be achieved by the former for a given computational time.

Fig. 5
figure 5

Feasible region from conventional McCormick, piecewise McCormick (PCM) and multiparametric disaggregation relaxation (MDT) for a quadratic term over a specific region. Notice that the underestimation from PCM is strictly tighter than MDT in the region \(x_i \in ]25,35[\backslash \{30\}\)

While MDT is not as tight as PCM, the latter generates considerably larger MILP problems for the same accuracy level. More specifically, with PCM we get exactly an exponential increase in the number of binary variables with the change of power and roughly the same behavior with respect to the number of total variables and constraints. In other words, the number of 0–1 variables grows linearly with the number of partitions. In contrast, the number of 0–1 variables for MDT grows logarithmically with the number of discrete points and linearly with the change of power, keeping problems tractable for a wider accuracy range. Notice that with MDT all problems can be solved with an optimality gap of less than 0.01 % in less than one hour, while PCM closes only to a gap of 6.63 % for (P3).

Since problem size is related to the number of parameterized/partitioned variables, one might be tempted to keep this number as low as possible. However, the results for (P3) give opposite results, since the worst performance is obtained for 3 parameterized variables (gap \(=\) 0.009 %), followed by 4 (gap \(=\) 0.0067 % in 1898 CPUs) and then 5 (gap \(=\) 0.0071 % in 520 CPUs, which improves to just 0.0020 % for \(p = (-3,{\ldots },-3)\), calculated using the best possible solution at time of termination).

7.2.1 Evaluation of Algorithm 1

Compared to Algorithm 2 in Tables 2, 3 and 4, Algorithm 1 generally leads to larger optimality gaps, particularly in the first iterations at coarse discretization levels, and it also requires more CPU time since two MILPs are solved at each iteration. Comparatively, solving (P) with a local solver can be done almost instantaneously. The lower bounding problem is the same as in Algorithm 2, while the upper bounding problem (\(\mathbf{P}^{\prime }\)) yields considerably worse bounds than the ones returned from the local NLP solver (recall that these were always global optimal solutions). As is illustrated in Fig. 6 for test problem (P3a), the solutions from both (\(\mathbf{P}^{\prime }\)) and (PR) typically become closer to the optimum with an increase in accuracy. However, it is clear that using a local NLP solver, as in Algorithm 2, is a superior approach. While it is possible for the solution of (\(\mathbf{P}^{\prime }\)) to be closer to the global optimum than the solution of (P) using a local NLP algorithm, in practice, this generally does not occur, especially when using the solution of (PR) as a starting point.

Fig. 6
figure 6

Comparison of Algorithms 1 and 2 for problem (P3a). The upper bound from Algorithm 1 for \(p = 4\) is equal to 30,000 (off scale)

7.3 Results for larger problems

As seen in Tables 1, 2, 3 and 4, when compared to the commercial solver BARON, multiparametric disaggregation is competitive in (P1), (P2) and (P4) but is orders of magnitude slower in (P3). This is to be expected given the very small size of problems and the reduced number of bilinear terms, which facilitates the spatial branch-and-bound procedure in BARON. In order to evaluate how both methods scale with problem size, we solved the five most difficult (higher optimality gap at termination when solved with BARON) water-using network design problems in Teles et al. [11], which also correspond to bilinear programs. In such problems, the discretized variables are concentrations featuring \(P = 3\) in Ex17 and \(P = 2\) in the other examples. The relative optimality tolerance was set to 0.01 % and the total computational time to 1-h.

Table 5 lists the results for global optimization solvers BARON and GloMIQO together with the implementation of algorithm 2 for MDT. While BARON fails to prove optimality in all five cases, GloMIQO and MDT are successful in two, which interestingly are not the same. GloMIQO takes roughly half an hour to tackle Ex14 and Ex15, while it takes just a few seconds for MDT to solve EX17 and Ex18. In the full set of problems from [11], MDT performs better than BARON 84 % of the time and slightly better than GloMIQO (53 %) when considering the CPUs performance metric. The average optimality gaps are however considerably lower, 0.088 % for MDT, 0.466 % for GloMIQO and 0.711 % for BARON.

Table 5 Comparison for water-using networks design problems [11]

In Table 6, several blending problems [45, 46] are solved using BARON, GloMIQO, and Algorithm 3 at a single level of discretization. These are multiperiod blending problems with varying numbers of tanks, time periods, and product qualities. Each problem is identified such that 6T-3P-2Q-029 is a 6 tank, 3 time period, 2 quality problem, with a unique 3-digit identifier to distinguish it from other problems of the same size. The results for MDT are reported for \(P = 0\) and \(p = -3\) and were solved using Gurobi [47] using 12 threads. As these problems are originally MINLPs, larger than those in Table 5, and 12 threads are utilized, the computational difference is much more significant. GloMIQO and BARON both use MILP solvers in their algorithms which can utilize multiple threads, but this effect is not significant as the subproblems being solved are generally very small. Because of the use of multiple threads and multiple CPUs in this comparison, wall times are reported for all three methods. While GloMIQO and BARON are unable to converge to an optimality gap of 0.1 % within the time limit of 2 h, the MDT is able to close the gap in all but two cases. In these two cases an additional level of discretization would close the gap, but these results show that if a reasonably fine discretization is chosen a priori, the optimality gaps can be significantly less than those of commercial global optimization solvers. For the one problem that BARON and GloMIQO were able to close the gap within the time limit, MDT outperformed them by approximately 17 and 1,400 times, respectively, although the problem is small enough and is solved fast enough that meaningful conclusions should not be drawn from this result alone.

Table 6 Comparison for multiperiod blending problems [45, 46]

8 Conclusions

This paper has presented a derivation based on disjunctive programming and exact linearization of the multiparametric disaggregation technique (MDT) for solving bilinear programming problems. The derivation, which was shown to be equivalent to applying the reformulation linearization technique (RLT), gives rise to an MILP approximation that yields an upper bound if feasible with respect to the original problem. A lower bounding MILP relaxation problem has also been derived, which can then be used as basis for rigorous global optimization algorithms. As has been shown with the smaller test problems, the relaxation from the MDT was shown to be as tight as the relaxation based on piecewise McCormick envelopes (PCM) for strictly bilinear terms but weaker for quadratic terms. The real advantage of MDT comes from the more favorable scaling of problem size as discretization is increased, which is translated into an ability to reach considerably lower optimality gaps and a clearly better computational performance. For large problems it was shown that multiparametric disaggregation can outperform global solvers BARON and GloMIQO.