Abstract
This paper proposes a parametric method for solving a generalized fractional programming problem which is called sum-of-ratios problem. The sum-of-ratios problems occur in many fields including computer vision, finance, engineering and management. Compared with other methods based on branch-and-bound procedure, our algorithm is based on Newton-like method for solving a system of nonlinear equations with parameters and it needs to solve convex programming problem in each iteration. We showed the global linear and local superlinear/quadratic rate of convergence of the algorithm. We demonstrated the practical efficiency of the algorithm by numerical experiments for various kinds of sum-of-ratios problem. In the numerical experiments, our method exhibited better solution quality and better convergence rate than other methods.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
The sum-of-ratios problem, which is to minimize (maximize) a sum of several fractional functions in convex set, is a non-convex optimization problem that is difficult to solve by the traditional optimization methods. The sum-of-ratios problem has attracted great interest over the 40 years for its practical and theoretical significance. The problem occurs in many important applications such as finite element method, computer graphics, finance, engineering and management [5, 6, 12, 15, 17]. In [12], many problems in projective geometry such as multi-view triangulation, camera resectioning and homography estimation were formulated as the sum-of-ratios problem and the branch-and-bound method,which is based on the recent developments in fractional programming and the theory of convex underestimators [2,3,4,5, 9, 14, 16, 17, 21, 24,25,26],was employed to find its global solution. In the method of [12], the number of variables increases as twice as the number of fractional functions and it requires to solve a second-order cone programming problem to obtain a lower bound of the optimal value in each iteration. In recent years, some algorithms has been developed for the global solution to these optimization problems [2, 3, 9, 11, 18, 19, 25]. Given any tolerance \(\varepsilon \), if the optimization problem is feasible, the algorithms return a solution which is at most \(\varepsilon \) far from the global optimum. But the methods based on branch-and-bound procedure require a lot of computations and have low convergence rate.Furthermore, it is not easy to find a reasonable branching strategy.
In this paper, we propose an efficient algorithm which transforms the sum-of-ratios problem into parametric convex programming problem and solves the convex programming problem in each iteration. Unlike the methods based on the branch-and-bound, our method does not need to introduce new variables and constraints, and it also does not need to use additional complicated computations for obtaining the upper bounds and lower bounds for numerators and denominators of every ratio, so it is easier to implement than the previous methods. This paper is organized as follows. In Section 2,we show that the sum-of-ratios problem is equivalent to some kind of parametric convex programming problem. In Section 3, it is about a Newton-like algorithm for finding optimal parameters associated with the optimal solution of the sum-of-ratios problem and its convergence characteristics were amply considered. In Section 4, several test problems in the previous studies are used to show the advantage of our algorithm over the other methods, and the experiment results show that our algorithm has better performance than other methods. Finally, Section 5 provides some discussions.
2 Equivalent parametric convex programming
The sum-of-ratios problem is to minimize the sum of fractional functions subject to convex constraints, which is formulated as follows.
where \( F_i(x)=\frac{f_i(x)}{h_i(x)},i=1,\cdots ,p,\quad X=\{x\in R^n|g_i(x)\le 0, i=1,\cdots ,m\}\), and \(f_i(x),g_i(x)\) and \(-h_i(x)\) are twice continuously differentiable convex functions.
It is assumed that the feasible set X is bounded, \(f_i(x)\ge 0\) and \(h_i(x)>0\) for every \(x\in X\), and that \(intX=\{x\in R^n|g_i(x)< 0, i=1,\cdots ,m\}\ne \emptyset \) (Slater condition). Even with these restrictions the above problem is NP-complete [8].
It is easy to see that the problem (1) is equivalent to the following problem.
Let \( \beta =(\beta _1, \cdots ,\beta _p)\) and \( \gamma =(\gamma _1, \cdots ,\gamma _p)\).
Theorem 1.
If \(({\bar{x}},{\bar{\beta }})\) is the solution of the problem (2), then there exists \({\bar{\gamma }}\) such that \({\bar{x}}\) is a solution of the following convex programming problem for fixed \(\gamma ={\bar{\gamma }}\) and \(\beta ={\bar{\beta }}\).
And \({\bar{x}}\) also satisfies the following system of equations for \(\gamma ={\bar{\gamma }}\) and \(\beta ={\bar{\beta }}\):
Proof.
The constraint \(F_i(x)\le \beta _i\) is equivalent to \(f_i(x)-\beta _i h_i(x)\le 0\). Let’s define the following function for the problem (2).
By Fritz-John optimality condition (Theorem 4.2.8 of [1]), there exist \({\bar{w}}, {\bar{\gamma }}=(\bar{\gamma }_{1},\cdots ,\bar{\gamma }_{p})\) and \({\bar{v}}=(\bar{v}_{1},\cdots ,\bar{v}_{m})\) such that
Suppose that \({\bar{w}}=0\). Then, by (7), we have \({\bar{\gamma }}=0\) because \( h_i(x)>0, i=1,\cdots ,p\) for all \(x\in X\). Hence, it follows from (6), (9), (10) and (12) that
where \(I({\bar{x}})=\{i|g_i({\bar{x}})=0, 1\le i\le m \}\). By Slater condition, there exists a point \(x'\) such that
Since \(g_i(x),i=1,\cdots ,m\) are convex, it follows from (15) that
Letting \(d=x'-{\bar{x}}\), from (16) and (14), we have \(\left(\sum \limits _{i\in I({\bar{x}})}\bar{v}_{i} \nabla g_i({\bar{x}})\right) ^Td<0\), which contradicts (13). Thus, we have \({\bar{w}}>0\).
Denoting \(\frac{{\bar{\gamma }}}{{\bar{w}}}\) and \(\frac{{\bar{v}}}{{\bar{w}}}\) by \({\bar{\gamma }}\) and \({\bar{v}}\), respectively, we see that (7) is equivalent to (4) and so (8) is equivalent to (5) because \(\bar{\gamma }_{i}>0, i=1,\cdots ,p\) by (4).
Given \(\gamma ={\bar{\gamma }}\) and \(\beta ={\bar{\beta }}\), (6), (9) and (10) is just the KKT condition for the problem (3). Since the problem (3) is convex programming for parameters \(\gamma >0\) and \(\beta \ge 0\), the KKT condition is also sufficient optimality condition and then \({\bar{x}}\) is the solution of (3) for \(\gamma ={\bar{\gamma }}\) and \(\beta ={\bar{\beta }}\). \(\square \)
Let \(\alpha =(\beta ,\gamma )\) denote parameter vector and let
where
and it is obvious that \(\gamma ^l_i>0, i=1,\cdots ,p.\)
Remark 1.
-
(i)
Since the feasible set X is bounded by the assumption, the \(\beta ^u, \gamma ^l\) and \(\gamma ^u\) are well defined, and the set \(\Omega \) is bounded.
-
(ii)
The \(\beta ^u, \gamma ^l\) and \(\gamma ^u\) involved in the definition of \(\Omega \) are only necessary for theoretical consideration but in solving the problem (1) at all.
Let \(x(\alpha )\) be the solution of the problem (3) for fixed \(\alpha \in \Omega \) and let
Let
and consider the following system of nonlinear equations:
Corollary 1.
If \({\bar{x}}\) is the solution for the problem (1), then there exists \({\bar{\alpha }}\in \Omega \) such that \({\bar{x}}=x({\bar{\alpha }})\) and \({\bar{\alpha }}\in {\hat{\Omega }}\).
Proof.
The proof follows from Theorem 1 because (5) and (4) implies (19) and (20), respectively. \(\square \)
Corollary 2.
If the problem (1) has a solution \({\bar{x}}\) , then \({\hat{\Omega }}\ne \emptyset \). And if \({\hat{\Omega }}\) consists of only one point \(\alpha ^*\), i.e. \({\hat{\Omega }}=\{\alpha ^*\}\), then \(x(\alpha ^*)\) is the solution of the problem (1).
Proof.
It follows from Corollary 1 that \({\hat{\Omega }}\ne \emptyset \). If \({\hat{\Omega }}=\{\alpha ^*\}\), the \(x(\alpha ^*)\) is the solution for the problem (1) because \({\bar{x}}=x(\alpha ^*)\) by Corollary 1. \(\square \)
It follows from Corollary 2 that the problem (1) is reduced to find the solution \(x(\alpha )\) of the problem (3) satisfying (19) and (20) for parameter vector \(\alpha \). If \({\hat{\Omega }}=\{\alpha ^*\}\), then solving the problem (1) is equivalent to finding the \(\alpha ^*\in {\hat{\Omega }}\).
In the next section, we establish the uniqueness of \({\hat{\Omega }}\) under some assumptions and propose an algorithm to find the unique element of \({\hat{\Omega }}\), i.e. the global optimal solution for the problem (1).
3 Algorithm and its convergence
In this section, first, we consider the system of equations to be satisfied by the parameters corresponding to the optimal solution of the problem (1) in Lemma 2 and Lemma 3. We show the existence and uniqueness of the solution to the system (Theorem 2, Corollary 3) and then construct a Newton-like algorithm ((46) or (59)) to solve the system. Second, we prove the global linear and local superlinear/quadratic convergence of the algorithm (Corollary 4, Theorem 3). Finally, we show that the solution of the convex programming problem (3) corresponding to the optimal parameter vector, which is any accumulation point of the sequence generated by the algorithm, is the solution of the original problem (1) under some assumptions (Theorem 5). Let
Then (19) and (20) can be rewritten as
Now, let’s derive Jacobian matrix \(\psi '(\alpha )\) to construct a Newton-like algorithm. If \(x(\alpha )\) is differentiable, we have the followings by (17) and (18):
for \(i=1,\cdots ,p, j=1,\cdots ,2p\). Therefore, defining A and B as
we have
To compute A, we need to find derivatives of \(x(\alpha )\). Since \(x(\alpha )\) is the optimal solution of the problem (3) for given \(\alpha =(\beta ,\gamma )\), there exists a Lagrange multiplier vector \(v(\alpha )\ge 0\) such that \(x(\alpha )\) and \(v(\alpha )\) satisfy the following equations:
letting \(F(x,\alpha )=\sum _{i=1}^{p}\gamma _i(f_i(x)-\beta _ih_i(x))\),\(x\in X\),\(\alpha =(\beta ,\gamma )\), \( \nabla G(x(\alpha ))=(\nabla g_1(x(\alpha )),\cdots ,\nabla g_m(x(\alpha )))\), we have
Let
Then (28) and (29) can be written by
By differentiating (30) and (31) with variables \(\alpha _j, j=1,\cdots ,2p\), we have
where
Since \(g_i(x), i=1,\cdots ,m\) don’t involve any parameters, we have
Therefore, by denoting
we can write (32) as
Hence, if \(J(\alpha )\) is nonsingular, we have
Thus, we have the following result.
Lemma 1.
If \(\nabla _x^2F+(\nabla _x^2G^1)^Tv^1 \) is nonsingular, we have for every \(j=1,\cdots ,2p\)
where \(M=(\nabla G^1)^T[\nabla _x^2F+(\nabla _x^2G^1)^Tv^1]^{-1}\nabla G^1.\)
Proof.
If \(\nabla _x^2F+(\nabla _x^2G^1)^Tv^1 \) is nonsingular,we have
by (32) and (35). By substituting (38) in (33) and taking (35) into account, we have
Therefore, we have
By substituting (39) in (38), we have (37). \(\square \)
Remark 2.
Since \(\gamma >0\) and \(\beta \ge 0\) for every \(\alpha =(\beta ,\gamma )\in \Omega \), if there exists a \(k\in \{1,\cdots ,p\}\) such that \(f_k(x)\) is strictly convex, then \(\nabla _x^2 F\) is positive definite and \(\nabla ^2g_i(x), i\in I(x)\) are positive semidefinite because of the convexity of \(f_i(x), -h_i(x), i=1,\cdots ,p\) and \(g_i(x), i=1,\cdots ,m\). It follows from (34)that \(\nabla _x^2F+(\nabla _x^2G^1)^Tv^1\) is also positive definite, which shows that (38) is well-defined.
In the above, we exploited the differentiability of \(x(\alpha )\) and \(v(\alpha )\). Now, for the differentiability of \(x(\alpha )\) and \(v(\alpha )\), it needs the following assumption.
(H1) In the problem (3) for each \(\alpha \in \Omega \), LICQ(linear independence constraint qualification), second-order sufficiency condition and strict complementarity condition are satisfied at \(x(\alpha )\)
Remark 3.
The assumption (H1) is often required when applying the classical implicit function theorem to obtain differentiability of the optimal solution in parametric optimization problem. Under the assumption (H1), we have the existence and the uniqueness of differentiable \(x(\alpha )\) and \(v(\alpha )\) satisfying (28) and (29) and have the non-singularity of the matrix \(J(\alpha )\) (Theorem 2.1 of [7]). Therefore, it follows that the matrix M defined in Lemma 1 is nonsingular.
For the sake of convenient analysis for a Newton-like algorithm to be presented below, the following assumption is needed.
(H2)\(\nabla G^1\) is nonsingular for each \(\alpha \in \Omega \), i.e. \(|I(x(\alpha ))|=n\) for each \(\alpha \in \Omega \).
Remark 4.
The assumption (H2) implies that \(\frac{\partial x}{\partial \alpha _j}=0, j=1,\cdots ,2p\) from (33) and (35). Therefore, it follows from (26) and (27) that \(\psi '(\alpha )=B\). As we will show below, the assumption (H1) and (H2) is a sufficient condition under which the system \(\psi (\alpha )=0\) has a unique solution and Newton method for the system always finds the global minimizer of the initial problem. It may be possible to apply the Newton method for the problems which don’t satisfy the assumption (H1)and (H2).
Lemma 2.
Under the assumption (H1)and (H2), \(\psi (\alpha )\) is strongly monotone with constant \(\delta >0\) in \(\Omega \) , where
Proof.
By (26) and Remark 4, the Jacobian matrix of \(\psi (\alpha )\) is as follows.
Then \(\psi '(\alpha )\) is positive definite because \(x(\alpha )\in X\) and \(h_i(x(\alpha )) >0,i=1,\cdots ,p\). Therefore, for any \(z\in R^{2p}\), we have
which completes the proof. \(\square \)
Let
Lemma 3.
The equation \(\psi (\alpha )=0\) is equivalent to the equation \(\alpha =A(\alpha )\). If the problem (1) has an optimal solution, the equation (22) has at least one solution in \(\Omega \).
Proof.
The first proposition is obvious from the definition of \(A(\alpha )\) and \(\psi (\alpha )\). It follows from Corollary 1 that there is \(\alpha \in \Omega \) such that \(\psi (\alpha )=0\). \(\square \)
Let’s introduce a mapping \(B_\lambda (\alpha )=\pi _\Omega (\alpha -\lambda \psi (\alpha ))\) to establish the existence and uniqueness of the solution of (22), where \(\lambda \) is a positive constant and \(\pi _\Omega (a)\) denotes the projection of a onto \(\Omega \), i.e. \(\pi _\Omega (a)\) is the solution of the following problem:
Lemma 4.
For all \(a\in R^p\) and for all \(x\in \Omega \),
Proof.
See Kinderlehrer and Stampcchia [13]. \(\square \)
Lemma 5.
Assume that \(\psi (\alpha )\) is differentiable, and satisfies (40) and Lipschitz condition with the constant M in \(\Omega \). Then \(B_\lambda :\Omega \rightarrow \Omega \) is a contractive mapping for all \(\lambda \in (0,2\delta /M^2)\).
Proof.
Since \(\psi (\alpha )\) is strongly monotone with the constant \(\delta \) by Lemma 2,
By (42), (43) and the Lipschitz condition, we have
which implies that
for all \(\lambda \in (0,2\delta /M^2)\), where \(q=\sqrt{1-2\lambda \delta +\lambda ^2M^2}< 1 \). \(\square \)
Theorem 2.
Assume that the problem (1) has an optimal solution. Suppose that \(\psi (\alpha )\) is differentiable, and satisfies (40) and the Lipschitz condition in \(\Omega \). The equation \(\alpha =B_\lambda (\alpha )\) for all \(\lambda \in (0,2\delta /M^2)\) is equivalent to the equation \(\psi (\alpha )=0\) and the equation (22) has a unique solution.
Proof.
By Lemma 5 and the contractive mapping principle, \(B_\lambda (\alpha )\) has only one fixed point \(\alpha ^*\) in \(\Omega \) for all \(\lambda \in (0,2\delta /M^2)\), i.e. \(B_\lambda (\alpha ^*)=\alpha ^*\). From (41), we have
for \(\alpha \in \Omega ,\) i.e. for every \(\alpha \in \)
Since the problem (1) has an optimal solution, there is a \({\bar{\alpha }}\in \Omega \) such that \(\psi ({\bar{\alpha }})=0\) by Lemma 3. Then, it follows from (44) that
By Lemma 2, \(\psi (\alpha )\) is strongly monotone with the constant \(\delta \) and so we have
This inequality and (45) together implies \(0\le [\psi (\alpha ^*)-\psi ({\bar{\alpha }})]^T({\bar{\alpha }}-\alpha ^*)\le -\delta \Vert \alpha ^*-{\bar{\alpha }}\Vert ^2\). Therefore, \(\Vert \alpha ^*-{\bar{\alpha }}\Vert =0\), i.e. \(\alpha ^*={\bar{\alpha }}\) . Thus , \(\psi (\alpha ^*)=0\) , which means that \(\alpha ^*\) is a solution of (22).
By the definition of \(B_\lambda (\alpha )\), it is obvious that \(\alpha \in \Omega \) with \(\psi (\alpha )=0\) is also the solution of \(\alpha =B_\lambda (\alpha ).\)
Suppose that \(\alpha '\in \Omega \) is a solution of (22) that is different from \(\alpha ^*\). Then, \(\alpha '=B_\lambda (\alpha ')\), which is contradiction because \(\alpha ^*\) is an unique fixed point of \(B_\lambda (\alpha )\) in \(\Omega \). Thus, \(\alpha ^*\) is a unique solution of the equation (22). \(\square \)
Corollary 3.
Suppose that the assumptions of Theorem 2 are satisfied. Then, the equation \(\alpha =A(\alpha )\) has an unique solution which is the unique solution of the equation (22).
Proof.
By Lemma 3, \(\alpha =A(\alpha )\) is equivalent to \(\psi (\alpha )=0\) which is in turn equivalent to \(\alpha =B_\lambda (\alpha )\) by Theorem 2. Therefore, it follows from Theorem 2 that equation \(\alpha =A(\alpha )\) has a unique solution which is the unique solution of the equation (22). \(\square \)
As shown in Lemma 3, \(\alpha \) is a fixed point of \(A(\alpha )\) if and only if \(\alpha \) is a root of \(\psi (\alpha )\) . We can construct the following fixed-point iteration to find a fixed point of \(A(\alpha )\) .
where \(x^{k}=x(\alpha ^{k})\).
Corollary 4.
Suppose that the assumptions of Theorem 2 are satisfied. Then the algorithm (46) is just the Newton method to solve the equation (22) and its local rate of convergence is superlinear or quadratic.
Proof.
The Newton method for the equation (22) is as following.
By (40), (17) and (18), the right-hand side of (47) is equal to \(A(\alpha ^k)\), i.e. the right-hand side of (46). Therefore, (47) means \(\alpha ^{k+1}=A(\alpha ^k)\) , that is, the fixed-point iteration method to find a fixed point of \(A(\alpha )\) is just the Newton method for solving the equation (22). Hence, the algorithm (46) has local superlinear or quadratic convergence rate. \(\square \)
Theorem 3.
Suppose that the assumptions of Theorem 2 are satisfied and that there exists \(L>0\) such that for every \(\alpha , \alpha '\in \Omega \)
and that there exists \({\tilde{M}}>0\) such that for every \(\alpha \in \Omega \)
Let
where \(\lambda _k\) is the greatest \( \xi ^i \) satisfying
and \(i\in \{0,1,2,\cdots \},\quad \xi \in (0,1),\quad \varepsilon \in (0,1).\) Then, the modified Newton method defined by (50) and (51) converges to the unique solution \(\alpha ^*\) of \(\psi (\alpha )\) with linear rate for any starting point \(\alpha ^0\in \Omega \) and the rate in the neighborhood of the solution is quadratic.
Proof.
We have already shown the existence and uniqueness of the solution \(\alpha \in \Omega \) for the equation \(\psi (\alpha )=0\). If there is k such that \(\psi (\alpha ^k)=0\), then \(\alpha ^k\) is a solution. So, it is assumed that \(\psi (\alpha ^k)\ne 0\) for every k . For \(\lambda \in [0,1]\) , we have the following by the Newton-Leibniz formula and (48).
where
by the Lipschitz condition, and \(\psi '(\alpha ^k)d^k=-\psi (\alpha ^k)\) by (50) . In view of (49), it follows from (50) and (52) that
Letting \({\bar{\lambda }}_k =\frac{1-\varepsilon }{L{\tilde{M}}^2\Vert \psi (\alpha ^k)\Vert }\), we have
for every \(\lambda \in (0,\min \{1,\bar{\lambda }_{k}\})\) . Then, (51), the definition of \(\lambda _k\), implies that
Since it follows from (53) that
\(\{\Vert \psi (\alpha ^k)\Vert \}\) is a monotonically decreasing sequence and so \(\{\bar{\lambda }_{k}\}\) increases monotonically. From (54), it follows that
Letting \(q=1-\varepsilon \min \{1,\xi \bar{\lambda }_{0}\}\) , then \(q<1\) and by (55) we have
Therefore, \(\{ \Vert \psi (\alpha ^{k})\Vert \}\) converges to zero with linear rate, which means that \(\{\alpha ^k\}\) converges to a solution of \(\psi (\alpha )=0\) for any starting point \(\alpha ^0\in \Omega \). Since \( \Vert \psi (\alpha ^{k})\Vert \rightarrow 0\) as \(k\rightarrow \infty \), we have \({\bar{\lambda }}_k\rightarrow \infty \) as \(k\rightarrow \infty \) and so, by (54), there exists \(k_0\) such that \(\lambda _k=1\) for every \(k\ge k_0\). Thus, in this case, (50) becomes the Newton method and the rate of convergence in the neighborhood of the solution is quadratic by the Lipschitz property (48). \(\square \)
Let us consider sufficient condition under which the assumptions of Theorem 3 are satisfied.
Theorem 4.
Suppose that \(\psi (\alpha )\) is differentiable and satisfies (40) in \(\Omega \). If \(h_i(x), i=1,\cdots ,p\) and \(x(\alpha )\) are Lipschitz continuous in the feasible set X and \(\Omega \), respectively, then \(\psi '(\alpha )\) is Lipschitz continuous and \([\psi '(\alpha )]^{-1}\) is bounded in \(\Omega \).
Proof.
By the Lipschitz continuity of \(h_i(x)\) and \(x(\alpha )\), there is a \(L_i>0\) and \(C>0\) such that, for all \(\alpha , \alpha '\in \Omega \),
Since, by (40), we have
there exists a constant \(L>0\) by (57) and (58) such that \(|\psi '(\alpha )-\psi '(\alpha ') |\le L |\alpha -\alpha ' |\) , which means the Lipschitz continuity of \(\psi '(\alpha )\) .
It is easy to see
Thus, from \(\frac{1}{h_i(x(\alpha ))}\le \gamma _i^u,\quad i=1,\cdots ,p\) and Remark 1, it follows that there exists a \({\tilde{M}}>0\) such that for every \(\alpha \in \Omega ,\quad |[\psi '(\alpha )]^{-1} |\le {\tilde{M}},\) that is, (49) is satisfied. \(\square \)
Under the assumption of Theorem 4, the modified Newton method defined by
can be rewritten component-wise as following.
On the basis of the above consideration, we construct an algorithm to find global solution of the problem (1) under the assumptions of Theorem 4 as following.
[Algorithm MN] Step 0. Choose \(\xi \in (0,1),\quad \varepsilon \in (0,1)\) and \(y^0\in X\). Let
Step 1. Find a solution \(x^k=x(\alpha ^k)\) of the problem (3) for \(\alpha =\alpha ^k.\)
Step 2. If \(\psi (\alpha ^k)=0\) , then \(x^k\) is a global solution and so stop the algorithm. Otherwise, let \(i_k\) denote the smallest integer among \(i\in \{0,1,2,\cdots \}\) satisfying
and let \(\lambda _k=\xi ^{i_k}\), where
Compute \( \alpha ^{k+1}=\alpha ^k+\lambda _k d^k\) which is equal to (59).
Step 3. Let \(k=k+1\) and go to step 1. \(\square \)
Remark 5.
Under the assumptions of Theorem 4, the step 2 of Algorithm MN is just the modified Newton method (50), (51). If the stepsize \(\lambda _k=\xi ^{i_k}\) in the step 2 is replaced by \(\lambda _k=1\), the step 2 is just the Newton method (47) or (46), and so we denote Algorithm MN as Algorithm N in this case. As shown in Theorem 3, Algorithm MN has global linear and local quadratic rate of convergence.
Remark 6.
If the assumption (H1) is satisfied, we can construct a Newton-like algorithm similar to Algorithm MN without the assumption (H2). In this case, the Newton direction in the step 2 is determined by \(d^k=-(\psi '(\alpha ^k))^{-1}\psi (\alpha ^k))\) using the formula (26), (27) and (37).
Even without differentiability of \(\psi (\alpha )\), we obtain the following result.
Theorem 5.
Suppose that \(\lambda _k\equiv 1\) (or \( \lambda _k\ge {\bar{\lambda }}>0\)) for every k. Then any accumulation point \(\alpha ^*\) of \(\{\alpha ^k\}\) generated by The algorithm N, i.e. (46) (or The algorithm MN, i.e. (59)) is the solution of the equation (22). Moreover, if the equation (22) has a unique solution, the \(x(\alpha ^*)\) is the solution of the problem (1).
Proof.
As shown in Remark 1, the parameter set \(\Omega \) is bounded. Therefore, the sequence \(\{\alpha ^k\}\) is bounded and it has a convergent subsequence by Bolzano-Weierstrass theorem. Then any accumulation point \(\alpha ^*\) of the subsequence is the solution of the equation (22) by (46) or (59). The Corollary 2 completes the proof. \(\square \)
4 Numerical experiments
Our algorithms N and MN were coded in MATLAB and numerical experiments had been carried out on a Pentium (R) 4 CPU 2 GHz RAM 0.97 GB personal computer using the optimization toolbox of MATLAB R2009a. We set \(\Vert \psi (\alpha ^k)\Vert <1e-6\) as the stopping criterion in numerical experiments. In the tables below, “ iter” is the total number of the convex programming problems solved during the algorithm’s work, and the row “N” and “MN” show the average performance of 100 runs with random initial points in our algorithm N and MN.
Example 1
Table 1 shows the computation result, which demonstrates the advantage of our algorithm over the previous algorithms based on the branch-and-bound method.
Example 2
The optimal solution and optimal value is (1, 2) and 4.0357143 , respectively. The results of our methods and previous methods [11, 19, 27] are shown in Table 2. This shows that our methods are much better than the previous methods.
Example 3
The test results are shown in the following table.
Example 4
The results of numerical experiments are given in Table 4, which shows the advantage of our methods over the previous methods.
Example 5
The numerical experiment results are shown in Table 5. It shows that the method MN is much better than previous methods. The algorithm N failed to find any global solution.
Example 6
([ 25 ]).
The numerical experiment results are shown in Table 6, which shows the advantage of our algorithms over the previous methods.
Example 7
The numerical experiment result is shown in Table 7.
Example 8
The numerical experiment results are shown in Table 8. Our results are better than those of previous methods.
Example 9
where \(a_1=0.25\) and \(a_2=-1.75.\) The numerical experiment results are shown in Table 9.
Example 10
The numerical experiment results are shown in Table 10.
Remark 7.
Example 5,6,8 and 9 do not satisfy the assumption of the problem (1), which is, that numerators are convex and denominators are concave in all ratios.In addition, Example 5,8 and 10 have nonconvex constraints. The solutions in [11] and [20] are nonfeasible. For those problems,the algorithm MN obtained much more improved optimal solutions than the existing methods based on branch-and-bound algorithm. Our algorithm MN always found global solution for any starting point in all the examples 1-10 and the algorithm N also gave the same good results for all the examples but for the example 5 which does not satisfy the assumption about objective and constraint functions. This shows that our algorithms can find more accurate global solutions for more general nonlinear sum of ratios problems.
Example 11.
We carried out numerical experiments for randomly generated minimization problems as following. The numerator and denominator of each term in objective function is
respectively, where
The feasible set is given by
where
Starting with randomly generated point in \([1,5]^n\), we carried out 50 runs of Algorithm N and Algorithm MN for fixed n and p. Both of algorithms N and MN were successful for all the problems. It is shown in Table 11 and Table 12.
As shown in the above tables, the number of iteration of Algorithm N and Algorithm MN was independent of the number of variables and fractional functions, and the run-time was proportional to the number of variables.
Remark 8.
It is worth mentioning that both of \(\Vert \psi (\alpha )\Vert \) and the objective function decreased monotonically at each iteration of our algorithms N and MN in all numerical experiments for Examples 1-11.
5 Discussions
In this paper, we presented a parametric optimization algorithm for the sum-of-ratios problem, which is a non-convex fractional programming problem that occurs in many fields, including computer graphics and management. Our approach is based on transforming the sum-of-ratios problem into parametric convex programming problem and it applied Newton-like algorithm to update parameters. The algorithm has global linear and local superlinear/quadratic rate of convergence. Through numerical experiments, we demonstrated that our method always finds the global optimum with a fast convergence rate. In fact, the experiments showed that the optimal point is usually obtained within the first few iterations and it also showed that the algorithm can find global solution for larger class of sum-of-ratios problem than those considered in this paper. Although there is no theoretical guarantee that our method can always find a global optimal solution, we expect that our approach will solve successfully several linear and nonlinear sum-of-ratios problems in practice.
References
M.S. Bazara, H.D. Sherali, C.M. Shetty, Nonlinear Programming: Theory and Algorithms, John Wiley and Sons, New York, 2006.
H.P. Benson, Branch-and-Bound Outer Approximation Algorithm for Sum-of-Ratios Fractional Programs, J.Optim. Theor. Appl. 146(2010)1-18.
H.P. Benson, Solving Sum of Ratios Fractional Programs via Concave Minimization, J.Optim. Theor. Appl. 135(2007)1-17.
H.P. Benson, Using concave envelopes to globally solve the nonlinear sum-of-ratios problem, J. Global Optim. 22(2002)343-364.
H.P. Benson, Global Optimization Algorithm for the Nonlinear Sum of Ratios Problem, J.Optim. Theor. Appl. 112(2002)1-29.
P. C. P. Carvalho, L. H. Figueiredo, J. Gomes, L. Velho, Mathematical Optimization in Computer Graphics and Vision, Elsevier, Morgan Kaufmann Publishers, 2008.
A. V. Fiacco, Sensitivity analysis for nonlinear programming using penalty methods, Math. Program. 10(1976)287-311.
R. W. Freund, F. Jarre, Solving the sum-of-ratios problem by an interior-point method, J. Global Optim. 19(2001)83-102.
Y. Ji, Y. Li, P. Lu, A global optimization algorithm for sum of quadratic ratios problem with coefficients, Appl. Math. Comput. 218(2012)9965-9973.
H.W. Jiao, A branch and bound algorithm for globally solving a class of nonconvex programming problems, Nonlinear Analysis 70(2009)1113-1123.
H.W. Jiao, Z.K. Wang, Y.Q. Chen, Global optimization algorithm for sum of generalized polynomial ratios problem, Appl. Math. Model. (2012)https://doi.org/10.1016/j.apm.2012.02.023
F. Kahl, S. Agarwal, M. K. Chandraker, D. Kriegman, S. Belongie, Practical Global Optimization for Multiview Geometry, Int.J.Comput.Vis. 79(2008)271-284.
D. Kinderlehrer, G. Stampcchia, An introduction to variational inequalities and their applications. Italiana: Academic Press, 1980
T. Kuno, A branch-and-bound algorithm for maximizing the sum of several linear ratios, J. Global Optim. 22(2002)155-174.
T. Munson, Mesh shape-quality optimization using the inverse mean-ratio metric, Math. Program. 110(2007)561-590.
S.J. Qu, K. C. Zhang, J. K. Zhao, An efficient algorithm for globally minimizing sum of quadratic ratios problem with nonconvex quadratic constraints, Appl. Math. Comput. 189(2007)1624-1636.
S. Schaible, J. Shi, Fractional programming: the sum-of- ratios case, Opt. Meth. Soft. 18(2003)219-229.
P. Shen, Y. Ma, Y. Chen, Global optimization for the generalized polynomial sum of ratios problem, J. Global Optim. 50(2011)439-455.
P. Shen, L. Jin, Using conical partition to globally maximizing the nonlinear sum of ratios, Appl. Math. Model. 34 (9)(2010)2396-2413.
P. Shen, Y. Duan, Y. Pei, A simplicial branch and duality bound algorithm for the sum of convex-convex ratios problem, J. Comput. Appl. Math. 223 (1)(2009)145-158.
P. Shen, C. Wang, Global optimization for sum of linear ratios problem with coefficients, Appl. Math. Comput. 176(2006)219-229.
Y. Wang, P. Shen, Z. Liang, A branch-and-bound algorithm to globally solve the sum of several linear ratios, Appl. Math. Comput. 168(2005)89-101.
P. Shen, G. Yuan, Global optimization for the sum of generalized polynomial fractional functions, Math. Methods Oper. Res. 65(2007)445-459.
M. Tawarmalani, N. V. Sahinidis, Semidefinite relaxations of fractional programs via novel convexification techniques, J. Global Optim. 20(2001)137-158.
C. Wang, S. Liu, P. Shen, Global optimization for sum of geometric fractional functions, Appl. Math. Comput. 216(2010)2263-2270.
W. Wu, R. Sheu, S. I. Birbil, Solving the sum-of-ratios problem by a stochastic search algorithm, J. Global Optim. 42(2008)91-109.
D. Yang, J. Shi, S. Wang, Conical partition algorithm for maximizing the sum of dc ratios, J. Global Optim. 31(2005)253–270.
Acknowledgements
The authors are grateful to the anonymous referees for their helpful comments and suggestions, which have greatly improved the earlier version of this paper.
Author information
Authors and Affiliations
Corresponding author
Additional information
Communicated by T S S R K Rao, Phd.
This work was supported by the National Natural Science Foundation under the fourth Five Years Program for Development of Science and Technology of DPR Korea.
Rights and permissions
About this article
Cite this article
Kim, Y., Jong, Y. & Yu, J. A parametric solution method for a generalized fractional programming problem. Indian J Pure Appl Math 52, 971–989 (2021). https://doi.org/10.1007/s13226-021-00102-y
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13226-021-00102-y