Abstract
In this article, we develop a quasi-Newton method to obtain nondominated solutions of fuzzy optimization problems. The objective function of the optimization problem under consideration is a fuzzy-number-valued function. The notion of generalized Hukuhara difference of fuzzy numbers, and hence generalized Hukuhara differentiability for multi-variable fuzzy-number-valued functions are used to develop the quasi-Newton method. The proposed technique produces a sequence of positive definite inverse Hessian approximations to generate the iterative points. A sequential algorithm and the convergence result of the proposed method are also given. It is obtained that the sequence in the proposed method has superlinear convergence rate. The method is also found to have quadratic termination property. Two numerical examples are provided to illustrate the developed technique.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
After the seminal work by Bellman and Zadeh [7], in 1970, fuzzy optimization problems (FOPs) are extensively studied to tackle the decision-making problems that are emanated from the imprecise environment. Recently, Luhandjula [28] projected a condensed and selective look at the known landscape of the theory and applications of fuzzy optimization. The edited monograph by Słowínski [36] and the article by Cadenas and Verdegay [8] and the references therein are the main stream of this topic. An insightful overview on the development of FOPs has been reported in the survey books by Lai and Hwang [24, 25] and by Lodwick and Kacprzyk [27].
Wu [41,42,43], Zhong [52] and Cheng [11] have studied duality theory for FOPs. Gong and Li [23] and Wu [48] have found the saddle point optimality conditions for FOPs. Recently, Ghosh and Chakraborty reported fuzzy geometrical view [10, 16, 17] on FOPs [18,19,20,21]. In order to solve an unconstrained FOP, Pirzada and Pathak [32] and Chalco-Cano et al. [9] developed a Newton method. Much similar to fuzzy optimization, Ghosh [14] derived a Newton method and quasi-Newton method [15] for interval optimization problems.
The practical situations of mathematical optimization models often encounter the problem of optimizing a fuzzy function over real data set. Precisely, this problem can be stated as ‘\(\displaystyle \min \nolimits _{x \in \mathbb {R}^n} \widetilde{f}(x)\)’, where \(\widetilde{f}\) is a fuzzy-number-valued function. In classical optimization theory, there are various numerical algorithms [1, 30, 51] for unconstrained optimization problems. However, those methods are evidently not applicable to solve FOPs. In this article, we attempt to develop a quasi-Newton method to solve an unconstrained FOP.
In the context of conventional optimization problems, towards searching the optimal points [4, 30] of a smooth optimization problem, it is natural to employ the derivative of the objective function. However, to find the optimal points of an FOP, not only the differentiability of fuzzy functions but also an appropriate ordering of fuzzy numbers has to be considered to find the optimal points, since unlike to the real numbers, fuzzy numbers are not linearly ordered [35].
There have been a wide literature on ordering of fuzzy numbers including the research articles referred in [12, 26, 29, 34, 39, 49]. Cheng [12] proposed a ranking of fuzzy numbers with the help of a distance method. Sun and Wu [38] presented a ranking based on fuzzy simulation analysis method. Peng et al. [31] studied ranking fuzzy variables in terms of credibility measure. Yao and Wu [50] defined ranking fuzzy numbers based on decomposition principle and signed distance. Wang [40] proposed a ranking of L-R fuzzy numbers based on a deviation degree. References of [40] reports the main stream on ordering of fuzzy numbers.
In this paper, the fuzzy-max ordering of fuzzy numbers of Ramík and Rimanek [34] will be used. There are two reasons behind this selection. First, it is a partial ordering in the space of fuzzy numbers [32]. Second, it has insightful association [44, 45] with the optimality notion on FOPs.
In order to find the notion of differentiability of fuzzy functions, Dubois and Prade [13] used the extension principle. Dubois an Prade [13] have also reported that a fuzzy extension of the usual definition of differentiation is of no use. As a result of this fact, Goetschel and Voxman [22] used a so-called ‘opposite’ of a fuzzy number to define fuzzy differentiation in analogy to the usual definition of differentiation. Recently, Hukuhara differentiability (H-differentiability) [3] received substantial attention in fuzzy optimization theory. Banks and Jacobs [3] developed H-differentiability for multifunctions. Puri and Ralescu [33] further extended the H-differentiability for fuzzy functions. The concept of H-fuzzy differentiation is rigorously discussed in [2] for the univariate and multivariate fuzzy functions. The H-differentiation is also studied by several authors in different fields. For instance, Wu et al. [47] applied the H-differentiability in fuzzy differential equations. In fuzzy optimization, Wu employed the H-derivative in order to find optimality conditions for multi-objective [46] and single objective [44, 45] fuzzy-valued functions. Bede and Gal [5, 37] proposed generalizations of H-differentiability (gH-differentiability) and its application in fuzzy differential equations. Bede and Stefanini [6] gave a generalized H-differentiability of fuzzy-valued functions. Pirzada and Pathak [32] derived a Newton method for FOP using H-derivative. Chalco-Cano et al. [9] used gH-differentiability to modify the Newton method in [32].
Since gH-derivative is the most general concept [9] for differentiability of fuzzy functions, in this paper, we employ the gH-derivative and its calculus [9, 44, 45] to derive a quasi-Newton method for solving an FOP.
It is to note that in the derivation of Newton method for FOPs [9, 32], the generating equation for iterative points involves computation of inverse of the concerned Hessian matrix. Computation of inverse-Hessian is truly cost effective. Thus, in this study, we attempt to propose a quasi-Newton method to sidestep the inherent computational expense in computing inverse-Hessian in the existing Newton method for FOPs. This sidestepping is done through a rank-two modification to an appropriate approximation of the pertaining inverse-Hessian.
The work in this study is organized as follows. Immediately next of this paragraph, list of symbols those are used throughout the article is given. In Sect. 2, the notations and terminologies are given which are used throughout the paper. In Sect. 3, we derive a quasi-Newton method for FOPs. Convergence analysis of the proposed method is presented in Sect. 4. Section 5 includes two illustrative numerical examples. Finally, we give a brief conclusions and scopes for future research in Sect. 6.
2 Preliminaries
This section provides basic definitions and notations which are employed in this article. Upper and lower case letters with a tilde (\(\widetilde{A}\), \(\widetilde{B}\), \(\widetilde{C}\), ... and \(\widetilde{a}\), \(\widetilde{b}\), \(\widetilde{c}\), ...) denote fuzzy subsets of \(\mathbb {R}\). The membership function of a fuzzy set \(\widetilde{A}\) of \(\mathbb {R}\) is represented by \(\mu (x|\widetilde{A})\), for x in \(\mathbb {R}\), with \(\mu (\mathbb {R}) \subseteq [0, 1]\).
2.1 Fuzzy numbers
Definition 1
(\(\alpha \)-cut of a fuzzy set [16]) The \(\alpha \)-cut of a fuzzy set \(\widetilde{A}\) of \(\mathbb {R}\), denoted \(\widetilde{A}(\alpha )\), is defined by
The sets \(\{x: \mu (x|\widetilde{A})> 0\}\) and \(\{x: \mu (x|\widetilde{A}) = 1\}\) are respectively called the support and the core of the fuzzy set \(\widetilde{A}\).
Definition 2
(Fuzzy number [16]) A fuzzy set \(\widetilde{N}\) of \(\mathbb {R}\) is called a fuzzy number if its membership function \(\mu \) has the following properties:
-
(i)
\(\mu (x|\widetilde{N})\) is upper semi-continuous,
-
(ii)
\(\mu (x|\widetilde{N}) = 0\) outside some interval [a, d], and
-
(iii)
there exist real numbers b and c such that \(a\le b\le c\le d\) and \(\mu (x|\widetilde{N})\) is increasing on [a, b] and decreasing on [c, d], and \(\mu (x|\widetilde{N}) = 1\) for each x in [b, c].
In particular, when \(b = c\), and the parts of the membership functions \(\mu (x|\widetilde{N})\) in [a, b] and [c, d] are linear, then the fuzzy number is called a triangular fuzzy number and is denoted by (a / c / d). We denote the set of all fuzzy numbers on \(\mathbb {R}\) by \(\mathcal {F}(\mathbb {R})\).
Since \(\mu (x|\widetilde{a})\) is upper semi-continuous for a fuzzy number \(\widetilde{a}\), the set \(\{x: \mu (x | \widetilde{a})\ge \alpha \}\) is closed for each \(\alpha \) in \(\mathbb {R}\). Thus, the \(\alpha \)-cut of a fuzzy number \(\widetilde{a}\) (the set \(\widetilde{a}(\alpha )\)) is a closed and bounded interval of \(\mathbb {R}\) for all \(\alpha \) in [0, 1]. We write \(\widetilde{a}(\alpha )\) \(=\left[ \widetilde{a}_\alpha ^L, \widetilde{a}_\alpha ^U\right] \).
Let \(\oplus \) and \(\odot \) denote the extended addition and multiplication. According to the extension principle [53], the membership function of \(\widetilde{a} \otimes \widetilde{b}\) (\(\otimes = \oplus \) or \(\odot \)) is defined by
For any \(\widetilde{a}\) and \(\widetilde{b}\) in \(\mathcal {F}(\mathbb {R})\), the \(\alpha \)-cut (for any \(\alpha \) in [0, 1]) of their addition and scalar multiplication can be obtained as [53]:
Definition 3
(Generalized Hukuhara difference [37]) Let \(\widetilde{a}\) and \(\widetilde{b}\) be two fuzzy numbers. If there exists a fuzzy number \(\widetilde{c}\) such that \(\widetilde{c}\) \(\oplus \) \(\widetilde{b} = \widetilde{a}\) or \(\widetilde{b} = \widetilde{a} \ominus \widetilde{c}\), then \(\widetilde{c}\) is said to be generalized Hukuhara difference (gH-difference for short) between \(\widetilde{a}\) and \(\widetilde{b}\). If such \(\widetilde{c}\) exists, then it is unique. Hukuhara difference between \(\widetilde{a}\) and \(\widetilde{b}\) is denoted by \(\widetilde{a} \ominus _{gH} \widetilde{b}\).
In terms of \(\alpha \)-cut, for all \(\alpha \in [0, 1]\), we have
2.2 Fuzzy functions
Let \(\widetilde{f}: \mathbb {R}^n \rightarrow \mathcal {F}(\mathbb {R})\) be a fuzzy function. We present the \(\alpha \)-cuts of this fuzzy function by
Here the functions \(\widetilde{f}_\alpha ^L\) and \(\widetilde{f}_\alpha ^U\) are two real valued functions and are called lower and upper functions, respectively.
With the help of gH-difference between two fuzzy numbers, gH-differentiability of a fuzzy function is defined as follows.
Definition 4
(gH-differentiability of fuzzy functions [5]) Let \(\widetilde{f}: \mathbb {R}^n \rightarrow \mathcal {F}(\mathbb {R})\) be a fuzzy function and \(x_0 = \left( x_1^0, x_2^0, \ldots , x_n^0\right) \) be an element of \(\mathbb {R}^n\). We define a fuzzy function \(\widetilde{h}_i: \mathbb {R} \rightarrow \mathcal {F}(\mathbb {R})\) for each \(i = 1, 2, \ldots , n\) as follows. Consider
We say \(\widetilde{h}_i\) is gH-differentiable if the following limit exists
If \(\widetilde{h}_i\) is gH-differentiable, then we say that \(\widetilde{f}\) has the ith partial gH-derivative at \(x_0\) and is denoted by \(\frac{\partial \widetilde{f}}{\partial x_i}(x_0)\).
The function \(\widetilde{f}\) is said to be gH-differentiable at \(x_0 \in \mathbb {R}^n\) if all the partial gH-derivatives \(\frac{\partial \widetilde{f}}{\partial x_1}(x_0)\), \(\frac{\partial \widetilde{f}}{\partial x_2}(x_0)\), \(\ldots \), \(\frac{\partial \widetilde{f}}{\partial x_n}(x_0)\) exists on some neighborhood of \(x_0\) and are continuous at \(x_0\).
Proposition 1
(See [9]) If a fuzzy function \(\widetilde{f}: \mathbb {R}^n \rightarrow \mathcal {F}(\mathbb {R})\) is gH-differentiable at \(x_0 \in \mathbb {R}^n\), then for each \(\alpha \in [0, 1]\), the real valued function \(\widetilde{f}^L_\alpha + \widetilde{f}^U_\alpha \) is differentiable at \(x_0\). Moreover,
Definition 5
(gH-gradient [9]) The gH-gradient of a fuzzy function \(\widetilde{f}: \mathbb {R}^n \rightarrow \mathcal {F}(\mathbb {R})\) at a point \(x_0 \in \mathbb {R}^n\) is defined by
We denote this gH-gradient by \(\nabla \widetilde{f} (x_0)\).
We define an m-times continuously gH-differentiable fuzzy function \(\widetilde{f}\) as a function whose all of the partial gH-derivatives of order m exist and are continuous. Then, we have the following immediate result.
Proposition 2
(See [9]) Let \(\widetilde{f}\) be a fuzzy function. Let at \(x_0 \in \mathbb {R}^n\), the function \(\widetilde{f} \) be m-times gH-differentiable. Then the real-valued function \(\widetilde{f}^L_\alpha + \widetilde{f}^U_\alpha \) is m-times differentiable at \(x_0\).
Definition 6
(gH-Hessian [9]) Let the fuzzy function \(\widetilde{f}\) be twice gH-differentiable at \(x_0\). Then, for each i, the function \(\frac{\partial \widetilde{f}}{\partial x_i}\) is gH-differentiable at \(x_0\). The second order partial gH-derivative can be calculated through computing all the \(\frac{\partial ^2 \widetilde{f}}{\partial x_i \partial x_j}\)’s. The gH-Hessian of \(\widetilde{f}\) at \(x_0\) can be captured by the square matrix
2.3 Optimality concept
As it is mentioned in Sect. 1 that in order to search the optimal solution of an FOP, we need a partial order relation between fuzzy numbers. Recently, fuzzy-max ordering of fuzzy numbers are extensively used [32, 44, 45] in fuzzy optimization, since it is a partial order relation in \(\mathcal {F}(\mathbb {R})\) [32]. Fuzzy-max ordering is defined as follows.
Definition 7
(Dominance relation between fuzzy numbers [32]) Let \(\widetilde{a}\) and \(\widetilde{b}\) be two fuzzy numbers. For any \(\alpha \in [0, 1]\), let \(\widetilde{a}(\alpha ) = \left[ \widetilde{a}^L_\alpha , \widetilde{a}^U_\alpha \right] \) and \(\widetilde{b}(\alpha ) = \left[ \widetilde{b}^L_\alpha , \widetilde{b}^U_\alpha \right] \). We say \(\widetilde{a}\) dominates \(\widetilde{b}\) if \(\widetilde{a}^L_\alpha \le \widetilde{b}^L_\alpha \) and \(\widetilde{a}^U_\alpha \le \widetilde{b}^U_\alpha \) for all \(\alpha \in [0, 1]\). If \(\widetilde{a}\) dominates \(\widetilde{b}\), then we write \(\widetilde{a} \preceq \widetilde{b}\). The fuzzy number \(\widetilde{b}\) is said to be strictly dominated by \(\widetilde{a}\), if \(\widetilde{a} \preceq \widetilde{b}\) and there exists \(\beta \in [0, 1]\) such that \(\widetilde{a}^L_\beta < \widetilde{b}^L_\beta \) or \(\widetilde{a}^U_\beta < \widetilde{b}^U_\beta \). If \(\widetilde{a}\) strictly dominates \(\widetilde{b}\), then we write \(\widetilde{a} \prec \widetilde{b}\).
With the help of Definition 7, the concept of non-dominated solution for FOPs is defined as follows.
Definition 8
(Non-dominated solution for fuzzy optimization [32]) Let \(\widetilde{f}: \mathbb {R}^n \rightarrow \mathcal {F}(\mathbb {R})\) be a fuzzy function and we intend to find a solution of ‘\(\displaystyle \min \nolimits _{x \in \mathbb {R}^n} \widetilde{f}(x)\)’. A point \(\bar{x} \in \mathbb {R}^n\) is said to be a locally non-dominated solution if there exists no \(x \in N_{\epsilon }(\bar{x})\) such that \(\widetilde{f}(x) \preceq \widetilde{f}(\bar{x})\), where \(N_{\epsilon }(\bar{x})\) denotes \(\epsilon \)-neighborhood of \(\bar{x}\). A local non-dominated solution is called a local solution of ‘\(\displaystyle \min \nolimits _{x \in \mathbb {R}^n} \widetilde{f}(x)\)’.
For local non-dominated solution, the following result is proved in [9].
Proposition 3
(See [9]) Let \(\widetilde{f}: \mathbb {R}^n \rightarrow \mathcal {F}(\mathbb {R})\) be a fuzzy function. If \(x^*\) is a local minimizer of the real valued \(\widetilde{f}^L_\alpha + \widetilde{f}^U_\alpha \) for all \(\alpha \in [0, 1]\), then \(x^*\) is a locally non-dominated solution of ‘\(\displaystyle \min \nolimits _{x \in \mathbb {R}^n} \widetilde{f}(x)\)’.
3 Quasi-Newton method
In this section, we consider to solve the following unconstrained Fuzzy Optimization Problem:
where \(\widetilde{f}: \mathbb {R}^n \rightarrow \mathcal {F}(\mathbb {R})\) is a fuzzy-number-valued function. On finding a nondominated solution of the problem, we note that the existing Newton method (see [9, 32]) requires computation of the inverse of the concerned Hessian. However, computation of the inverse of Hessian is cost effective. Thus in this article, we intend to develop a quasi-Newton method to sidestep the computational difficulty of the Newton method. Towards this end, for the FOP we assume that at each of the following generated sequential points \(x_k\), the function \(\widetilde{f}\) and its gH-gradient and gH-Hessian are well-defined. Therefore, according to Propositions 1 and 2, we can calculate \(\widetilde{f}^L_\alpha (x_k)\), \(\widetilde{f}^U_\alpha (x_k)\), \(\nabla \widetilde{f}^L_\alpha (x_k)\), \(\nabla \widetilde{f}^U_\alpha (x_k)\), \(\nabla ^2 \widetilde{f}^L_\alpha (x_k)\) and \(\nabla ^2 \widetilde{f}^U_\alpha (x_k)\) for all \(\alpha \in [0, 1]\), for all \(k = 0, 1, 2, \cdots \). Hence, we can have a quadratic approximations of the the lower and the upper functions \(\widetilde{f}^L_\alpha \) and \(\widetilde{f}^U_\alpha \).
Let the quadratic approximation models of \(\widetilde{f}^L_\alpha \) and \(\widetilde{f}^U_\alpha \) at \(x_{k+1}\) be
and
which satisfy the interpolating conditions
The derivatives of \(h_\alpha ^L\) and \(h_\alpha ^U\) yield
In the next, the Newton method (see [9, 32]) attempts to find x in terms of \(x_{k+1}\) as follows
where \(\phi (x) = \int _0^1 \left( f_\alpha ^L(x) + f_\alpha ^U(x)\right) \mathrm {d}\alpha \).
However, due to inherent computational difficulty to find \([\nabla ^2 \phi (x_{k+1})]^{-1}\), it is often suggested that consider an appropriate approximation.
Let \(A_{k+1}\) be an approximation of \([\nabla ^2 \phi (x_{k+1})]^{-1}\). Then, from (3.1), setting \(x = x_k\), \(\delta _k = x_{k+1} - x_k\), \(\beta _{\alpha k}^L = \nabla f_\alpha ^L(x_{k+1}) - \nabla f_\alpha ^L(x_{k})\) and \(\beta _{\alpha k}^U = \nabla f_\alpha ^U(x_{k+1}) - \nabla f_\alpha ^U(x_{k})\), we obtain
According to Proposition 3, to obtain nondominated solutions of (FOP), we need to have solutions of the equation \(\nabla (f_\alpha ^L + f_\alpha ^U)(x) = 0\).
In order to capture solutions of this equation, much similar to the Newton method [32], the above procedure clearly suggests to consider the following generating sequence
If \(A_k\) is an appropriate approximation of the inverse Hessian \(\left[ \nabla ^2 \phi (x_{k+1})\right] ^{-1}\) and \(\nabla (f_\alpha ^L + f_\alpha ^U)(x_{k+1}) \thickapprox 0\), then the equation \(x_{k+1} = x_k + A_{k+1} \varPhi _k\) reduces to
which is the generating equation of the Newton method [32] and hence obviously will converge to the minimizer of \(f_\alpha ^L + f_\alpha ^U\).
As we observe, the key point of the above method is to appropriately generate \(A_{k}\)’s. Due to the inherent computational difficulty to find the inverse of the Hessian \(\nabla ^2 \phi (x_{k+1})\) we consider an approximation that should satisfy
In this article we attempt to introduce a simple rank-two update of the sequence \(\{A_k\}\) that satisfy the above quasi-Newton equation (3.2).
Let \(A_k\) be the approximation of the kth iteration. We attempt to update \(A_k\) into \(A_{k+1}\) by adding two symmetric matrices, each of rank one as follows:
where \(u_k\) and \(v_k\) are two vectors in \(\mathbb {R}^n\), and \(p_k\) and \(q_k\) are two scalars which are to be determined by the quasi-Newton equation (3.2). Therefore, we now have
Evidently, \(v_k\) and \(w_k\) are not uniquely determined, but their obvious choices are \(v_k = \delta _k\) and \(w_k = A_k \varPhi _k\). Putting this values in (3.3), we obtain
Therefore,
We consider this equation to generate the sequence of above mentioned inverse Hessian approximation.
Therefore, accumulating all, we follow the following sequential way (Algorithm 1) to obtain an efficient solution of the considered (FOP) \(\min _{x \in \mathbb {R}^n} \widetilde{f}(x)\).
4 Convergence analysis
In this section, the convergence analysis of the proposed quasi-Newton method is performed. At first, we show the superlinear convergence rate of the proposed method. Then, we explore that the used search directions are descent and the Hessian approximations are positive definite. Finally, the quadratic termination property of the method is proved.
Theorem 1
(Superlinear convergence rate)
Let \(\widetilde{f}: \mathbb {R}^n \rightarrow \mathcal {F}(\mathbb {R})\) be thrice continuously gH-differentiable on \(\mathbb {R}^n\). Suppose that \(\bar{x}\) in \(\mathbb {R}^n\) satisfies the following three conditions:
-
(i)
\(\bar{x}\) is a local minimizer of \(\widetilde{f}^L_\alpha \) and \(\widetilde{f}^U_\alpha \),
-
(ii)
\(\int _0^1 \nabla ^2 \big (f_\alpha ^L(x) + f_\alpha ^U(x)\big ) \mathrm {d}\alpha \) is symmetric positive definite, and
-
(iii)
\(\int _0^1 \nabla ^2 \big (f_\alpha ^L(x) + f_\alpha ^U(x)\big ) \mathrm {d}\alpha \) is Lipschitzian with constant \(\gamma ^L+\gamma ^U\).
Then the iteration sequence \(\{x_k\}\) in Algorithm 1 converges to \(\bar{x}\) superlinearly if and only if
where \(\phi (x) = \int _0^1\left( f_\alpha ^L(x) + f_\alpha ^U(x)\right) d\alpha \).
Proof
According to the hypothesis (i), at \(\bar{x}\) we have
From hypothesis (ii), the Hessian matrix
is positive definite and a symmetric matrix.
With the help of hypothesis (iii), the function
is found to be Lipschitzian at \(\bar{x}\) with constant \(\gamma ^L + \gamma ^U\). Mathematically, there exists a neighborhood \(N_\epsilon (\bar{x})\) where
Towards proving the result, the following equivalence will be proved.
With the help of quasi-Newton equation (3.1), we have
Therefore,
It is evident to note that
Since \(\nabla ^2 \phi (\bar{x})\) is positive definite, we have \(\xi > 0\) and \(m \in \mathbb {N}\) such that for all \(k \ge m\)
Hence we now have
where \(c_k = \frac{\Vert x_{k+1} - \bar{x}\Vert }{\Vert x_k - \bar{x}\Vert }\). This inequality gives
This completes the proof of superlinear convergence of the sequence \(\{x_k\}\) in Algorithm 1.
Conversely, since \(\{x_k\}\) converges superlinearly to \(\bar{x}\) and \(\nabla \phi (\bar{x}) = 0\), we must have \(\beta > 0\) and \(p \in \mathbb {N}\) such that
Again due to superlinear convergence of \(\{x_k\}\), we have
Since \(\lim _{k \rightarrow \infty } \frac{\Vert x_{k+1} - x_k\Vert }{\Vert x_k - \bar{x}\Vert } = 1\), this inequality implies \(\lim _{k \rightarrow \infty } \frac{\Vert \varPhi _k\Vert }{\Vert x_{k+1} - x_k\Vert } = 0\). Hence the result follows. \(\square \)
Theorem 2
(Positive definiteness of the inverse Hessian approximation)
Let \(x_0 \in \mathbb {R}^n\) and the objective function \(\widetilde{f}\) in Algorithm 1 is twice continuously gH-differentiable. If
then the inverse Hessian approximations \(A_k\)’s are positive definite and the search directions \(d_k = - A_k \nabla \phi (x_k)\) are descent for all \(k = 1, 2, \cdots \).
Proof
If all \(A_k\)’s are positive definite, then according to the hypothesis, the search directions \(d_k = - A_k \nabla \phi (x_k)\) are trivially descent, since
Thus, we prove that all \(A_k\)’s are positive definite.
According to Algorithm 1, the matrix \(A_0\) is assumed to be positive definite.
Let \(A_1, A_2, \ldots , A_m\) are positive definite. We will show that \(A_{m+1}\) is positive definite. Then, by the principle of mathematical induction, the result will follow.
Let \(x \ne 0\) be a vector in \(\mathbb {R}^n\). With the help of the rank-two update of \(A_k\)’s in Algorithm 1, we obtain
Since \(A_m\) is a positive definite matrix, by Cholesky decomposition there exists a positive definite symmetric matrix \(D_m\) such that \(A_m = D_m~~D_m\).
Let \(a = D_m~x\) and \(b = D_m~ \varPhi _m\). Then
We note that
-
(i)
\(\delta _m^t~\varPhi _m = \alpha _m~d_m^t~\left( \nabla \phi (x_{m+1} - \nabla \phi (x_m)\right) = \alpha _m~ \nabla \phi (x_m)^t A_m \nabla \phi (x_m) > 0\), since \(d_m = - A_m \nabla \phi (x_m)\) is a descent direction, and hence \(\alpha _m > 0\), and \(d_m^t \nabla \phi (x_{m+1}) = 0\), and
-
(ii)
\(b^t~b > 0\), since \(\varPhi _m \ne 0\).
Eq. (4.1) gives
Below we prove that out of the two terms \(\frac{(a^t~a)(b^t~b) - (a^t~b)^2}{b^t~b}\) and \(\frac{(x^t~\delta _m)^2}{\delta _m \varPhi _m}\) in \(x^t~A_{m+1}~x\) at least one is strictly positive.
Since \(x \ne 0\), the equality in Cauchy–Schwarz inequality holds only when a || b, equivalently, \(x ~||~ \varPhi _m\).
Then there exists \(\lambda \ne 0\) such that \(x = \lambda \varPhi _m\) and hence \(\frac{(x^t \delta _m)^2}{\delta _m \varPhi _m} = \lambda ^2 \delta _m^t \varPhi _m > 0\). This yields the result. \(\square \)
Theorem 3
(Quadratic termination property)
Let \(\widetilde{f}(x) = \tfrac{1}{2}~ x^t ~\widetilde{A}_{n\times n}~ x \oplus \widetilde{B}_{n\times 1}^t~x \oplus \widetilde{C}_{1\times 1}\) be a quadratic fuzzy function with the end point matrices of \(\widetilde{A}(\alpha )\) are such that the Hessian matrices \(\int _0^1 \nabla ^2 f_\alpha ^L(x) d\alpha \) and \(\int _0^1 \nabla ^2 f_\alpha ^U(x) d\alpha \) are positive definite. Let H be the Hessian of
Then with the help of exact line search, the generated sequence \(\{\delta _k\}\) in Algorithm 1 satisfies the following properties for \(i = 0, 1, \ldots , m\) where \(m \le n-1 :\)
-
(i)
\(A_{i + 1} \varPhi _j = \delta _j\), \(j = 0, 1, 2, \ldots , i\),
-
(ii)
\(\delta _i^t~H~\delta _j = 0\), \(j = 0, 1, 2, \ldots , i - 1\), and
-
(iii)
The method terminates at \(m+1 \le n\) steps. If \(m = n - 1\), then \(A_n = H^{-1}\).
Proof
We note that according to the hypothesis,
is positive definite.
With the help of induction principle, we show the parts (i) and (ii).
The proof is trivial for \(i = 0\). Let the parts (i) and (ii) be true for some i. We will show that the results hold for \(i+1\).
As \(\nabla \phi (x_{i+1}) \ne 0\), by exact line search, and \(\varPhi _k = \nabla \phi (x_{k+1}) - \nabla \phi (x_k) = H (x_{k+1} - x_k)\) \(= H \delta _k\), \(k = 1, 2, \ldots , i\), the induction hypothesis implies that for \(j \le i\),
Thus, with the help of \(\delta _{i+1} = -\alpha _{i+1} A_{i+1} \nabla \phi (x_{i+1})\), it follows that
which completes the proof of the part (ii) holding for \(i+1\).
Below we show that part (i) holds for \(i+1\), i.e.,
From Algorithm 1, the rank-two update of \(A_{k}\)’s clearly shows
Thus, part (i) is true when \(j = i+1\).
From part (ii), it follows from the induction hypothesis that for \(j \le i\),
Therefore,
Hence, the part (i) follows.
Lastly, from part (ii), we note that the directions \(\delta _i\), \(i = 0, 1, 2, \ldots , m\) are conjugate. Hence, Algorithm 1 must terminate after \(m \le n\) steps. If \(m = n -1\), linear independence of \(\delta _i\)’s, \(i = 0, 1, \ldots , n-1\) clearly shows
which implies \(A_n = H^{-1}\). \(\square \)
5 Illustrative examples
In this section, two illustrative examples are presented to explore the computational procedure of Algorithm 1. In the first example, we consider a fuzzy quadratic form as the objective function. In the second, a general nonlinear fuzzy function is taken as the objective function.
Example 1
Consider the following quadratic fuzzy optimization problem:
Let us consider the initial approximation to the minimizer as \(x_0 = (x_1^0, x_2^0) = (0, 0)\). With the help of fuzzy arithmetic, the lower and the upper function can be obtained as
and
Therefore,
Here
Considering the initial matrix \(A_0 = I_2\), we calculate the sequence \(\{x_k\}\), \(x_k = (x_1^k, x_2^k)\), through the following equations:
The initial iteration (\(k = 0\))
The second iteration (\(k = 1\))
Hence \(\bar{x} = (-1, \tfrac{3}{2})\) is a nondominated solution of the considered problem. It is important to note that the method converged at the second iteration since the objective function is a quadratic fuzzy function. This convergence at second iteration supports the result in Theorem 3.
Example 2
Consider the following nonlinear fuzzy optimization problem:
With the help of fuzzy arithmetic, the lower and the upper function can be obtained as
and
Therefore,
Here
The following Table 1 exhibits the performance of Algorithm 1 on this example with the initial point as \(x_0 = (0,3)\), initial approximation of the Hessian as \(A_0 = I_2\) and the termination scalar as \(\epsilon = 0.01\). It is found that (2.12, 1.06) is a solution of the considered problem.
6 Conclusion
In this paper, a quasi-Newton method with rank-two modification has been derived to find a non-dominated solution of an unconstrained fuzzy optimization problem. In the optimality concept, the fuzzy-max ordering of a pair of fuzzy numbers has been used. The gH-differentiability of fuzzy functions have been employed to find the non-dominated solution point. An algorithmic implementation and the convergence analysis of the proposed technique has also been presented. The technique is found to have superlinear convergence rate and quadratic termination property. Two numerical examples have been explored to illustrate the proposed technique.
It is to note that unlike Newton method for fuzzy optimization problems [9, 32], the proposed method is derived without using the inverse of the concerned Hessian matrix. Instead, a sequence of positive definite inverse Hessian approximation \(\{A_k\}\) is used which satisfies the quasi-Newton equation \( A_{k+1} \varPhi _k = \delta _k\). Thus the derived method sidestepped the inherent computational difficulty to compute inverse of the concerned Hessian matrix. In this way, the proposed method is made as more efficient than the existing Newton method [9, 32].
The method can be easily observed as much similar to the classical DFP method [30] for the conventional optimization problem. In the similar way of the presented technique, it can be further extended to a BFGS-like method [32] for fuzzy optimization. Instead of the rank-two modification in the approximation of the inverse of the Hessian matrix, a rank-one modification can also be done. It is also to be observed that in Algorithm , we make use the exact line search technique along the descent direction \(d_k = - A_k \nabla \phi (x_k)\). However, an inexact line search technique [30] could have also been used. A future research on this rank-one modification and inexact line search technique for fuzzy optimization problem can be performed and implemented on Newton and quasi-Newton method.
It is important to observe that the proposed technique generates only one non-dominated solution for a fuzzy optimization problem. Development of a method which can capture complete nondominated solution set is an important issue for fuzzy optimization problem. Our next step of research would be focused on capturing complete nondominated solution of a fuzzy optimization problem with the help of the proposed quasi-Newton method.
Abbreviations
- \(\mathbb {R}\) :
-
The set of all real numbers
- \(\widetilde{A}\), \(\widetilde{B}\), ...and \(\widetilde{a}\), \(\widetilde{b}\), ...:
-
Fuzzy sets on \(\mathbb {R}\)
- \(\mathcal {F}(\mathbb {R})\) :
-
The set of all fuzzy numbers on \(\mathbb {R}\)
- \(\mu (.|\widetilde{A})\) :
-
Membership function of the fuzzy set \(\widetilde{A}\)
- \(\widetilde{A}(\alpha )\) :
-
\(\alpha \)-cut of the fuzzy set \(\widetilde{A}\)
- (l / a / r):
-
A triangular fuzzy number at a
- \({[}{\widetilde{a}}_\alpha ^{L}, {\widetilde{a}}_\alpha ^{U}{]}\) :
-
\(\alpha \)-cut of the triangular fuzzy number \({\widetilde{a}}\)
- \(\oplus \) :
-
Extended addition
- \(\ominus \) :
-
Extended substraction
- \(\odot \) :
-
Extended multiplication
- \(\ominus _{gH}\) :
-
Generalized Hukuhara difference
- \({[}\widetilde{f}_\alpha ^L(x), \widetilde{f}_\alpha ^U(x){]}\) :
-
\(\alpha \)-cut of the fuzzy function \(\widetilde{f}\) at x
- \(\widetilde{f}_\alpha ^L\) :
-
Lower function of the \(\alpha \)-cut of the fuzzy function \(\widetilde{f}\)
- \(\widetilde{f}_\alpha ^U\) :
-
Upper function of the \(\alpha \)-cut of the fuzzy function \(\widetilde{f}\)
- \(\tfrac{\partial \widetilde{f}}{\partial x_i}(x)\) :
-
Partial gH-derivative of \(\widetilde{f}\) with respect to \(x_i\) at x
- \(\nabla \widetilde{f}(x)\) :
-
gH-gradient of \(\widetilde{f}\) at x
- \(\nabla ^2 \widetilde{f}(x)\) :
-
gH-Hessian of \(\widetilde{f}\) at x
- \(\widetilde{a} \preceq \widetilde{b}\) :
-
\(\widetilde{a}^L_\alpha \le \widetilde{b}^L_\alpha \) and \(\widetilde{a}^U_\alpha \le \widetilde{b}^U_\alpha \) for all \(\alpha \in [0, 1]\)
- \(\widetilde{a} \prec \widetilde{b}\) :
-
\(\widetilde{a} \preceq \widetilde{b}\) and \(\exists \) \(\beta \in [0, 1]\) such that \(\widetilde{a}^L_\beta < \widetilde{b}^L_\beta \) or \(\widetilde{a}^U_\beta < \widetilde{b}^U_\beta \).
References
Andrei, N.: Hybrid conjugate gradient algorithm for unconstrained optimization. J. Optim. Theory Appl. 141(2), 249–264 (2009)
Anastassiou, G.A.: Fuzzy Mathematics: Approximation Theory, vol. 251. Springer, Berlin (2010)
Banks, H.T., Jacobs, M.Q.: A differential calculus for multifunctions. J. Math. Anal. Appl. 29(2), 246–272 (1970)
Bazaraa, M.S., Sherali, H.D., Shetty, C.M.: Nonlinear Programming: Theory and Algorithms, 3rd edn. Wiley, Hoboken (2006)
Bede, B., Gal, S.G.: Generalizations of the differentiability of fuzzy-number-valued functions with applications to fuzzy differential equations. Fuzzy Sets Syst. 151(3), 581–599 (2005)
Bede, B., Stefanini, L.: Generalized differentiability of fuzzy-valued functions. Fuzzy Sets Syst. 230, 119–141 (2013)
Bellman, R.E., Zadeh, L.A.: Decision making in a fuzzy environment. Manag. Sci. 17(4), B141–B164 (1970)
Cadenas, J.M., Verdegay, J.L.: Towards a new strategy for solving fuzzy optimization problems. Fuzzy Optim. Decis. Mak. 8(3), 231–244 (2009)
Chalco-Cano, Y., Silva, G.N., Rufián-Lizana, A.: On the Newton method for solving fuzzy optimization problems. Fuzzy Sets Syst. 272, 60–69 (2015)
Chakraborty, D., Ghosh, D.: Analytical fuzzy plane geometry II. Fuzzy Sets Syst. 243, 84–109 (2014)
Cheng, Z.: Duality theory in fuzzy mathematical programming problems with fuzzy coefficients. Comput. Math. Appl. 49(11–12), 1709–1730 (2005)
Cheng, C.-H.: A new approach for ranking fuzzy numbers by distance method. Fuzzy Sets Syst. 95, 307–317 (1998)
Dubois, D., Prade, H.: Towards fuzzy differential calculus part 3: Differentiation. Fuzzy Sets Syst. 8(3), 225–233 (1982)
Ghosh, D.: Newton method to obtain efficient solutions of the optimization problems with interval-valued objective functions. J. Appl. Math. Comput. 53(1–2), 709–731 (2017)
Ghosh, D.: A quasi-Newton method with rank-two update to solve interval optimization problems. Int. J. Appl. Comput. Math. 3(3), 1719–1738 (2017)
Ghosh, D., Chakraborty, D.: Analytical fuzzy plane geometry I. Fuzzy Sets Syst. 209, 66–83 (2012)
Ghosh, D., Chakraborty, D.: Analytical fuzzy plane geometry III. Fuzzy Sets Syst. 283, 83–107 (2016)
Ghosh, D., Chakraborty, D.: A method to capture the entire fuzzy non-dominated set of a fuzzy multi-criteria optimization problem. Fuzzy Sets Syst. 272, 1–29 (2015)
Ghosh, D., Chakraborty, D.: A new method to obtain fuzzy Pareto set of fuzzy multi-criteria optimization problems. J. Intel. Fuzzy Syst. 26(3), 1223–1234 (2014)
Ghosh, D., Chakraborty, D.: Quadratic interpolation technique to minimize univariable fuzzy functions. Int. J. Appl. Comput. Math. 3(2), 527–547 (2016)
Ghosh, D., Chakraborty, D.: Fuzzy ideal cone: a method to obtain complete fuzzy non-dominated set of fuzzy multi-criteria optimization problems with fuzzy parameters. In: Proceedings of IEEE International Conference on Fuzzy Systems 2013, FUZZ IEEE 2013, IEEE Xplore, pp. 1–8 (2013)
Goetschel, R., Voxman, W.: Elementary fuzzy calculus. Fuzzy Sets Syst. 18(1), 31–43 (1986)
Gong, Z.-T., Li, H.-X.: Saddle point optimality conditions in fuzzy optimization problems. In: Cao, B.-Y., Zhang, C.-Y., Li, T.-F. (eds.) Fuzzy Info and Engineering, vol. 54, pp. 7–14. Springer, ASC, Berlin (2009)
Lai, Y.-J., Hwang, C.-L.: Fuzzy mathematical programming: methods and applications. Lecture Notes in Economics and Mathematical Systems, vol. 394. Springer, New York (1992)
Lai, Y.-J., Hwang, C.-L.: Fuzzy multiple objective decision making: methods and applications. Lecture Notes in Economics and Mathematical Systems, vol. 404. Springer, New York (1994)
Lee-Kwang, H., Lee, J.-H.: A method for ranking fuzzy numbers and its application to decision-making. IEEE Tr Fuzzy Syst. 7(6), 677–685 (1999)
Lodwick, W.A., Kacprzyk, J.: Fuzzy Optimization: Recent Advances and Applications, vol. 254. Physica, New York (2010)
Luhandjula, M.K.: Fuzzy optimization: milestones and perspectives. Fuzzy Sets Syst. 274, 4–11 (2015)
Mitchell, H.B., Schaefer, P.A.: On ordering fuzzy numbers. Int. J. Intel. Syst. 15(11), 981–993 (2000)
Nocedal, J., Wright, S.J.: Numerical Optimization, 2nd edn. Springer, New York (2006)
Peng, J., Liu, H., Gang, S.: Ranking fuzzy variables in terms of credibility measure. In: Fuzzy System Knowledge Discovery, Springer, pp. 217–220 (2006)
Pirzada, U.M., Pathak, V.D.: Newton method for solving the multi-variable fuzzy optimization problem. J. Optim. Theory. Appl. 156(3), 867–881 (2013)
Puri, M.L., Ralescu, D.A.: Differentials of fuzzy functions. J. Math. Anal. Appl. 91(2), 552–558 (1983)
Ramík, J., Rimanek, J.: Inequality relation between fuzzy numbers and its use in fuzzy optimization. Fuzzy Sets Syst. 16(2), 123–138 (1985)
Ramík, J.: Optimal solutions in optimization problem with objective function depending on fuzzy parameters. Fuzzy Sets Syst. 158, 1873–1881 (2007)
Słowínski, R. (ed.): Fuzzy Sets in Decision Analysis, Operations Research and Statistics. Kluwer, Boston (1998)
Stefanini, L.: A generalization of Hukuhara difference and division for interval and fuzzy arithmetic. Fuzzy Sets Syst. 161, 1564–1584 (2010)
Sun, H., Wu, J.: A new approach for ranking fuzzy numbers based on fuzzy simulation analysis method. Appl. Math. Comp. 174(1), 755–767 (2006)
Wang, X., Kerre, E.E.: Reasonable properties for the ordering of fuzzy quantities. Fuzzy Sets Syst. 118(3), 375–385 (2001)
Wang, Z.-X., Liu, Y.-L., Fan, Z.-P., Feng, B.: Ranking L-R fuzzy number based on deviation degree. Inf. Sci. 179(13), 2070–2077 (2009)
Wu, H.-C.: Duality theory in fuzzy linear programming problems with fuzzy coefficients. Fuzzy Optim. Decis. Mak. 2(1), 61–73 (2003)
Wu, H.-C.: Duality theory in fuzzy optimization problems. Fuzzy Optim. Decis. Mak. 3(3), 345–365 (2004)
Wu, H.-C.: Duality theory in fuzzy optimization problems formulated by the Wolfe’s primal and dual pair. Fuzzy Optim. Decis. Mak. 6(3), 179–198 (2007)
Wu, H.-C.: The Karush–Kuhn–Tucker optimality conditions for the optimization problem with fuzzy-valued objective function. Math. Methods Oper. Res. 66(2), 203–224 (2007)
Wu, H.-C.: The optimality conditions for optimization problems with fuzzy-valued objective functions. Optimization 57(3), 473–489 (2008)
Wu, H.-C.: The optimality conditions for optimization problems with convex constraints and multiple fuzzy-valued objective functions. Fuzzy Optim. Dec. Mak. 8(3), 295–321 (2009)
Wu, C., Song, S., Lee, S.S.: Approximate solutions, existence and uniqueness of the Cauchy problem of fuzzy differential equations. J. Math. Anal. Appl. 202(2), 629–644 (1996)
Wu, H.-C.: Saddle point optimality conditions in fuzzy optimization problems. Fuzzy Optim. Decis. Mak. 3(3), 261–273 (2003)
Yager, R.R., Dimitar, F.: On ranking fuzzy numbers using valuations. Int. J. Intel. Syst. 14(12), 1249–1268 (1999)
Yao, J.S., Wu, K.: Ranking fuzzy numbers based on decomposition principle and signed distance. Fuzzy Sets Syst. 116(2), 275–288 (2000)
Zeng, L.C., Schaible, S., Yao, J.C.: Hybrid steepest descent methods for zeros of nonlinear operators with applications to variational inequalities. J. Optim. Theory Appl. 141(1), 75–91 (2009)
Zhong, Y., Shi, Y.: Duality in fuzzy multi-criteria and multi-constraint level linear programming: a parametric approach. Fuzzy Sets Syst. 132(3), 335–346 (2002)
Zimmermann, H.J.: Fuzzy Set Theory and its Applications, 4th edn. Springer, New York (2001)
Acknowledgements
The first author gratefully acknowledges the financial support through Early Career Research Award (ECR/2015/000467), Science and Engineering Research Board, Government of India.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Ghosh, D., Ghosh, D. & Bhuiya, S.K. A quasi-Newton method with rank-two update to solve fuzzy optimization problems. SeMA 75, 285–303 (2018). https://doi.org/10.1007/s40324-017-0134-0
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s40324-017-0134-0