Abstract
Multiobjective optimization problems typically have conflicting objectives, and a gain in one objective very often is an expense in another. Using the concept of Pareto optimality, we investigate a multiobjective bilevel optimization problem (say, P). Our approach consists of proving that P is locally equivalent to a single level optimization problem, where the nonsmooth Mangasarian–Fromovitz constraint qualification may hold at any feasible solution. With the help of a special scalarization function introduced in optimization by Hiriart–Urruty, we convert our single level optimization problem into another problem and give necessary optimality conditions for the initial multiobjective bilevel optimization problem P.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Bilevel optimization is an important research area, and many researchers have made contributions [1–15]. It is a sequence of two optimization problems in which the feasible region of the upper-level problem is determined implicitly by the solution set of the lower-level problem. The origin of the bilevel optimization problem can be traced back to von Stackelberg [10], who used it to model the market economy in 1934.
Bard [2] studied the linear bilevel optimization problem and developed first-order necessary optimality conditions. Under semi-Lipschitz property, Zhang [14] extended the classic approach to study nonsmooth problem data and derived existence and optimality conditions for the problems in terms of the graph of the solution multifunction to the lower level problem. Dempe [4] and Outrata [8] derived necessary conditions for the case, where the solution set of the lower level problem is a singleton. Using the optimal value function in the lower-level problem, Ye and Zhu [12, 13] reformulated the bilevel problem as a single level nonconvex optimization problem, where the nonsmooth Mangasarian–Fromovitz constraint qualification does not hold at any feasible solution; under the partial calmness condition, they derived optimality conditions for the general bilevel optimization problems, without convexity assumption on the lower level problem and without the assumption that the solution set of the lower level problem be a singleton. Later, Bao, Gupta and Mordukhovich [17], and Zhang, Truong and Zhang [15] used the variational approach and got necessary conditions directly by using advanced tools of variational analysis such as the extremal principle and the separation theorems for convex sets. For more details on bilevel optimization, see Bard [3], Dempe [5, 6], Vicente and Calamai [11], and Shimizu et al. [9].
Multiobjective optimization problems typically have conflicting objectives, and a gain in one objective very often is an expense in another. We investigate a multiobjective bilevel optimization problem using the optimistic approach. This means we assume that the leader presupposes cooperation of the follower in the sense that the latter will choose in each time that solution in the solution set of his/her parametric optimization problem which is best suited with respect to the leader’s objective function. Using the concept of Pareto optimality, together with a special scalarization function introduced in optimization by Hiriart–Urruty [20, 21], we give necessary optimality conditions. Several intermediate optimization problems are introduced to help us in our investigation.
The outline of the paper is as follows: preliminary results are described in Sect. 2; necessary optimality conditions are established in Sect. 3; conclusions are given in Sect. 4.
2 Preliminaries
Let C⊂ℝn be a pointed (C∩−C={0}), closed, and convex cone with nonempty interior introducing a partial order in ℝn, and let A be a nonempty subset of ℝn. \(\overline{z}\in A\) is said to be a Pareto (respectively, a weak Pareto) minimal vector of A with respect to C iff
(respectively, \(A\subset \overline{z}+\mathbb{R}^{n}\backslash -\mathrm{int}\, C\)), where int denotes the topological interior. For a subset S of ℝn, we consider the function
where d(y,S):=inf{∥u−y∥:u∈S}. This function is introduced in Hiriart–Urruty [20, 21], and used by Ciligot–Travain [19], and Amahroq and Taa [16].
Let us recall the following result of [20].
Proposition 2.1
[20]
Let S⊂ℝn be a closed and convex cone with nonempty interior and S≠ℝn. The function Δ S is convex, positively homogeneous, 1-Lipschitzian, decreasing on ℝn with respect to the order introduced by S.
Of course, (ℝn\S)={y∈ℝn:Δ S (y)>0}, int S={y∈ℝn:Δ S (y)<0} and the boundary of S is the set bd S={y∈ℝn:Δ S (y)=0}.
As a direct consequence of Proposition 2.1, one has the following result.
Proposition 2.2
[19]
Let S⊂ℝn be a nonempty, closed, and convex cone with nonempty interior. Then for all y∈ℝn,
Here, the set ∂ ca f(x) designs the subdifferential of convex analysis of f at x.
For all the rest, for a locally Lipschitz mapping f:ℝn→ℝ, the set ∂f(x) denotes the Clarke generalized Jacobian of f at x; i.e.,
Recall the following interesting results, which are due to Clarke [18]. For more details, we refer the interested reader to [18].
Proposition 2.3
[18]
Suppose that {f i } i∈{1,…,n} be a finite collection of functions each of which is Lipschitz near \(\overline{x}\). The function h defined by
is easily seen to be Lipschitz near \(\overline{x}\) as well. Moreover,
where \(I ( \overline{x} ) :=\{i:f_{i} (\overline{x} ) =h ( \overline{x} )\}\) and conv denotes the convex hull.
Proposition 2.4
[18]
Suppose that T be separable, and {f t } t∈T be a collection of locally Lipschitz functions f t at \(\overline{x}\). Set
Then h is Lipschitz near \(\overline{x}\) and
3 Necessary Optimality Conditions
Consider the following multiobjective bilevel optimization problem:
where, for each \(x\in \mathbb{R}^{n_{1}}\), S(x) is the solution set of the following parametric optimization problem (the lower level problem)
where \(f:\mathbb{R}^{n_{1}}\times \mathbb{R}^{n_{2}}\longrightarrow \mathbb{R}\), \(g_{s}:\mathbb{R}^{n_{1}}\times \mathbb{R}^{n_{2}}\longrightarrow \mathbb{R}\), s∈S:={1,…,q}, \(G_{j}:\mathbb{R}^{n_{1}}\times \mathbb{R}^{n_{2}}\rightarrow \mathbb{R}\), j∈J:={1,…,p}, and \(F_{i}:\mathbb{R}^{n_{1}}\times \mathbb{R}^{n_{2}}\rightarrow \mathbb{R}\), i∈I:={1,…,n}, are given locally Lipschitz functions; n 1≥1 and n 2≥1 are integers.
A pair \((\overline{x},\overline{y})\) is said to be a local efficient (respectively, weak local efficient) solution of (P) iff there exists a neighborhood V of \((\overline{x}, \overline{y})\) such that \(F( \overline{x},\overline{y})\) is a local Pareto (respectively, weak local Pareto) minimal vector of \(F( \overline{S}\cap V)\) where
In order to derive optimality conditions, we consider a new single level problem which is locally equivalent to the bilevel multiobjective problem (P) at the optimal solution.
Denote by
the feasible region of the lower level problem (P x ).
Let \((\overline{x},\overline{y})\) be a local weak efficient solution of (P). Then there exist neighborhoods U 0 of \(\overline{x}\) and V 0 of \(\overline{y}\) such that
Throughout this paper, we assume that the set-valued map Y is uniformly bounded around \(\overline{x}\); there exists a bounded neighborhood \(U(\overline{x},\overline{y})\) of \(( \overline{x},\overline{y})\) such that ⋃ x∈U Y(x) is bounded. Here,
Taking \(U_{\overline{x}}:=U_{0}\cap U\), the set \({\bigcup}_{x\in U_{\overline{x}}}Y ( x )\) is also bounded. Thus, \(\mathrm{cl}{\bigcup }_{x\in U_{\overline{x}}}Y ( x )\) is compact. Then
is a nonempty compact set that contains an open neighborhood of \(\mathrm{cl}\bigcup_{x\in U_{\overline{x}}}Y(x)\). Here, cl and \(\overline{B}_{R^{n_{2}}}\) denote respectively the closure and the closed unit ball of \(R^{n_{2}}\). The following lemmas are crucial for our investigations; we prove that a local weak efficient solution of (P) is also a local weak efficient solution of
where
and
Taking \(x\in \mathbb{R}^{n_{1}}\), we consider the optimal value function of the lower level problem (P x ) defined by
Lemma 3.1
Proof
Let \(x\in U_{\overline{x}}\) and y∈Y(x). Since Θ is compact,
Since \(x\in U_{\overline{x}}\), one has Y(x)⊆Θ and
Consequently,
By definition of the optimal value function V(x), one has
Then
□
Lemma 3.2
Proof
-
Let us prove that
$$\left \{ \begin{array}{l}( x,y ) \in \mathbb{R}^{n_{1}}\times \mathbb{R}^{n_{2}}: \\[3pt]x\in U_{\overline{x}},y\in Y ( x ) , \\[3pt]f ( x,y ) -V ( x ) =0.\end{array} \right \} \subseteq \left \{ \begin{array}{l}( x,y ) \in \mathbb{R}^{n_{1}}\times \mathbb{R}^{n_{2}}: \\[3pt]x\in U_{\overline{x}},y\in Y ( x ) , \\[3pt]\varPsi ( x,y ) =0.\end{array} \right \} .$$
Let \(x\in U_{\overline{x}}\) and y∈Y(x) be such that
Consequently, y is a global minimizer of (P x ) for fixed x. Since
one deduces that
Suppose that Ψ(x,y)>0. Then there exists z∈Θ such that
That is,
Thus,
Then z∈Y(x) such that
A contradiction, since y is a global minimizer of (P x ) for fixed x. One concludes that
-
Let us prove that
$$\left \{\begin{array}{l}( x,y ) \in \mathbb{R}^{n_{1}}\times \mathbb{R}^{n_{2}}: \\[3pt]x\in U_{\overline{x}},y\in Y ( x ) , \\[3pt]f ( x,y ) -V ( x ) =0.\end{array}\right \}\supseteq \left \{\begin{array}{l}( x,y ) \in \mathbb{R}^{n_{1}}\times \mathbb{R}^{n_{2}}: \\[3pt]x\in U_{\overline{x}},y\in Y ( x ) , \\[3pt]\varPsi ( x,y ) =0.\end{array} \right \} .$$
Let \(x\in U_{\overline{x}}\) and y∈Y(x) be such that
Then
Since Y(x)⊆Θ, one gets also
That is, for every z∈Y(x), one has
Since
one has
Consequently,
Thus,
Since
one deduces that f(x,y)−V(x)=0. □
Lemma 3.3
If \((\overline{x},\overline{y})\) is a local weak efficient solution of (P), then the solution set of the problem \(\max_{z\in \varTheta }\psi (\overline{x},\overline{y},z)\) is given by \(S(\overline{x})\).
Proof
Since \((\overline{x},\overline{y})\) is a local weak efficient solution of (P), one has
For any \(\overline{z}\in S(\overline{x} )\), one gets
and
Consequently,
To get the result, since \(S( \overline{x})\subset \varTheta \), it suffices to prove that
By contrary, suppose that there exists z∈Θ such that
Then
Thus,
A contradiction with \(\overline{y}\in S ( \overline{x})\) □
Remark 3.1
Under the following hypotheses (H1), (H2), (H3), (H4) and (H5), the optimization problem (P) has at least one optimal solution.
- (H1)::
-
F i (.,.) is lower semicontinuous (l.s.c.) on \(\mathbb{R}^{n_{1}}\times \mathbb{R}^{n_{2}}\) for all i∈I;
- (H2)::
-
G j (.,.) is lower semicontinuous (l.s.c.) on \(\mathbb{R}^{n_{1}}\times \mathbb{R}^{n_{2}}\) for all j∈J;
- (H3)::
-
g s (.,.) is lower semicontinuous (l.s.c.) on \(\mathbb{R}^{n_{1}}\times \mathbb{R}^{n_{2}}\) for all s∈S;
- (H4)::
-
Ψ(.,.) is lower semicontinuous (l.s.c.) on \(\mathbb{R}^{n_{1}}\times\mathbb{R}^{n_{2}}\);
- (H5)::
-
The problem (P∗) has at least one feasible solution and its feasible set is bounded.
Especially, under these conditions, \(\overline{S}\) is a nonempty compact set and F i (.,.) is lower semicontinuous on \(\mathbb{R}^{n_{1}}\times \mathbb{R}^{n_{2}}\) for all i∈I.
Remark 3.2
The function Ψ(.,.) is locally Lipschitz near \((\overline{x},\overline{y})\).
Let
Let m 0:=p+q+1,m:=n+m 0,Π 0:={1,…,m 0} and Π:={1,…,m}. We denote by G, g, π, and \(\overleftrightarrow{\psi}\) the mappings defined as follows:
and
Here,
and
For \(t^{\ast }= ( \mu^{\ast },\upsilon^{\ast },\gamma^{\ast } )\in \mathbb{R}_{+}^{p}\times \mathbb{R}_{+}^{q}\times \mathbb{R}_{+}\), we consider the set
Let \(\overline{u}:=(\overline{x},\overline{y})\in E\). In the following theorem, we will need
and
Definition 3.1
We say that the nonsmooth Abadie constraint qualification holds at \(( \overline{x},\overline{y})\in E\) iff
Here, \(K ( E,\overline{u})\) denotes the contingent cone to E at \(\overline{u}\).
Remark 3.3
In general,
The following theorem gives necessary optimality conditions.
Theorem 3.1
Let \(\overline{u}= ( \overline{x},\overline{y} ) \in E\) be a local weak efficient solution of (P). Suppose that the nonsmooth Abadie constraint qualification holds at \(\overline{u}\). Then there exist \(y^{\ast }\in ( -\mathbb{R}_{+}^{n} )^{\circ }\setminus \{ 0 \}\) such that
Proof
Since \(\overline{u}= ( \overline{x},\overline{y} )\in E\) is a local weak efficient solution of (P), it is also a local weak efficient solution of (P∗) with respect to \(\mathbb{R}_{+}^{n}\). Setting
one deduces that \(\overline{u}\) minimizes \(\varDelta _{-\mathrm{int}\,\mathbb{R}_{+}^{n}}\circ \overleftrightarrow{F}\) over the feasible set E. Since \(\varDelta _{-\mathrm{int}\, \mathbb{R}_{+}^{n}}\circ \overleftrightarrow{F}\) is a locally Lipschitz function, denote by k 0>0 the Lipschitz constant of \(\varDelta _{-\mathrm{int}\, \mathbb{R}_{+}^{n}}\circ \overleftrightarrow{F}\). Then
Since the nonsmooth Abadie constraint qualification holds at \(\overline{u}\),
Thus,
where
denotes the convex cone generated by \(\varDelta (\overline{u} ) \times \partial \pi (\overline{u})\).
Thus, there exists \(u^{\ast }\in \partial(\varDelta _{-\mathrm{int}\, \mathbb{R}_{+}^{n}}\circ \overleftrightarrow{F} ) ( \overline{u} )\) such that the function \(d\mapsto \langle u^{\ast },d \rangle +\delta_{C^{0}(\overline{u})}(d)\) attains its minimum at 0, where \(C^{0} ( \overline{u})\) is the polar cone of \(C(\overline{u})\) and \(\delta_{C^{0}(\overline{u})}\) is the indicator function of \(C^{0} ( \overline{u})\).
Applying the chain rule [18], there exist \(v^{\ast }\in \partial_{ca}\varDelta _{-\mathrm{int}\,\mathbb{R}_{+}^{n}} ( 0 )\) such that
Since \(\varDelta _{-\mathrm{int}\, \mathbb{R}_{+}^{n}}(.)\) is a convex function and \(\varDelta _{-\mathrm{int}\,\mathbb{R}_{+}^{n}}(0)=0\) we have for all v∈ℝn
and hence for all \(v\in - ( \mathbb{R}_{+}^{n} )\)
That is \(v^{\ast }\in (-(\mathbb{R}_{+}^{n}))^{\circ }\). We conclude from Proposition 2.2 that v ∗≠0. □
In the following lemma, we prove that any local weak efficient solution of (P∗) is also a local weak efficient solution of the unconstrained optimization problem
Lemma 3.4
Let \((\overline{x},\overline{y})\in E\) be a local weak efficient solution of (P∗). Then \((\overline{x},\overline{y})\) is a local weak efficient solution of \(( \mathrm{P}_{1}^{\ast})\) with respect to \(\mathbb{R}_{+}^{m}\).
Proof
Suppose the contrary. One can find sequences \(( x_{n},y_{n} )\rightarrow ( \overline{x},\overline{y})\) such that
Thus,
Since \((\overline{x},\overline{y})\in E\), one has
Consequently,
Then
A contradiction with the fact that \((\overline{x},\overline{y})\in E\) be a local weak efficient solution of (P∗). □
The theorem below uses Lemma 3.4 to get necessary optimality conditions.
Theorem 3.2
Let \((\overline{x},\overline{y})\in E\) be a local weak efficient solution of (P). Then there exist \(y^{\ast }\in( -\mathbb{R}_{+}^{m} )^{\circ }\setminus\{0\}\) such that
Proof
Since \((\overline{x},\overline{y})\in E\) is a local weak efficient solution of (P), it is also a local weak efficient solution of \((\mathrm{P}_{1}^{\ast})\) with respect to \(\mathbb{R}_{+}^{m}\).
The proof of this theorem consists of several steps.
-
Let us prove that \((\overline{x},\overline{y})\) solves locally the following scalar convex minimization problem:
$$\left \{ \begin{array}{l}\text{Minimize}\quad \varDelta _{-\mathrm{int}\, \mathbb{R}_{+}^{m} } \bigl(\overleftrightarrow{\psi }_{1} ( x,y ) -\overleftrightarrow{\psi }_{1} ( \overline{x},\overline{y} ) ,\ldots,\overleftrightarrow{\psi }_{m} ( x,y ) -\overleftrightarrow{\psi }_{m} ( \overline{x},\overline{y} ) \bigr) \\[3pt]\text{subject to}\quad ( x,y ) \in \mathbb{R}^{n_{1}}\times \mathbb{R}^{n_{2}}.\end{array} \right .$$
By assumption, \((\overline{x},\overline{y})\in E\) is a local weak efficient solution of \((\mathrm{P}_{1}^{\ast})\) with respect to \(\mathbb{R}_{+}^{m}\); there exists a neighborhood V of \((\overline{x},\overline{y})\) such that
Hence by Proposition 2.1,
Since \(\varDelta _{-\mathrm{int}\, \mathbb{R}_{+}^{m} } ( 0 )=0\), it follows that \((\overline{x},\overline{y})\) solves locally the problem
-
Set
$$\overleftrightarrow{\psi }_{\overline{u}} ( x,y ) :=\overleftrightarrow{\psi } ( u ) -\overleftrightarrow{\psi } ( \overline{u} ) .$$
As \(\varDelta _{-int}\, \mathbb{R}_{+}^{m}\) and \(\overleftrightarrow{\psi }\) are locally Lipschitz, then there exists α≥1 such that \(\varDelta _{-\mathrm{int}}\, \mathbb{R}_{+}^{m}\circ \overleftrightarrow{\psi }\) is locally Lipschitz with a Lipschitz constant α. Consequently, \(\varDelta _{-\mathrm{int}}\, \mathbb{R}_{+}^{m}\circ \overleftrightarrow{\psi }_{\overline{u}}\) is Lipschitzian. Then
Applying the chain rule [18], there exist \(y^{\ast }\in \partial \varDelta _{-\mathrm{int}\, \mathbb{R}_{+}^{m} } ( 0 )\) such that
Since \(\varDelta _{-\mathrm{int}\, \mathbb{R}_{+}^{m} }( . )\) is a convex function and \(\varDelta _{-\mathrm{int}\, \mathbb{R}_{+}^{n} } (0 ) =0\) we have for all v∈ℝm
and hence for all \(v\in - ( \mathbb{R}_{+}^{m})\)
That is \(v^{\ast }\in ( -\mathbb{R}_{+}^{m} )^{\circ }\). From Proposition 2.2, we have that v ∗≠0.
Thus, there exist \(v^{\ast }\in ( -\mathbb{R}_{+}^{m} )^{\circ }\setminus \{ 0 \}\) such that
Applying the sum rule [18], we obtain
Finally, there exist \(y^{\ast }\in ( -\mathbb{R}_{+}^{m} )^{\circ }\setminus \{ 0 \}\) such that
□
Remark 3.4
Using Propositions 2.3 and 2.4, one gets
and
Moreover, setting
and
one obtains
4 Conclusions
As a hierarchical optimization problem, the multiobjective bilevel problem (P) combines decisions of the so-called leader and the so-called follower. While the leader has the first choice and the follower reacts optimally on the leaders selection, the leaders aim consists in finding such a selection which, together with the followers response, minimizes the mapping F with respect to a given cone. With the help of the concept of Pareto optimality, together with a special scalarization function introduced by Hiriart–Urruty, we give necessary optimality conditions. Our approach consists of proving that (P) is locally equivalent to a single level optimization problem, where the nonsmooth Mangasarian–Fromovitz constraint qualification may hold at any feasible solution.
References
Babahadda, H., Gadhi, N.: Necessary optimality conditions for bilevel optimization problems using convexificators. J. Glob. Optim. 34, 535–549 (2006)
Bard, J.F.: Optimality conditions for the bilevel programming problem. Nav. Res. Logist. Q. 31, 13–26 (1984)
Bard, J.F.: Practical Bilevel Optimization: Algorithms and Applications. Kluwer Academic, Dordrecht (1998)
Dempe, S.: A necessary and sufficient optimality condition for bilevel programming problem. Optimization 25, 341–354 (1992)
Dempe, S.: Foundations of Bilevel Programming. Kluwer Academic, Dordrecht (2002)
Dempe, S.: Annotated bibliography on bilevel programming and mathematical programs with equilibrium constraints. Optimization 52, 333–359 (2003)
Dempe, S., Gadhi, N.: Second order optimality conditions for bilevel set optimization problems. J. Glob. Optim. 47, 233–245 (2010)
Outrata, J.V.: On necessary optimality conditions for Stackelberg problems. J. Optim. Theory Appl. 76, 306–320 (1993)
Shimizu, K., Ishizuka, Y., Bard, J.F.: Nondifferentiable and Two-Level Mathematical Programming. Kluwer Academic, Boston (1997)
Stackelberg, H.v.: Marktform und Gleichgewicht. Springer, Berlin (1934)
Vicente, L.N., Calamai, P.H.: Bilevel and multilevel programming: A bibliography review. J. Glob. Optim. 5, 291–306 (1994)
Ye, J.J., Zhu, D.L.: Optimality conditions for bilevel programming problems. Optimization 33, 9–27 (1995)
Ye, J.J., Zhu, D.L.: A note on optimality conditions for bilevel programming problems. Optimization 39, 361–366 (1997)
Zhang, R.: Problems of hierarchical optimization in finite dimensions. SIAM J. Optim. 4, 521–536 (1995)
Zhang, R., Truong, B., Zhang, Q.: Multistage hierarchical optimization problems with multi-criterion objectives. J. Ind. Manag. Optim. 7, 103–115 (2011)
Amahroq, T., Taa, A.: On Lagrange–Kuhn–Tucker multipliers for multiobjective optimization problems. Optimization 41, 159–172 (1997)
Bao, T.Q., Gupta, P., Mordukhovich, B.S.: Necessary conditions in multiobjective optimization with equilibrium constraints. J. Optim. Theory Appl. 135, 179–203 (2007)
Clarke, F.H.: Optimization and Nonsmooth Analysis. Wiley-Interscience, New York (1983)
Ciligot-Travain, M.: On Lagrange–Kuhn–Tucker multipliers for Pareto optimization problem. Numer. Funct. Anal. Optim. 15, 689–693 (1994)
Hiriart-Urruty, J.B.: Tangent cones, generalized gradients and mathematical programming in Banach spaces. Math. Oper. Res. 4, 79–97 (1979)
Hiriart-Urruty, J.B., Lemaréchal, C.: Convex Analysis and Minimization Algorithms I. Springer, Berlin (1993)
Acknowledgements
This work has been supported by the Alexander-von Humboldt foundation.
Author information
Authors and Affiliations
Corresponding author
Additional information
Communicated by Boris Mordukhovich.
Rights and permissions
About this article
Cite this article
Gadhi, N., Dempe, S. Necessary Optimality Conditions and a New Approach to Multiobjective Bilevel Optimization Problems. J Optim Theory Appl 155, 100–114 (2012). https://doi.org/10.1007/s10957-012-0046-1
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10957-012-0046-1