Fakultät für Mathematik und Informatik
Preprint 2009-02
Ayalew Getachew Mersha, Stephan Dempe
Feasible Direction Method for
Bilevel Programming Problem
ISSN 1433-9307
Ayalew Getachew Mersha, Stephan Dempe
Feasible Direction Method for Bilevel
Programming Problem
TU Bergakademie Freiberg
Fakultät für Mathematik und Informatik
Institut für Numerische Mathematik und Optimierung
Akademiestr. 6 (Mittelbau)
09596 FREIBERG
http://www.mathe.tu-freiberg.de
ISSN 1433 – 9307
Herausgeber:
Dekan der Fakultät für Mathematik und Informatik
Herstellung:
Medienzentrum der TU Bergakademie Freiberg
FEASIBLE DIRECTION METHOD FOR BILEVEL
PROGRAMMING PROBLEM
AYALEW GETACHEW MERSHA AND STEPHAN DEMPE
Abstract. In this paper, we investigate the application of feasible direction
method for an optimistic nonlinear bilevel programming problem. The convex
lower level problem of an optimistic nonlinear bilevel programming problem
is replaced by relaxed KKT conditions. The feasible direction method developed by Topkis and Veinott [22] is applied to the auxiliary problem to get a
Bouligand stationary point for an optimistic bilevel programming problem.
1. Introduction
In this paper we consider the following bilevel programming problem.
F (x, y) → “ min ”
x
(1.1)
≤
G(x)
0, y ∈ Ψ(x)
with
(1.2)
y ∈ Ψ(x) := arg min{f (x, y) : g(x, y) ≤ 0}
y
where F, f : Rn+m → R, g : Rn+m → Rp , G : Rn → Rl are at least twice continuously differentiable functions. The following assumption is made throughout this
paper.
Assumption 1. f (x, .), gi (x, .), i = 1, . . . , p are convex for every x.
If the solution of the lower level problem (1.2) corresponding to any parameter
x is not unique, the upper level problem (1.1) is not a well defined optimization
problem. In this case the bilevel programming problem may not be solvable. This
situation is depicted by examples constructed in [2] and [4, example on page 121].
The following Lemma substantiates this claim.
Lemma 1.1. [3] If Ψ(x) is not a singleton for all parameters x, the leader may
not achieve his infimum objective function value.
To overcome such an unpleasant situation there are three strategies available for
the leader. The first strategy is to replace min with inf in the formulation of problem
(1.1) and define optimal solutions. The second strategy is to allow cooperation
between the two players: the upper level decision maker, the leader, and the lower
level decision maker, the follower. This resulted in the so called optimistic or weak
bilevel programming problem. The third is a conservative strategy. In this case the
Work of the first author was supported by DAAD(Deutscher Akademischer Austausch Dienst)
with a scholarship grant.
1
2
AYALEW GETACHEW MERSHA AND STEPHAN DEMPE
leader is forced to bound the damage caused by a follower’s “unfavorable” choice.
Define the following sets,
M
M (X)
M (x)
I(x◦ , y◦ )
:= {(x, y) : G(x) ≤ 0, g(x, y) ≤ 0},
:= {x ∈ Rn : ∃y s.t. (x, y) ∈ M },
:= {y ∈ Rm : g(x, y) ≤ 0, }
:= {i : gi (x◦ , y◦ ) = 0}.
Definition 1.2. The lower level problem (1.2) is said to satisfy the MangasarianFromowitz Constraint Qualification (MFCQ) at (x◦ , y◦ ), y◦ ∈ Ψ(x◦ ) if
r ∈ Rm r> ∇y gi (x◦ , y◦ ) < 0, ∀i ∈ I(x◦ , y◦ ) 6= ∅.
The Lagrangian function for the lower level problem is given by
L(x, y, λ) := f (x, y) + λ> g(x, y).
Consider the set of Lagrange multipliers
Λ(x, y) := {λ : λ ≥ 0, λ> g(x, y) = 0, ∇y L(x, y, λ) = 0}.
Let
J(λ) := {j : λj > 0}.
Definition 1.3. The lower level problem (1.2) is said to satisfy a strong sufficient optimality condition of second order (SSOC) at a point (x◦ , y◦ ) if for each
λ ∈ Λ(x◦ , y◦ ) and for every nonzero element of the set
r ∈ Rm r> ∇y gi (x◦ , y◦ ) = 0, ∀i ∈ J(λ)
we have
rT ∇2yy L(x◦ , y◦ , λ)r > 0.
Definition 1.4. The constant rank constraint qualification (CRCQ) is valid for
problem (1.2) at a point (x◦ , y◦ ) if there is an open neighborhood Wε (x◦ , y◦ ), ε > 0
of (x◦ , y◦ ) such that for each subsets I ⊆ I(x◦ , y◦ ), the family of gradient vectors
{∇y gi (x, y) : i ∈ I} has the same rank for all (x, y) ∈ Wε (x◦ , y◦ ).
Assume the following is satisfied:
Assumption 2. The set M is nonempty and compact.
2. Optimistic bilevel programming
If the follower agrees to cooperate with the leader in the decision making process,
we get the weak bilevel programming problem:
(2.1)
ϕ◦ (x) −→ min
x
s.t. G(x) ≤ 0
where
(2.2)
ϕ◦ (x) := min{F (x, y) : y ∈ Ψ(x)}
y
As can be seen from the formulation, this problem is weaker that the original bilevel
programming problem.
FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM
3
Definition 2.1. [4] A point (x∗ , y ∗ ) ∈ Rn × Rm is a local solution of problem (2.1),
(2.2) if x∗ ∈ {x : G(x) ≤ 0}, y ∗ ∈ Ψ(x∗ ) with
F (x∗ , y) ≥ F (x∗ , y ∗ ), ∀y ∈ Ψ(x∗ )
and there exists a sufficiently small positive number > 0, with
ϕ◦ (x∗ ) ≤ ϕ◦ (x), ∀x ∈ B (x∗ ).
If = ∞ then (x∗ , y ∗ ) is a global solution.
Now we further require to replace the optimistic bilevel programming problem
with another version of hierarchical problem. Consider the following problem
(2.3)
min{F (x, y) : G(x) ≤ 0, y ∈ Ψ(x)}
x,y
with the properties F, g, h explained at the beginning of this paper. The relation
between problem (2.3) and (2.1), (2.2) follows. If Assumption 2 and MFCQ satisfied, the global optimal solutions of (2.3) and (2.1), (2.2) coincide [4]. With respect
to a local optimal solution, Dutta and Dempe [6] showed that every local optimal
solution of the optimistic bilevel problem is a local optimal solution of problem
(2.3) as the following result shows.
Theorem 2.2. [6] Let the lower level problem (1.2) satisfy Assumption 2 and
MFCQ and let x̄ be a local optimal solution of problem (2.1) with corresponding
ȳ ∈ arg min{F (x̄, y) : y ∈ Ψ(x̄)}.
y
Then (x̄, ȳ) is a local optimal solution of problem (2.3).
Problem (2.3) is used to derive the necessary optimality condition for the optimistic bilevel programming problem (2.1), (2.2) because it can be reformulated to
a single level problem provided that the lower level problem is convex and satisfies
the Mangasarian-Fromowitz constraint qualification.
3. Piecewise continuously differentiable functions
Definition 3.1. A function y : Rn → Rm is a P C 1 function at x◦ if it is
continuous and there exists an open neighborhood V of x◦ and a finite number of continuously differentiable functions yi : V → Rm , i = 1, . . . , b such that
y(x) ∈ {y1 (x), . . . , yb (x)} for all x ∈ V . The function y is a P C 1 function on some
open set O provided it is a P C 1 function at every point x◦ ∈ O.
Theorem 3.2. [18] Let the lower level problem (1.2) at x = x◦ be convex satisfying (MFCQ), (SSOC) and (CRCQ) at a stationary solution y◦ . Then, the locally
uniquely determined function y(x) ∈ Ψ(x), and hence F (x, y(x)) is a P C 1 function.
It has been shown in [11] that P C 1 functions are locally Lipschitz continuous.
The directional differentiability of the solution function y(x) and hence F (x, y(x))
is by product of the result in [16]. The following result shows that P C 1 functions
are directional differentiable and locally Lipschitz continuous.
Theorem 3.3. [4, Theorem 4.8] Let f˜ : Rn → R be a P C 1 function. Then f˜ is
directionally differentiable and locally Lipschitz continuous.
4
AYALEW GETACHEW MERSHA AND STEPHAN DEMPE
Define
K(i) := {x : y(x) = yi (x)}
and the index set of essentially active functions
Iys (x) := {i : y(x) = yi (x), x ∈ cl int K(i)} ⊆ {1, . . . , b}.
The usual tangent cone to the set K(i) is defined by
xk − x
TK(i) (x) := d : ∃xk → x, xk ∈ K(i) ∀k, tk & 0 such that
→d .
tk
Once the set of essentially active selection function is known, it is shown in Scholtes
[20, Proposition A.4.1] that the generalized Jacobian of y(·) is given by
∂ ◦ y(x) = conv{∇yi (x) : i ∈ Iys (x)}.
The generalized directional derivative of solution function y(·) is given by
y ◦ (x; d) = sup{hd, zi : z ∈ ∂ ◦ y(x)}.
4. Optimality
∗
∗
It is known that a point (x , y ) is said to be a local optimal solution for a given
optimization problem if it is the best point in an small neighborhood of (x∗ , y ∗ ).
Definition 4.1. [4]: A feasible point (x∗ , y ∗ ) is said to be a local optimal solution
of problem (2.3) if there exists an open neighborhood U (x∗ , y ∗ ) such that
F (x∗ , y ∗ ) ≤ F (x, y), ∀(x, y) ∈ U (x∗ , y ∗ ), y ∈ Ψ(x), G(x) ≤ 0.
Remember that the validity of conditions MFCQ, SSOC and CRCQ in the lower
level problem implies that the solution function y(x) is a P C 1 function: a continuous
selection of continuously differentiable functions y1 , . . . , yb . The so called Clarke and
Bouligand stationarity are defined as:
Definition 4.2. [20] Let the lower level problem satisfy MFCQ, SSOC and CRCQ
at a point (x◦ , y◦ ), y◦ ∈ Ψ(x◦ ). Such a point (x◦ , y◦ ) is said to be a Clarke stationary
point for the optimistic bilevel programming problem if ∀d, ∃i ∈ Iys (x◦ ) such that
d ∈ TK(i) (x◦ ),
∇x F (x◦ , y◦ )d + ∇y F (x◦ , y◦ )∇yi (x◦ )d ≥ 0.
Definition 4.3. Let conditions MFCQ, SSOC and CRCQ be satisfied for the lower
level problem under consideration. A point (x◦ , y◦ ) is said to be a Bouligand
stationary point for problem (1.1), (1.2) if
(4.1)
∇x F (x◦ , y◦ )d + ∇y F (x◦ , y◦ )y 0 (x◦ ; d) ≥ 0
∀d : d> ∇Gi (x◦ ) ≤ 0, ∀i : Gi (x◦ ) = 0.
Or in other words ∀j ∈ Iys (x◦ ) we have
(4.2)
∇x F (x◦ , y◦ )d + ∇y F (x◦ , y◦ )∇yj (x◦ )d ≥ 0
∀d : d> ∇Gi (x◦ ) ≤ 0, ∀i : Gi (x◦ ),
where Iys (x◦ ) is the set of essentially active indices.
FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM
5
If the lower level problem (1.2) is convex satisfying MFCQ, it can be replaced
by its sufficient and necessary conditions and hence we get the following problem:
F (x, y) −→
(4.3)
min
x,y,λ
G(x)
≤
0
∇y L(x, y, λ)
=
0
g(x, y)
≤
0
λ
≥
0
λi gi (x, y)
=
0, i = 1, . . . , p.
Problems (1.1), (1.2) and problem (4.3) are equivalent if the lower level problem
is convex satisfying MFCQ at each feasible point, the solution function is uniquely
determined and strongly stable in the sense of Kojima [12], and a global solution is
considered. Without the convexity assumption problem (4.3) has a larger feasible
set. Hence even the global solution of the aforementioned problems may not coincide. The following simple relation can be established between problem (1.1),(1.2)
and problem (4.3).
Theorem 4.4. Let (x̄, ȳ, λ̄) be a feasible point of (4.3). If the lower level problem
(1.2) is convex satisfying MFCQ then the primal part, (x̄, ȳ) is feasible in (1.1).
Proof. Let (x̄, ȳ, λ̄) be a feasible point of problem (4.3). Then for x̄, G(x̄) ≤ 0,
validity of MFCQ in the lower level problem, the pair (ȳ, λ̄) solves the system
∇y L(x̄, y, λ) = 0
g(x̄, y) ≤ 0
λ≥0
λi gi (x̄, y) = 0, i = 1, . . . , p.
Due to the convexity of the parametric lower level problem the above system is both
sufficient and necessary condition for ȳ ∈ Ψ(x̄). Hence (x̄, ȳ) is a feasible point for
problem (1.1).
Problem (4.3) is one element of special classes of mathematical programming
problems called Mathematical Programs with Equilibrium Constraints (MPEC).
MPECs are special classes of nonlinear programming problems with equilibrium
conditions in the constraints. This means an expression of this type:
θ1 (x) ≥ 0, θ2 (x) ≥ 0, θ1i (x)θ2i (x) = 0, i = 1, . . . , m
appear in the constraints, where θk : Rn → Rm , k = 1, 2 are differentiable functions.
MPECs are considered as one of the ill conditioned problems in optimization theory
due to the existence of the complementarity term. Neither known regularity conditions like MFCQ, LICQ and Slater Constraint Qualification nor their nonsmooth
variants are satisfied at any point of the feasible set of MPEC [19]. MPECs are used
to model many economics and mechanic applications[17]. There are a number of
investigations done during the last couples of decades. Consequently, many results
were reported. The notable contribution we need to mention here is the successful
attempt to solve MPEC as an ordinary nonlinear programming problem [7, 21].
Also a number of papers address question of deriving optimality conditions (see for
instance [19]). The interested reader can consult the monographs [14, 17].
6
AYALEW GETACHEW MERSHA AND STEPHAN DEMPE
Scholtes [21] investigated a general version of problem (4.3) via a regularization
method. He relaxed the complementary constraint and then solved the resulting
parameterized nonlinear programming problem for decreasing sequence of the relaxation parameter. The validity of MFCQ in the lower level problem implies the
multiplier set is a nonempty convex compact polyhedron[10]. However, it is not
necessarily a singleton.
To obtain a stationary solution of an optimistic bilevel programming problem with
convex lower level problem, we have to show the nonexistence of descent directions
at the current incumbent stationary solution of the corresponding MPEC. As it
is clearly shown in [4, Theorem 5.15], a stationary solution for the corresponding
MPEC may not be a local solution for the optimistic bilevel programming problem
even under strong assumptions: convexity and validity of regularity conditions for
the lower level problem. We approach the ill-conditioned problem (4.3) by relaxing the stationarity condition ∇y L(x, y, λ) = 0 and the complementarity condition
λi gi (x, y) = 0, i = 1, 2, ..p.
Consequently, we consider the following parametric auxiliary problem NLP() for
> 0:
F (x, y) −→
(4.5)
min
x,y,λ
G(x)
k∇y L(x, y, λ)k∞
≤
≤
0
g(x, y)
≤
0
λ
≥
0
−λi gi (x, y)
≤
, i = 1, . . . , p
Before we proceed, we need to establish the relation between problem (1.1) and
problem (4.5).
Theorem 4.5. Let the parametric lower level problem (1.1) be convex satisfying MFCQ. If (x∗ , y ∗ ) is a local optimal solution for problem (1.1), (1.2) then
(x∗ , y ∗ , λ∗ ) is local optimal solution for problem (4.5) for = 0 and for all
λ∗ ∈ Λ(x∗ , y ∗ ) := {λ : λ ≥ 0, λT g(x∗ , y ∗ ) = 0, ∇y L(x∗ , y ∗ , λ) = 0}.
Proof. Assume that there exists λ̄ ∈ Λ(x∗ , y ∗ ) such that (x∗ , y ∗ , λ̄) is not a local optimal solution of problem (4.5) with = 0. Then there exists a sequence
{(xk , yk , λk )} converging to (x∗ , y ∗ , λ) with
F (xk , yk ) < F (x∗ , y ∗ )
G(xk ) ≤ 0
(4.6)
g(xk , yk ) ≤ 0
∇y L(xk , yk , λk ) = 0
λk ∈ Λ(xk , yk ).
Since the KKT conditions are necessary and sufficient for problem (1.2), we have
yk ∈ Ψ(xk ), ∀k. This implies (xk , yk ) is feasible in (1.1), (1.2). Therefore (x∗ , y ∗ ) is
not a local optimal solution of problem (1.1), (1.2) which is a contradiction to the
hypothesis of the theorem.
We make the following assumption for the upper level constraint.
FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM
7
Assumption 3. The upper level constraint qualification is said to be satisfied at a
point x if
{d : d> ∇Gi (x) < 0, ∀i : Gi (x) = 0} =
6 ∅
The existence of descent direction at a point in the neighborhood of a nonstationary point is shown in the next theorem.
Theorem 4.6. Let (x̄, ȳ, λ̄) be a feasible solution of problem (4.5) for = 0. Let the
lower level problem (1.2) be convex satisfying MFCQ, SSOC and CRCQ. Assume
that the upper level constraint set satisfies Assumption 3. If (x̄, ȳ) is not a Bouligand
stationary point of the problem (1.1) then there exists d with d> ∇x Gi (x̄) < 0, ∀i :
Gi (x̄) = 0 and δ > 0 such that
∇x F (x, y)d + ∇y F (x, y)∇yj (x)d < 0, ∀(x, y) ∈ Bδ (x̄, ȳ)
for some j = 1, . . . , w.
Proof. Let the lower level problem be convex satisfying MFCQ, SSOC and CRCQ.
Let (x̄, ȳ, λ̄) be a feasible solution of problem (4.5) for = 0. If (x̄, ȳ) is not a
Bouligand stationary point of the problem (1.1) then there exists a direction d
satisfying
∇Gi (x̄)> d ≤ 0, i : Gi (x̄) = 0
and
∇x F (x̄, ȳ)d + ∇y F (x̄, ȳ)y 0 (x̄; d) < 0.
Due to MFCQ, SSOC and CRCQ the function y(·) ∈ {y1 (·), . . . , yw (·)} is a P C 1
function and hence directionally differentiable (Theorem 3.3). Therefore, there is
j ∈ {1, . . . , w} such that
y 0 (x̄; d) = ∇yj (x̄)d.
Hence for this j we have
∇x F (x̄, ȳ)d + ∇y F (x̄, ȳ)∇yj (x̄)d < 0.
Since yj (·) is continuously differentiable there exists δ > 0 such that
∇x F (x, y)d + ∇y F (x, y)∇yj (x)d < 0, ∀(x, y) ∈ Bδ (x̄, ȳ).
The following theorem is one of the key results that relates the MPEC (Problem
4.3) and its relaxed nonlinear programming problem NLP():
Theorem 4.7. Let {(xk , yk , λk )} be a local optimal point of problem (4.5) for = k
where k → 0. Let (x∗ , y ∗ , λ∗ ) be a limit point of the sequence {(xk , yk , λk )}. If
∇Gi (x∗ )d < 0, i : Gi (x∗ ) = 0,
(4.7)
(d, r) 6= ∅
∇gj (x∗ , y ∗ )(d, r) < 0, j : gj (x∗ , y ∗ ) = 0
holds, then (x∗ , y ∗ , λ∗ ) is a Bouligand stationary solution for problem (4.5) for
= 0.
8
AYALEW GETACHEW MERSHA AND STEPHAN DEMPE
Proof. Let (x∗ , y ∗ , λ∗ ) be not a Bouligand stationary solution for problem (4.5) for
= 0. Then by definition the system
∇x F (x∗ , y ∗ )d + ∇y F (x∗ , y ∗ )r < 0
∇x Gi (x∗ )d ≤ 0, ∀i : Gi (x∗ ) = 0
k∇(∇y L(x∗ , y ∗ , λ∗ ))(d, r, γ)k∞ = 0
∇x gj (x∗ , y ∗ )d + ∇y gj (x∗ , y ∗ )r = 0, ∀j : gj (x∗ , y ∗ ) = 0
γi ≥ 0, i : λ∗i = 0
∇(λ∗i gi (x∗ , y ∗ ))(d, r, γ) = 0
has solution (d, r, γ). Let > 0 be arbitrary small. Due to the assumption (4.7)
ˆ r̂, γ̂)
there is (d,
∇x F (x∗ , y ∗ )dˆ + ∇y F (x∗ , y ∗ )r̂ < 0
Gi (x∗ ) + ∇x Gi (x∗ )dˆ < 0, ∀i
ˆ r̂, γ̂)k∞ ≤ /2
k∇(∇y L(x∗ , y ∗ , λ∗ ))(d,
gj (x∗ , y ∗ ) + ∇x gj (x∗ , y ∗ )dˆ + ∇y gj (x∗ , y ∗ )r̂ < 0, ∀j : gj (x∗ , y ∗ ) = 0
λ∗i + γ̂i > 0, ∀i
ˆ r̂, γ̂) ≤ /2
−∇(λ∗i gi (x∗ , y ∗ ))(d,
ˆ r̂)k sufficiently small. Since {(xk , yk , λk )} converges to (x∗ , y ∗ , λ∗ )
where k(d, r)−(d,
there exist k > 0 such that the following holds for k ≥ k
∇x F (xk , yk )dˆ + ∇y F (xk , yk )r̂ < 0
Gi (xk ) + ∇x Gi (xk )dˆ < 0, ∀i
ˆ r̂, γ̂)k∞ ≤ /2
k∇(∇y L(xk , yk , λk ))(d,
gj (xk , yk ) + ∇x gj (xk , yk )dˆ + ∇y gj (xk , yk )r̂ < 0, ∀j : gj (x∗ , y ∗ ) = 0
λ∗i + γ̂i > 0, ∀i
ˆ r̂, γ̂) ≤ /2
−∇((λi )k gi (xk , yk ))(d,
which implies (xk , yk , λk ) is not a local solution of problem (4.5) resulting in a
contradiction.
The converse of the above theorem may not be true as the following simple
example shows.
Example 4.8. [21]: Consider the following simple example.
min −y
s.t xy
≤
x, y
≥
0
Clearly, every point on the positive x-axis is a local minimizer of N LP (0). But the
relaxed problem N LP (), > 0 has no local solution.
It is clear that the feasible set of N LP () does not expand as → 0. Hence the
following theorem follows:
Theorem 4.9. Let {(xk , yk , λk )} be a sequence of local optimal solutions of problem
(4.5) corresponding to k . If k → 0 then there is a monotone increasing subsequence
of the sequence {F (xk , yk )}.
FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM
9
Proof. If the sequence {k } converges to zero then there is a subsequence {ki } of
{k} such that ki > ki+1 , ∀i. which implies S(ki ) ⊆ S(ki+1 ), ∀i where S(k ) is the
feasible set of problem (4.5) corresponding to k . The theorem follows from the
property of min operator.
Those presented results are the base of convergence analysis of the algorithm
given next. The verification of optimality or stationarity is based on the size of
Initialization: Initialize with (x◦ , y◦ ) ∈ {(x, y) : G(x) ≤ 0, g(x, y) ≤ 0},
k := 0, s∗ < w∗ < 0, δ, σ, α∗ ∈ (0, 1)
Step 1: Choose a maximal set of indices Ik ⊆ {i : gi (xk , yk ) = 0} with
{∇y gi (xk , yk ) = 0, i ∈ Ik } is linearly independent. Go to Step 2.
Step 2: Solve the following system of inequalities to find a multiplier:
kLy (xk , yk , λ)k∞ ≤ −λi gi (xk , yk ) ≤ , ∀j = 1, . . . , p
(4.8)
λ ≥ 0, λj = 0, j ∈
/ Ik .
Step 3 : Find direction a (dk , rk ) using the direction finding problem (4.11)
(see the next page) with sk < s∗ . If there is no solution goto Step 6
otherwise proceed to Step 4.
Step 4 : Using Armijo rule, find a step size which satisfies the following:
F (xk + tdk , yk + trk ) ≤ F (xk , yk ) + δtsk
Gi (xk + tdk ) ≤ 0, i = 1, . . . , l
(4.9)
gj (xk + tdk , yk + trk ) ≤ 0, j = 1, . . . , p.
Step 5: Set
xk+1 = xk + tk dk , yk+1 = yk + tk rk ,
put k:=k+1 and goto step 1.
Step 6: Put s∗ := s∗ /2. If s∗ ≤ w goto Step 3 otherwise goto Step 7.
Step 7: Choose new set Ik (σ) ⊆ {i : −σ ≤ gi (xk , yk ) = 0} with
{∇y gi (xk , yk ), i ∈ Ik (σ)} is linearly independent. If all such Ik (σ) are already tried then go to Step 9 otherwise go to step 8.
Step 8: Solve the following optimization problem
(4.10)
k∇y L(xk , yk , λ)k2 → min
λ
λi = 0, ∀i ∈
/ Ik (σ)
goto Step 3.
Step 9: Put := α. If ≤ ∗ then stop other wise goto Step 2.
Algorithm 4.1: Feasible direction algorithm applied to bilevel programming problem with convex lower level problem.
the directional derivative. If the directional derivative is sufficiently small (in the
absolute value), we compute an approximate Lagrange multiplier by considering
the “near” active constraints in the lower level problem. The solution of problem
(4.10) is uniquely determined and converges to the true multiplier λ∗ ∈ Λ(x∗ , y ∗ )
for the sequence {(xk , yk )} produced by the algorithm with limit point (x∗ , y ∗ ), if
Ik (σ) ∈ {I : {j : λ∗j > 0} ⊆ I ⊆ Ig (x∗ , y ∗ ) with {∇y gi (xk , yk ) : i ∈ Ik (σ)} is linearly
independent. The main purpose of doing Step 7 is to exclude Clarke stationary
10
AYALEW GETACHEW MERSHA AND STEPHAN DEMPE
points that are not Bouligand stationary points. However, the choice of σ in Step
7 seems to be difficult. The direction in Step 3 of Algorithm 4.1 is found by solving
the following linear optimization problem:
s −→ min
(4.11a)
d,r,s,γ
∇x F (xk , yk )d + ∇y F (xk , yk )r ≤ s
(4.11b)
∇x Gi (xk )d + Gi (x) ≤ s, i = 1, ..., l
(4.11c)
(4.11d)
∇y L(xk , yk , λk ) +
∇2xy L(xk , yk , λk )d
+
∇2yy L(xk , yk , λk )r
+ γ T ∇y g(xk , yk ) ≤ (4.11e) −∇y L(xk , yk , λk ) − ∇2xy L(xk , yk , λk )d − ∇2yy L(xk , yk , λk )r − γ T ∇y g(xk , yk ) ≤ (4.11f)
−λT g(xk , yk ) − γ T g(xk , yk ) − λT ∇x g(xk , yk )d − λT ∇y g(xk , yk )r ≤ (4.11g)
∇x gi (xk , yk )d + ∇y gi (xk , yk )r = 0, i ∈ Ik
(4.11h)
gj (xk , yk ) + ∇x gj (xk , yk )d + ∇y gj (xk , yk )r ≤ 0, j ∈
/ Ik
(4.11i)
−λ − γ ≤ 0
(4.11j)
kdk∞ ≤ 1
(4.11k)
krk∞ ≤ 1
(4.11l)
γ ≥ 0.
Problem (4.11) has a negative optimal objective function value at all non stationary
terms of the feasible sequence {(xk , yk )} provided that a suitable index set Ik is
chosen in Step 2 of the algorithm.
Theorem 4.10. Let the lower level problem be convex satisfying MFCQ, CRCQ
and SSOC. Let (x̄, ȳ, λ̄) be a feasible point of problem (4.3). If
(x, y) ∈ Bδ (x̄, ȳ) ∩ {(x, y) : y ∈ Ψ(x)}, δ > 0,
then there exists > 0 and I ⊆ {i : gi (x̄, ȳ) = 0}, with {∇y gi (x̄, ȳ) : i ∈ I} linearly
independent such that the following system is satisfied:
k∇y L(x, y, λ̄)k∞ ≤ (4.12)
λ̄ ≥ 0, λ̄i = 0, ∀i ∈
/ I.
Proof. Since the lower level problem satisfies MFCQ, the Λ(x̄, ȳ) is nonempty and
bounded for ȳ ∈ Ψ(x̄). Moreover, ȳ ∈ Ψ(x̄) and convexity implies that
k∇y L(x̄, ȳ, λ̄)k∞ = 0
(4.13)
λ̄> g(x̄, ȳ) = 0
λ̄ ≥ 0
for λ̄ ∈ Λ(x̄, ȳ). From this there exists an index set I such that
k∇y L(x̄, ȳ, λ̄)k∞ = 0
λ̄i = 0, ∀i ∈
/I
with {i : λi > 0} ⊂ I ⊂ Ig (x̄, ȳ). Consider the sequence
{(xk , yk )} ⊆ Bδ (x̄, ȳ) ∩ {(x, y) : y ∈ Ψ(x)}
FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM
11
with {(xk , yk )} converging to (x̄, ȳ). Fix such an index set I. Then we have
X
k∇y L(xk , yk , λ̄)k∞ = k∇y f (xk , yk ) +
λ̄i ∇y gi (xk , yk )k∞
i∈I
=
k∇y f (x̄, ȳ) +
X
λ̄i ∇y gi (x̄, ȳ)
i∈I
+∇(∇y f (x̄, ȳ) +
X
λ̄i ∇y gi (x̄, ȳ)((xk , yk )> − (x̄, ȳ)> )k∞
i∈I
≤
+ok((xk , yk )> − (x̄, ȳ)> )k∞
X
k∇y f (x̄, ȳ) +
λ̄i ∇y gi (x̄, ȳ)k∞
i∈I
{z
|
}
=0
+k∇(∇y f (x̄, ȳ) +
X
λ̄i ∇y gi (x̄, ȳ))((xk , yk )> − (x̄, ȳ)> )k∞
i∈I
>
+ok((xk , yk ) − (x̄, ȳ)> )k∞ .
Put
= k∇(∇y f (x̄, ȳ) +
X
λ̄i gi (x̄, ȳ)((xk , yk )> − (x̄, ȳ)> )k∞
i∈I
+ok((xk , yk )> − (x̄, ȳ)> )k∞ .
The following result is adapted from a result reported in [5]. It shows that
the sequence of step size parameters is bounded away from zero if the sequence of
optimal function values of problem (4.11) converges to a non zero number.
Theorem 4.11. Let the lower level problem be convex satisfying MFCQ. Suppose
{xk , yk , λk , dk , rk , k , sk , tk , Ik } be a sequence produced by the algorithm with k → 0.
If the sequence {xk , yk , λk } converges to (x∗ , y ∗ , λ∗ ) and limk→∞ sk = s < 0 then
limk→∞ tk > 0.
Proof. Assume that aforementioned positive step size converges to zero. Then, the
sequence of maximal possible step sizes τk must tend to zero too. Define τk by
τk = min{max{t : Gi (xk + tdk ) ≤ 0}, max{t : gj (xk + tdk , yk + tdk ) ≤ 0}}.
Then there exists
i : Gi (xk ) = Gi (xk + τk dk ) = 0
or
(4.14)
j : gj (xk , yk ) = gj (xk + τk dk , yk + τk dk ) = 0.
If the sequence {xk }, {dk } with kdk k = 1 converges respectively to x̄ and d¯ then the
¯ Observe that for the above
sequence { Gi (xk +τkτdkk )−Gi (xk ) } converges to ∇Gi (x̄)d.
taken i we have
implies
Gi (xk +τk dk )−Gi (xk )
τk
= 0. But ∇Gi (xk )dk ≤ sk for Gi (x̄) = 0. This
Gi (xk + τk dk ) − Gi (xk )
= ∇Gi (x̄)d¯ ≤ s < 0
k→∞,τk ↓0
τk
which is a contradiction. The other case, namely for condition (4.14) can be shown
analogously.
0=
lim
12
AYALEW GETACHEW MERSHA AND STEPHAN DEMPE
The next result shows that the directional derivative converges to zero.
Theorem 4.12. Let F be bounded below, the lower level problem be convex satisfying MFCQ, SSOC, CRCQ and the upper level constraint satisfy condition {d :
d> ∇Gi (x) < 0} 6= ∅. Let {xk , yk , λk , dk , rk , k , sk , tk , Ik } be a sequence computed
by the algorithm where k → 0. Then the sequence {sk } has zero as the only accumulation point.
Proof. Suppose there exist a subsequence that does not converge to zero. i.e it
converges to a number say s < 0. Assume without loss of generality that such a
subsequence be {sk }. From Theorem 4.11, we have limk→0 tk > 0. But from Step
4 of the algorithm we have
F (xk + tk dk , yk + tk dk ) − F (xk , yk ) ≤ δsk tk , ∀k
which implies F (xk , yk ) → −∞ as k → ∞. This is a contradiction to the fact that
F is bounded below.
Assume that λ̄ is obtained from Step 2 of the proposed algorithm. Let (x̄, ȳ, λ̄)
be a solution of (4.3) and let s = 0 be the corresponding optimal objective function
value of problem (4.11). Then we have the following result.
Theorem 4.13. Let (x, y) be a point satisfying G(x) ≤ 0, y ∈ Ψ(x). Let the
feasible set of the lower level problem satisfy MFCQ, SSOC, and CRCQ. Moreover,
let the optimal objective function value of problem (4.11) with = 0 be zero and λ
be obtained by solving (4.8). Then (x, y) is a Clarke stationary point for problem
(2.3).
Proof. From y ∈ Ψ(x) we have Λ(x, y) 6= ∅. That means the system
T
λ g(x, y) = 0
∇y L(x, y, λ) = 0
λ≥0
has a solution. Moreover, if = 0 then from (4.11d) we have
(4.15)
∇2yy L(x, y, λ)r + γ T ∇y g(x, y) = −∇2xy L(x, y, λ)d.
Also λi > 0 implies gi (x, y) = 0. In this case (4.11f) gives us
λi ∇x gi (x, y)d + λi ∇y gi (x, y)r ≥ 0
and hence we get
∇x gi (x, y)d + ∇y gi (x, y)r ≥ 0
for = 0. On the other hand from (4.11h) and since the optimal objective function
value is zero we have that gi (x, y) = 0 implies ∇x gi (x, y)d + ∇y gi (x, y)r ≤ 0.
Therefore
(4.16)
∇x gi (x, y)d = −∇y gi (x, y)r
∀i : λi > 0. From (4.15) and (4.16) we derive
λ ∈ arg max{∇x L(x, y, λ)d : λ ∈ Λ(x, y)}
and r = y 0 (x; d). Again from gi (x, y) < 0 we have λi = 0. This means, from
(4.11i) we get γi ≥ 0. But in (4.11f) = 0 results in −γi gi (x, y) ≤ 0 giving γi ≤ 0.
FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM
13
Therefore γi = 0. Since ∀d, Gi (x) + ∇x Gi (x)d < 0 there is λ ∈ Λ(x, y) such that
r = y 0 (x; d) is computed by solving (4.15), (4.16) such that the system
∇x F (x, y)d + ∇y F (x, y)y 0 (x; d) < 0
Gi (x) + ∇x Gi (x)d < 0
has no solution. Or in other words ∀d, Gi (x) + ∇x Gi (x)d < 0, i : 1, . . . , l, we have
∇x F (x, y)d + ∇y F (x, y)y 0 (x; d) ≥ 0.
Hence (x, y) is a Clarke stationary point (2.3) in the sense of Definition 4.2.
Remark 4.14. In the above theorem one can not guarantee that (x̄, ȳ) is a Bouligand
stationary point because the computed direction depends on the fixed multiplier. If
the multiplier changes we can not guarantee the nonexistence of descent direction.
Theorem 4.15. Let k → 0, {xk , yk , λk , γk , sk } be a solution of problem (4.5) for
= k . Let {xk , yk , λk , γk , sk } converge to (x∗ , y ∗ , λ∗ , γ ∗ , 0). If (x∗ , y ∗ ) satisfies
SSOC, MFCQ, CRCQ and there exists
d ∈ {d : ∇x Gi (x∗ )d < 0, ∀i : Gi (x∗ ) = 0}
then (x∗ , y ∗ ) is a Bouligand stationary point of the optimistic bilevel programming
problem (2.3).
Proof. Let (x∗ , y ∗ ) be not a Bouligand stationary point. Then there must exist
d∗ ∈ {d : ∇x Gi (x∗ )d < 0, ∀i : Gi (x∗ ) = 0}
with
∇x F (x∗ , y ∗ )d∗ + ∇y F (x∗ , y ∗ )y 0 (x∗ ; d∗ ) < 0.
Since the lower level problem satisfies MFCQ, SSOC, and CRCQ then there is
λ∗ ∈ Λ(x∗ , y ∗ ) with (r∗ , γ ∗ ), r∗ = y 0 (x∗ ; d∗ ) is the solution of the following system
of linear equations and inequalities:
2
∇ L(x∗ , y ∗ , λ∗ )r + γ T ∇y g(x∗ , y ∗ ) + ∇2xy L(x∗ , y ∗ , λ∗ )d∗ = 0
yy ∗ ∗ ∗
∇x gi (x , y )d + ∇y gi (x∗ , y ∗ )r = 0, ∀i : λ∗i > 0
∇ g (x∗ , y ∗ )d∗ + ∇y gi (x∗ , y ∗ )r ≤ 0, ∀i : gi (x∗ , y ∗ ) = λ∗ = 0
x i
γi ≥ 0, γi (∇x gi (x∗ , y ∗ )d∗ + ∇y gi (x∗ , y ∗ )r) = 0, ∀i : gi (x∗ , y ∗ ) = λ∗i = 0
Since the (x∗ , y ∗ ) satisfies MFCQ and from the definition of Λ(., .),
∇y L(x∗ , y ∗ , λ∗ ) = 0
∗
∗ ∗
λ∗ ≥ 0
λ ∈ Λ(x , y ) ⇒
∗T
λ g(x∗ , y ∗ ) = 0
which implies that the optimal objective function value of (4.11) for = 0 is s∗ < 0.
Since (xk , yk ) → (x∗ , y ∗ ) for → 0, > 0 then there exist k:
− ≤ ∇y L(xk , yk , λ∗ ) ≤
λ∗ ≥ 0
∀k ≥ k.
∗T
−λ g(xk , yk ) ≤ 14
AYALEW GETACHEW MERSHA AND STEPHAN DEMPE
Then the following is satisfied
∇x gi (xk , yk )d∗ + ∇y gi (xk , yk )r = 0, i ∈ Ik
∗
g
/ Ik
i (xk , yk ) + ∇x gi (xk , yk )d + ∇y gi (xk , yk )r ≤ s, ∀i ∈
− ≤ ∇2yy L(xk , yk , λ∗ ) + ∇2xy L(xk , yk , λ∗ )d∗ + γ ∗T ∇y gi (xk , yk ) ≤ − ≤ λi gi (xk , yk ) + γi∗ gi (xk , yk ) + λ∗i (∇y gi (xk , yk )d∗ + ∇y gi (xk , yk )r)
gi (xk , yk ) + γi∗ (∇x gi (xk , yk )d∗ + ∇y gi (xk , yk )r) ≥ 0
∗
γi ≥ 0, ∀i : λ∗i = 0, gi (xk , yk ) ≥ −
which implies that the point (d∗ , r, λ∗ , γ ∗ ) is feasible in problem (4.11). But this
means that optimal objective function value of (4.11) at the limit is s ≤ s∗ <
0, ∀k ≥ k which implies the sequence {sk } does not converge to zero. This is a
contradiction to the fact that sk → 0. Hence the limit (x∗ , y ∗ ) is B-stationary for
the bilevel programming problem(2.3).
5. Numerical Experience
In this section we consider the relaxed KKT reformulation N LP () to numerically verify our claim in the previous sections of this paper. In this section, we
present our experience with numerical experiments of the proposed algorithm in order to substantiate the theory we have established in previous sections. We present
two instances of numerical experiments. The first concerns verifying literature solution by directly solving problem (4.5) for a decreasing sequence of the relaxation
parameter . The second concerns the application of the feasible direction method.
The latter is purely the implementation of our algorithm. The numerical test is conducted on some exemplary test problems from Leyffer’s MacMPEC collection [13].
MacMPEC collection is a collection of mathematical programs with equilibrium
constraints coded in AMPL. We have recoded them in AMPL since we consider
the relaxed problem (4.5). The experiment was conducted on Medion MD 95800,
Intelr Pentiumr M Processor 730, 1.6 GHz with 512 MB RAM under Microsoft
Windows operating system. We used the optimization solvers SNOPT, DONLP2,
LOQO, and MINOS that can be downloaded along with AMPL student edition.
5.1. Relaxation method. We consider problem (4.5) and the solution approach
in Scholtes[21] solved by the same solver for a decreasing sequence of relaxation parameters . On most instances of test problems our relaxation method and Scholtes
regularization produce a sequence of iterates that converges to a local solution of
the corresponding optimistic bilevel programming reported in the literature with
the exception of problems ex9.1.3, ex9.1.7 and ex9.1.8. Our method produced
a sequence of iterates that converge to the local solution of the optimistic bilevel
programming ex9.1.7 reported in the original literature. However, with the same
initial point, the regularization approach in Scholtes[21] fails to converge to the
literature solution of the optimistic bilevel programming problem. We obtained a
better solution for problems ex9.1.3, ex9.1.8 and bard3. Table 1 gives information regarding the comparison of best function values. The first column shows
serial number of considered problems, the second column indicates the name of the
problem. The third, the forth and the fifth columns show respectively best function values from literature, using our approach and using Scholtes regularization
scheme [21]. The result seems to confirm the relevance of relaxing the stationarity
constraint of problem (4.5).
FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM
No
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
Name
bard1
bard3
ex9.1.1
ex9.1.2
ex9.1.3
ex9.1.4
ex9.1.5
ex9.1.6
ex9.1.7
ex9.1.8
ex9.1.9
ex9.2.1
ex9.2.2
ex9.2.3
ex9.2.4
ex9.2.5
ex9.2.6
ex9.2.7
ex9.2.8
ex9.2.9
TuMiPh
MD06
FLZl98
Table 1.
scheme.
15
Best function value obtained from
ours
Scholtes
literature approach approach
source comment
17
17
17
[1]
-12.6787 -15.8164 -15.8164
[1]
-13
-14
-14
[8]
-16
-16
-16
[8]
-29
-58
-58
[8]
-37
-37
-37
[8]
-1
-1
-1
[8]
-49
-49
-49
[8]
-26
-26
0
[8]
better solution
-1.75
-3.25
-3.25
[8]
3.111
3.111
3.111
[8]
17
17
17
[8]
100
100
100
[8]
5
5
5
[8]
.5
.5
.5
[8]
5
5
5
[8]
-1
-1
-1
[8]
17
17
17
[8]
1.5
1.5
1.5
[8]
2
2
2
[8]
22.5
22.5
22.5
[23]
-20
-20
-20
[15]
1
1
1
[9]
Comparison our approach with Scholtes regularization
5.2. Feasible direction method. Initially Algorithm 4.1 was implemented in
Matlab 7. It is known that Matlab’s programming facility made it suitable for such
an iterative method. However, Matlab’s optimization solvers performed poorly at
least for our problems. It was observed that a wrong multiplier is calculated from
problems (4.8) and (4.10). Having observed this limitation in Matlab solvers, the
algorithm was re-implemented in AMPL using AMPL’s script facility. The algorithm solves three problems alternately. The multiplier problem, the direction
finding problem and the choice of the step size parameter. The step size t is chosen
as the maximum of { 12 , 14 , 18 , . . .} such that the system
F (xk + tdk , yk + trk ) ≤ F (xk , yk ) + δtsk
Gi (xk + tdk ) ≤ 0, i = 1, . . . , l
gj (xk + tdk , yk + trk ) ≤ 0, j = 1, . . . , p
is satisfied. For the current incumbent primal point (xk , yk ) the algorithm first
computes the corresponding dual point λ either from linear inequalities (4.8) or
quadratic optimization problem (4.10) depending on the size of s at hand. Once the
appropriate triplets (xk , yk , λk ) found, the direction finding problem (4.11) is solved
by CPLEX, a linear programming solver, which can be downloaded from AMPL web
16
AYALEW GETACHEW MERSHA AND STEPHAN DEMPE
site, and problem (4.10) is solved by Spellucci’s DONLP2, a nonlinear programming
solver. In our AMPL implementation, this is done by alternatively solving different
problems. For the test problems listed in Table 1, literature solutions are obtained.
However, for some problems such as ex9.2.7 the solution slow convergence was
observed. This indicates that some researches and more experiments has to be
done in order to improve the performance of this algorithm. To better illustrate
this we consider the following example. This example is taken from the monograph
[8].
Example 5.1. Consider
min
x≥0,y≥0
(x − 5)2 + (2y + 1)2
where y solves
min (y − 1)2 − 1.5xy
y≥0
s.t.
(5.1)
−3x + y ≤ −3
x − .5y ≤ 4
x + y ≤ 7.
The convex lower level problem is replaced by the its KKT condition and then the
stationarity condition is relaxed. The feasible direction method is applied to the
problem
min (x − 5)2 + (2y + 1)2
s.t.
(5.2)
2(y − 1) − 1.5x + λ1 − .5λ2 + λ3 ≤ −2(y − 1) + 1.5x − λ1 + .5λ2 − λ3 ≤ −3x + y + 3 ≤ 0
x + 0.5y − 4 ≤ 0
x+y−7≤0
−λ1 (−3x + y + 3) ≤ −λ2 (x + 0.5y − 4) ≤ −λ3 (x + y − 7) ≤ FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM
17
The direction finding problem is given by
s −→ min
d,r,s,γ
2(x − 5)d + 4(2y + 1)r ≤ s
−x − d ≤ s
2(y − 1) − 1.5x + λ1 − .5λ2 + λ3 − λ4 − 1.5d + 2r + γ1 − 0.5γ2 + γ3 − γ4 ≤ −2(y − 1) + 1.5x − λ1 + .5λ2 − λ3 + λ4 + 1.5d − 2r − γ1 + 0.5γ2 − γ3 + γ4 ≤ = 0 if − 3x + y + 3 = 0
−3x + y + 3 − 3d + r
≤ s if − 3x + y + 3 6= 0
= 0 if x − 0.5y − 4 = 0
x − 0.5y − 4 + d − 0.5r
≤ s if x − 0.5y − 4 6= 0
= 0 if x + y − 7 = 0
x+y−7+d+r
≤ s if x + y − 7 6= 0
−λ1 (−3x + y + 3) − γ1 (−3x + y + 3) − λ1 (−3d + r) ≤ −λ2 (x − 0.5y − 4) − γ2 (x − 0.5y − 4) − λ2 (d − .5r) ≤ −λ3 (x + y − 7) − γ3 (x + y − 7) − λ3 (d + r) ≤ λ4 y + γ4 y + λ4 r ≤ −λ1 − γ1 − s ≤ 0
−λ2 − γ2 − s ≤ 0
−λ3 − γ3 − s ≤ 0
−λ4 − γ4 − s ≤ 0
For this particular problem the initial point is taken to be (x◦ , y◦ ) = (1.1, 0.3),
(λ1 , λ2 , λ3 , λ4 ) = (0, 0, 0, 0), σ = 10−1 , δ = 0.3, = 10−4 . The solution reported in
the literature is (1, 0)(see [8, Problem 9.3.8]). The first iteration is the computation
of the multiplier. Then the direction finding problem is solved using CPLEX.
The step size parameter is chosen in the sense of Armijo. These three alternating
phase of computation continues until the stopping criterion ( = 10−6 for this
problem) satisfied. For this particular problem, the algorithm converges to the
above mentioned literature solution after several iterations.
6. Conclusion
In this paper, a descent like algorithm based on the framework of a feasible
direction method is proposed to compute Bouligand stationary points of optimistic
bilevel programming problems with convex lower level problem. It has been shown
and numerically evidenced that in addition to relaxing complementarity conditions,
relaxing the stationarity conditions of the auxiliary MPEC problem (4.3) has a
significant advantage on computing the appropriate multiplier, and hence finding
Bouligand stationary points of the optimistic bilevel programming problems. The
algorithm is implemented for some bilevel programming problems. Its performance,
though in its infant stage, is found promising.
References
[1] J. F. Bard. Convex two-level optimization. Mathematical Programming, 40:15–27, 1988.
18
AYALEW GETACHEW MERSHA AND STEPHAN DEMPE
[2] J. F. Bard. Some properties of the bilevel programming problem. Journal of Optimization
Theory and Applications, 68:371–378, 1991.
[3] J. F. Bard. Practical Bilevel Optimization: Algorithms and Applications. Kluwer Academic
Publishers, Dordrecht, 1998.
[4] S. Dempe. Foundations of Bilevel programming. Kluwer Academic Publishers, Dordrecht et
al., 2002.
[5] S. Dempe and H. Schmidt. On an algorithm solving two-level programming problems with
nonunique lower level solutions. Computational Optimization and Applications, 6:227–249,
1996.
[6] J. Dutta and S. Dempe. Bilevel programming with convex lower level problem. In S. Dempe
and V. Kalashnikov, editors, Optimization with Multivalued Mappings: Theory, Applications and Algorithms, volume 2 of Springer Optimization and Its Applications, pages 51–71.
Springer US, 2006.
[7] R. Fletcher and S. Leyffer. Numerical experience with solving MPECs as NLPs. Technical
Report NA/210, Department of Mathematics, University of Dundee, 2002.
[8] C. A. Floudas, P. M. Pardalos, C. S. Adjiman, W. R. Esposito, Z. H. Gümüs, S. T. Harding,
J. L. Klepeis, C. A. Meyer, and C. A. Schweiger. Handbook of test problems in local and global
optimization. Number 33 in Nonconvex Optimization and Its Applications. Kluwer Academic
Publishers, Dordrecht, 1999.
[9] C. A. Floudas and S Zlobec. Multilevel Optimization: Algorithms and Applications, volume 20, chapter Optimality and duality in parametric convex lexographic programming,
pages 359–379. Kluwer Academic Publishers, Dordrecht, 1998.
[10] J. Gauvin. A necessary and sufficient regularity condition to have bounded multiplier in non
convex programming. Mathematical Programming, 12:136–138, 1977.
[11] W.W. Hager. Lipschitz continuity for constrained processes. SIAM Journal on Control and
Optimization, 17:321–269, 1979.
[12] M. Kojima. Strongly stable stationary solutions in nonlinear programs. In S.M. Robinson,
editor, Analysis and Computation of Fixed Points, pages 93–138. Academic Press, New York,
1980.
[13] S. Leyffer. MacMPEC ampl collection of Mathematical Programs with Equilibrium Constraints . Argonne National Laboratory, 2000. http://www-unix.mcs.anl.gov/~leyffer/
MacMPEC/.
[14] Z.-Q. Luo, J.-S. Pang, and D. Ralph. Mathematical Programs with Equilibrium Constraints.
Cambridge University Press, Cambridge, 1996.
[15] A.Getachew Mersha and S. Dempe. Linear bilevel programming with upper level constraints
depending on the lower level solution. Applied Mathematics and Computation, 180:247–254,
2006.
[16] R. Mifflin. Semismooth and semiconvex functions in constrained optimization. SIAM Journal
on Control and Optimization, 15:959–972, 1977.
[17] J. Outrata, M. Kočvara, and J. Zowe. Nonsmooth Approach to Optimization Problems with
Equilibrium Constraints. Kluwer Academic Publishers, Dordrecht, 1998.
[18] D. Ralph and S. Dempe. Directional derivatives of the solution of a parametric nonlinear
program. Mathematical Programming, 70:159–172, 1995.
[19] H. Scheel and S. Scholtes. Mathematical programs with equilibrium constraints: stationarity,
optimality, and sensitivity. Mathematics of Operations Research, 25:1–22, 2000.
[20] S. Scholtes. Introduction to piecewise differentaible equations. Preprint 53-1994, Universität
Karlsruhe, Karlsruhe, 1994.
[21] S. Scholtes. Convergence properties of a regularization scheme for mathematical programs
with complementarity constraints. SIAM Journal on Optimization, 11:918–936, 2001.
[22] D. M. Topkis and A.F. Veinott. On the convergence of of some feasible direction algorithms
for nonlinear programming. SIAM Journal on Control and Optimization, 5:268–279, 1967.
[23] H. Tuy, A. Migdalas, and N.T. Hoai-Phuong. A novel approach to bilevel nonlinear programming. Journal of Global Optimization, 38:527–554, 2007.
FEASIBLE DIRECTION METHOD FOR BILEVEL PROGRAMMING PROBLEM
19
Department of Mathematics and Computer Science, Technical University Bergakademie
Freiberg, Germany
E-mail address: [email protected]
Department of Mathematics and Computer Science, Technical University Bergakademie
Freiberg, Germany
E-mail address: [email protected]
© Copyright 2026 Paperzz