Improved Feasible SQP Algorithm for Nonlinear Programs with

1496
JOURNAL OF COMPUTERS, VOL. 8, NO. 6, JUNE 2013
Improved Feasible SQP Algorithm for Nonlinear
Programs with Equality Constrained SubProblems
Zhijun Luo1, Guohua Chen3 and Simei Luo4
Department of Mathematics & Applied Mathematics, Hunan University of Humanities, Science and Technology, Loudi,
China
Email: [email protected]
Zhibin Zhu2
School of Mathematics and Computing Science, Guilin University of Electronic Technology, Guilin, China
Email: [email protected]
Abstract—This paper proposed an improved feasible
sequential quadratic programming (FSQP) method for
nonlinear programs. As compared with the existing SQP
methods which required solving the QP sub-problem with
inequality constraints in single iteration, in order to obtain
the feasible direction, the method of this paper is only
necessary to solve an equality constrained quadratic
programming sub-problems. Combined the generalized
projection technique, a height-order correction direction is
yielded by explicit formulas, which can avoids Maratos
effect. Furthermore, under some mild assumptions, the
algorithm is globally convergent and its rate of convergence
is one-step superlinearly. Numerical results reported show
that the algorithm in this paper is effective.
Index Terms—Nonlinear programs, FSQP method, Equality
constrained quadratic programming, Global convergence,
Superlinear convergence rate
I. INTRODUCTION
Consider the following nonlinear programs
min f ( x)
s.t. g j ( x) ≤ 0, j ∈ I = {1, 2,
Where
f ( x), g j ( x) : R n → R ( j ∈ I ) are
, m},
(1)
continuously
differentiable functions. Denote the feasible set for (1) by
X = {x ∈ R n | g j ( x) ≤ 0, j ∈ I } .
The Lagrangian function associated with (1) is defined
as follows:
m
L ( x, λ ) = f ( x ) + ∑ λ j g j ( x )
j =1
A point x ∈ X is said to be a KKT point of (1), if it is
satisfies the equalities
This work was supported in part by the National Natural Science
Foundation (11061011) of China, and the Educational Reform Research
Fund of Hunan University of Humanities, Science and Technology
(NO.RKJGY1030), corresponding author, E-mail: [email protected] .
© 2013 ACADEMY PUBLISHER
doi:10.4304/jcp.8.6.1496-1503
m
∇f ( x) + ∑ λ j ∇g j ( x) = 0,
j =1
λ j g j ( x) = 0, j ∈ I ,
where λ = (λ1 , , λm )T is nonnegative, and λ is said to
be the corresponding KKT multiplier vector.
Method of Sequential Quadratic Programming (SQP)
is an important method for solving nonlinearly
constrained optimization [1, 2, 18]. It generates
iteratively the main search direction d 0 by solving the
following quadratic programming (QP) sub-problem:
1
min ∇f ( x)T d + d T Hd
2
s.t. g j ( x) + ∇g j ( x)T d ≤ 0, j ∈ I ,
(2)
where H ∈ R n× n is a symmetric positive definite matrix.
However, such type SQP algorithms have two serious
shortcomings:
1) SQP algorithms require that the relate QP subproblems (2) must be consistency;
2) There exists Matatos effect.
Many efforts have been made to overcome the
shortcomings through modifying the quadratic subproblem (2) and the direction d [4, 5, 7, 8]. Some
algorithms solve the problem (1) by using the idea of
filter method or trust-region [13, 16, 17].
For the problem (2), it is also a hot topic to solve the
QP problem like (2) in the field of optimization. By using
the idea of active constraints set, some algorithms solve
step by step a series of corresponding QP problems with
only equality constraints to obtain the optimum solution
to the QP sub-problem (2). P. Spellucci [6] proposed a
new method, the d 0 is obtained by solving QP subproblem with only equality constraints:
JOURNAL OF COMPUTERS, VOL. 8, NO. 6, JUNE 2013
1497
1
min ∇f ( x)T d + d T Hd
2
s.t. g j ( x) + ∇g j ( x)T d = 0, j ∈ A ⊆ I ,
complementarity. Furthermore, its global and superlinear
convergence rate is obtained under some suitable
conditions.
This paper is organized as follows: In Section II, we
state the algorithm; the well-defined of our approach is
also discussed, the accountability of which allows us to
present global convergence guarantees under common
conditions in Section III, while in Section IV we deal
with superlinear convergence. Finally, in Section V,
numerical experiments are implemented.
(3)
where the so-called working set A ⊆ I is suitably
determined. If d 0 = 0 and λ ≥ 0 ( λ is said to be the
corresponding KKT multiplier vector.), the algorithm
stops. The most advantage of these algorithms is merely
necessary to solve QP sub-problems with only equality
constraints. However, if d 0 = 0 , but λ < 0 , the algorithm
will not implement successfully. In [10], proposed an
SQP method for general constrained optimization. Firstly,
make use of the technique which handle the general
constrained optimization as an inequality parametric
programming, then, consider a new quadratic
programming with only equality constraints as follow:
1
min ∇f ( x)T d + d T Hd
2
s.t. g j ( x) + ∇g j ( x)T d = − min{0, π ( x)}, j ∈ J ( x).
Where π ( x) is a suitable vector, J ( x) is a suitable
approximate active set. But the QP problems may no
solution under some conditions. Recently, Zhu [14]
Consider the following QP sub-problem:
1
min ∇f ( x)T d + d T Hd
2
s.t. p j ( x) + ∇g j ( x)T d = 0, j ∈ L.
The active constraints set of (1) is denoted as follows:
I ( x) = { j ∈ I | g j ( x) = 0, j ∈ I }.
approximate active, which guarantees to hold that if
d 0 = 0 , then x is a KKT point of (1), i.e., if d 0 = 0 , then
it holds that λ ≥ 0 . Depended strictly on the strict
complementarity, which is rather strong and difficult for
testing, the superlinear convergence properties of the
SQP algorithm are obtained. For avoiding the superlinear
convergence depend strictly on the strict complementarity,
Another some SQP algorithms (see [15]) have been
proposed, however it is regretful that these algorithms are
infeasible SQP type and nonmonotone. In [16], a feasible
SQP algorithm is proposed. Using generalized projection
technique, the superlinear convergence properties are still
obtained under weaker conditions without the strict
complementarity.
We will develop an improved feasible SQP method for
solving optimization problems based on the one in [14].
The traditional FSQP algorithms, in order to prevent
iterates from leaving the feasible set, and avoid Maratos
effect, it needs to solve two or three QP sub-problems
like (2). In our algorithm, per single iteration, it is only
necessary to solve an equality constrained quadratic
programming, which is very similar to (4). Obviously, it
is simpler to solve the equality constrained QP problem
than to solve the QP problem with inequality constraints.
In order to void the Maratos effect, combined the
generalized projection technique, a height-order
correction direction is computed by an explicit formula,
and it plays an important role in avoiding the strict
(5)
Now, the following algorithm is proposed for solving
the problem (1).
Algorithm A:
Step 0 Initialization:
Given a starting point x 0 ∈ X , and an initial
symmetric positive definite matrix H 0 ∈ R n×n . Choose
1
parameters ε 0 ∈ (0,1), α ∈ (0, ), τ ∈ (2,3) . Set k = 0 ;
2
Step 1. Computation of an approximate active set J k .
Step 1.1. For the current point x k ∈ X , set i = 0,
ε i ( x k ) = ε 0 ∈ (0,1) .
(4)
where p j ( x ) is a suitable vector, L is a suitable
© 2013 ACADEMY PUBLISHER
II. DESCRIPTION OF ALGORITHM
Step
1.2.
If
det( Ai ( x k )T Ai ( x k )) ≥ ε i ( x k )
,
let
J k = J ( x ), Ak = A( x ), i ( x ) = i , and go to Step 2.
Otherwise go to Step 1.3, where
k
k
k
J i ( x k ) = { j ∈ I | −ε i ( x k ) ≤ g j ( x k ) ≤ 0},
Ai ( x k ) = (∇gi ( x k ), j ∈ J i ( x k )).
(6)
1
Step 1.3. Let i = i + 1, ε i ( x k ) = ε i −1 ( x k ) , and go to
2
Step1. 2.
Step 2. Computation of the vector d 0k .
Step 2.1
Bk = ( AkT Ak ) −1 AkT , v k = (v kj , j ∈ J k ) = − Bk ∇f ( x k ),
k
k
⎪⎧ −v j , v j < 0
p kj = ⎨
k
k
⎪⎩ g j ( x ), v j ≥ 0
p k = ( p kj , j ∈ J k ).
(7)
Step 2.2 Solve the following equality constrained QP
Sub problem at x k :
1
min ∇f ( x k )T d + d T H k d
2
k
k T
s.t. p j + ∇g j ( x ) d = 0, j ∈ J k .
Let
d 0k
be
the
KKT
point
of
(8)
(8),
and
b = (b , j ∈ J k ) be the corresponding multiplier vector.
k
k
j
If d 0k = 0 , STOP. Otherwise, CONTINUE;
1498
JOURNAL OF COMPUTERS, VOL. 8, NO. 6, JUNE 2013
Step 3
Computation of the feasible direction with descent d k :
~
J k ,i +1 ≡ J k ,i
Lk for i large enough. From (6) and (13),
with i → ∞ , we obtain
~
d k = d 0k − δ k Ak ( AkT Ak ) −1 ek
Where ek = (1,
δk =
L k = I ( x k ), det( AIT( xk ) AI ( xk ) ) = 0.
(9)
This is a contradiction to H 2.2, which shows that the
statement is true.
2) Suppose K is an infinite index set such that
{x k }k∈K → x* . We suppose that the conclusion is false,
,1)T ∈ R| J k | , and
|| d 0k || (d 0k )T H k d 0k
, μ k = −( AkT Ak ) −1 AkT ∇f ( x k )
2 | μ kT ek ||| d 0k || +1
i.e., there exists K ' ⊆ K (| K ' |= ∞) , such that
ε k ,i → 0, k ∈ K ' , k → ∞.
k
Step 4. Computation of the high-order revised direction
k
d :
d k = − Ak ( Ak T Ak )−1 (|| d0k ||τ ek + g Jk ( xk + d k )),
(10)
~
Let Lk = J k ,ik −1 . From the definition of ε k ,ik , it holds,
for k ∈ K ' , k large enough, that
~
det( AT~ AT~ ) = 0, − 2ε k ,ik ≤ g j ( x k ) ≤ 0, j ∈ Lk . (14)
where
Lk
g J k ( x k + d k ) = g J k ( x k + d k ) − g J k ( x k ) − ∇ g J ( x k )T d k .
Lk
Since there are only finitely many choices for sets
k
Step 5. Line search:
Compute tk , the first number t in the sequence
1 1 1
{1, , , ,...} satisfying
2 4 8
~
Lk ⊆ I
, it is sure that there exists K '' ⊆ K ' (| K '' |= ∞) ,
~
~
such that Lk ≡ L, (k ∈ K '' ) , for k
~
f ( x k + td k + t 2 d k ) ≤ f ( x k ) + α t∇f ( x k )T d k ,
g j ( x + td + t d ) ≤ 0, j ∈ I .
k
k
2
(11)
Denote A = {∇g j ( x* ) | j ∈ L} , then, let k ∈ K '' , k → ∞ ,
from (14), it holds that
~
Step 6. Update:
Obtain H k +1 by updating the positive definite matrix H
using
some
quasi-Newton
formulas.
Set
x k +1 = x k + tk d k + t 2 d k , and k = k + 1 . Go back to step 1.
~
~
det( AT A) = 0, g j ( x* ) = 0, j ∈ L ⊆ I ( x* ).
(12)
k
large enough.
~
This is a contradiction to H 2.2, too, which shows that
the statement is true.
k
Throughout this paper, following basic assumptions
are assumed.
H2.1 The feasible set X ≠ Φ , and functions f ( x),
g j ( x), j ∈ I are twice continuously differentiable.
H2.2 ∀ x ∈ X , the vectors {∇g j ( x), j ∈ I ( x)} are
linearly independent.
d 0k = 0 , then x is a KKT point of (1). If d 0k ≠ 0 , then
k
x k computed in step 4 is a feasible direction with descent
of (1) at x k .
Proof.
By the KKT conditions of QP sub-problem (8), we
have
∇f ( x k ) + H k d 0k + Ak b k = 0,
p kj + ∇g j ( x k )T d 0k = 0, j ∈ J k ,
Lemma 2.1 Suppose that H2.1and H2.2 hold, then
1) For any iteration, there is no infinite cycle in step 1.
2) If a sequence {x k } of points has an accumulation
_
point, then there exists a constant ε > 0 such that
_
ε k ,ik > ε for k large enough.
Proof.
1) Suppose that the desired conclusion is false, that is
to say, there exists some k, such that there is an infinite
cycle in Step 1, then we obtain, ∀i = 1, 2, , that Ak ,i is
not of full rank, i.e., it holds that
det( AkT,i Ak ,i ) = 0, i = 1, 2,
Lemma 2.2 For the QP sub-problem (8) at x k , if
,
(13)
If d 0k = 0 , we obtain
∇f ( x k ) + Ak b k = 0, p kj = 0, j ∈ J k ,
Thereby, from (7) and x k ∈ X , ∀k implies that
g j ( x k ) = 0, v kj ≥ 0, j ∈ J k .
In addition, we have b k = − Bk ∇f ( x k ) = v k , in a word,
we obtain
∇f ( x k ) + Ak b k = 0, g j ( x k ) = 0, b kj ≥ 0, j ∈ J k ,
Let b kj = 0, j ∈ I \ J k , which shows that x k is a KKT
point of (1).
If d 0k ≠ 0 , we have
g J k ( x k )T d k = AkT d k = − p k − δ k ek ,
And by (6), we can know that J k ,i +1 ⊆ J k ,i . Since there
are only finitely many choices for J k ,i ⊆ I , it is sure that
© 2013 ACADEMY PUBLISHER
∇f ( x ) d 0 = − ( d 0 ) H k d 0 + b p ,
k
So,
T
k
k
T
k
kT
k
JOURNAL OF COMPUTERS, VOL. 8, NO. 6, JUNE 2013
1499
∇f ( x k )T d k = ∇f ( x k )T d 0k − δ k ∇f ( x k )T Ak ( AkT Ak ) −1 ek
prove that d 0* = 0 . Suppose by contradiction that d 0* ≠ 0 .
1
1
≤ − (d 0k )T H k d 0k + b kT p k ≤ − (d 0k )T H k d 0k < 0.
2
2
Thereby, we know that d k is a feasible descent
direction of (1) at x k .
Then, from Lemma 2.2, it is obvious that d * is welldefined, and it holds that
III. GLOBAL CONVERGENCE OF ALGORITHM
∇f ( x* )T d * < 0, ∇g j ( x* )T d * < 0, j ∈ I ( x* ) ⊆ J
Thus, from (16), it is easy to see that the step-size tk
obtained in step 5 are bounded away from zero on
K , i .e .
In this section, firstly, it is shown that Algorithm A
given in section 2 is well-defined, that is to say, for every
k, that the line search at Step 5 is always successful
Lemma 3.1 The line search in step 5 yields a stepsize
1
tk = ( )i for some finite i = i (k ) .
2
Proof.
It is a well-known result according to Lemma 2.2. For
(11),
s f ( x k + td k + t 2 d k ) − f ( x k ) − α t∇f ( x k )T d k
= ∇f ( x k )T (td k + t 2 d k ) + o(t ) − α t∇f ( x k )T d k
= (1 − α )t∇f ( x k )T d k + o(t ).
For (12), if
tk ≥ t* = inf{tk , k ∈ K } > 0, k ∈ K .
In addition, from (11) and Lemma 2.2, it is obvious
that { f ( x k )} is monotonous decreasing. So, according to
assumption
H 2.1, the fact that {x k }K → x* implies that
f ( x k ) → f ( x* ), k → ∞.
(18)
So, from (11), (16), (17), it holds that
0 = lim( f ( x k ) − f ( x* )) ≤ lim(α tk ∇f ( x k )T d k )
x∈K
x∈K
1
≤ α t*∇f ( x* )T d * < 0,
2
which is a contradiction thus lim d 0k = 0 . Thus, x* is a
KKT point of (1).
j ∉ I (x ), g j ( x ) < 0;
k
j ∈ I ( xk ), g j ( xk ) = 0, ∇g j ( xk )T d k < 0,
IV. THE RATE OF CONVERGENCE
so we have
g j ( x k + td k + t 2 d k ) = ∇f ( x k )T (td k + t 2 d k ) + o(t )
= α t∇g j ( x k )T d k + O(t ).
In the sequel, the global convergence of Algorithm A
is shown. For this reason, we make the following
additional assumption.
H3.1 {x k } is bounded, which is the sequence
generated by the algorithm, and there exist constants
b ≥ a > 0 , such that a || y ||2 ≤ yT H k y ≤ b || y ||2 , for all k
and all y ∈ R n .
Since there are only finitely many choices for sets
J k ⊆ I , and the sequence {d 0k , d1k , d k , v k , b k } is bounded,
we can assume without loss of generality that there exists
a subsequence K, such that
x → x , H k → H* , d → d , d → d , d → d ,
*
(17)
x →∞
k
k
(16)
k
0
*
0
k
b k → b* , v k → v * , J k ≡ J ≠ Φ , k ∈ K ,
*
k
*
(15)
where J is a constant set.
Now we discuss the convergent rate of the algorithm,
and prove that the sequence {x k } generated by the
algorithm is one-step super-linearly convergent under
some mild conditions without the strict complementarily.
For this purpose, we add some regularity hypothesis.
H 4.1 The sequence {x k } generated by Algorithm A is
bounded, and possess an accumulation point x* , such
that the KKT pair ( x* , u* ) satisfies the strong secondorder sufficiency conditions, i.e.,
d T ∇ 2xx L( x* , u * )d > 0,
∀d ∈ Ω {d ∈ R n : d ≠ 0, ∇g I + ( x* )T d = 0},
L( x, u ) = f ( x ) + ∑ u j g j ( x), I + = { j ∈ I : u *j > 0}.
j∈ I
Lemma 4.1 Suppose that assumptions H 2.1-H 3.1 hold,
then,
1) There exists a constant ζ > 0 , such that
|| ( AkT Ak ) −1 ||≤ ζ ;
2)
lim d 0k = 0; lim d k = 0; lim d k = 0;
k →∞
k →∞
k →∞
Theorem 3.2 The algorithm either stops at the KKT
point x k of the problem (1) in finite number of steps, or
generates an infinite sequence {x k } any accumulation
3)
point x* of which is a KKT point of the problem (1).
Proof.
The first statement is easy to show, since the only
stopping point is in step 3. Thus, assume that the
algorithm generates an infinite sequence {x k } , and (15)
holds. According to Lemma 2.2, it is only necessary to
Proof.
1) By contradiction, suppose that sequence
{|| ( AkT Ak ) −1 ||} is unbounded, then there exists an infinite
subset K, such that
|| ( AkT Ak ) −1 ||→ ∞, (k ∈ K ).
© 2013 ACADEMY PUBLISHER
|| d k ||∼|| d 0k ||, || d k ||= O(|| d k ||2 ),
||d k − d 0k ||=O (|| d 0k ||3 ), || d k ||= O(|| d k ||2 ).
.
1500
JOURNAL OF COMPUTERS, VOL. 8, NO. 6, JUNE 2013
In view of the boundedness of {x k } and J k being a
subset of the finite set I = {1, 2, , m} as well as
Lemma 2.1, we know that there exists an infinite index
set K ' ⊆ K such that
x k → x, J k ≡ J ' , ∀k ∈ K ' , det( AkT Ak ) ≥ ε , ε k ≥ ε .
As a result,
lim(
AkT Ak ) = ∇g J ' ( x)T ∇g J ' ( x),
'
k ∈K
−1
continuously differentiable, for k
Hence, we obtain || ( A Ak ) ||→|| ∇g J ' ( x) ∇g J ' ( x) ||,
T
this contradict || ( AkT Ak ) −1 ||→ ∞, (k ∈ K ). So the first
conclusion 1) follows.
2) We firstly show that lim d 0k = 0 .
k →∞
We suppose by contradiction that lim d 0k ≠ 0,
enough, if v < 0 , we have
p j ( x k ) + ∇g j0 ( x* )T d 0k = −v kj0 + ∇g j0 ( x* )T d 0k
0
1
≥ − v kj0 > 0.
2
Otherwise,
p j ( x k ) + ∇g j0 ( x* )T d 0k
k →∞
then
1
= g j0 ( x k ) + ∇g j0 ( x* )T d 0k ≤ − β < 0, (v kj0 ≥ 0)
2
which is contradictory with (8) and the fact j0 ∈ J k . So,
J k ≡ I* (for k large enough).
Prove that b k → uI* = (u *j , j ∈ I* ), v k → (u *j , j ∈ I* ) .
For the v k → (u *j , j ∈ I* ) statement, we have the
there exist an infinite index set K and a constant σ > 0
such that || d 0k ||> σ holds for all k ∈ K . Taking notice
following results from the definition of v k ,
v k → − − B*∇f ( x* ) = −( A*T A* ) −1 A*T ∇f ( x* )
of the boundedness of {x k } , by taking a subsequence if
necessary, we may suppose that
x k → x, J k ≡ J ' , ∀k ∈ K .
Using Taylor expansion, we analyze the first search
inequality of Step 5, combining the proof of Theorem 3.2,
the fact that x k → x* , (k → ∞) implies that it is true.
i.e.
The proof of lim d = 0; lim d = 0 are elementary
k
k
k →∞
k →∞
from the result of 1) as well as formulas (9) and (10).
3) The proof of 3) is elementary from the formulas (9),
(10) and assumption H2.1.
Lemma
4.2.
Let H2.1 to H4.1 holds,
lim || x k +1 − x k ||= 0 . Thereby, the entire sequence {x k }
k →∞
converges to x* i.e. x k → x* , k → ∞ .
Proof.
From the Lemma 4.1, it is easy to see that
lim || x k +1 − x k ||= lim(|| tk d k + tk2 d k ||)
k →∞
k →∞
≤ lim(|| d k || + || d k ||) = 0
In addition, since x* is a KKT point of (1), it is
evident that
∇f ( x* ) + A*uI* = 0, uI* = − B*∇f ( x* )
.
Moreover, together with Theorem 1.1.5 in [4], it shows
that x k → x* , k → ∞
Lemma 4.3 It holds, for k large enough, that
1) J k ≡ I ( x* ) I* , b k → uI* = (u*j , j ∈ I* ), v k → (u *j , j ∈ I* )
2) I + ⊆ Lk = { j ∈ J k : g j ( x k ) + ∇g j ( x k )T d 0k = 0} ⊆ J k .
Proof.
1) Prove J k ≡ I* .
On one hand, from Lemma 2.1, we know, for k large
enough, that I* ⊆ J k . On the other hand, if it doesn’t
hold that J k ⊆ I* , then there exist constants j0 and
β > 0 , such that
g j0 ( x* ) ≤ − β < 0, j0 ∈ J k .
So, according to d 0k → 0 and the functions g j ( x),
−1
uI* = −( A A* ) A ∇f ( x ).
T
*
T
*
*
Otherwise, from (8), the fact that d 0k → 0 implies that
∇f ( x k ) + H k d 0k + Ak b k = 0, b k → − B*∇f ( x* ) = u I* .
The claim holds.
2) For lim( x k , d 0k ) = ( x* , 0) , we have Lk ⊆ I ( x* ) .
k →∞
Furthermore, it has lim u Ik+ = uI*+ > 0 , so the proof is
x →∞
finished.
In order to obtain super-linear convergence, a crucial
requirement is that a unit step size is used in a
neighborhood of the solution. This can be achieved if the
following assumption is satisfied.
H4.2 Let || (∇ 2xx L( x k , u Jkk ) − H k )d k ||= o(|| d k ||) , where
L( x, u Jkk ) = f ( x) +
k →∞
© 2013 ACADEMY PUBLISHER
large
k
j0
0
det(∇g J ' ( x)T ∇g J ' ( x)) ≥ ε > 0.
T
k
( j ∈ I ) are
∑u
j∈ J k
k
Jk
g j ( x) .
Lemma 4.4 Suppose that Assumption H 2.1 to H 4.2
are all satisfied. Then, the step size in Algorithm A
always one, i.e. tk ≡ 1 , if k is large enough.
Proof.
It is only necessary to prove that
f ( x k + d k + d k ) ≤ f ( x k ) + α∇f ( x k )T d k ,
g j ( x k + d k + d k ) ≤ 0, j ∈ I .
For
(12)
k
if
j ∈ I \ I*
we
have
(19)
(20)
g j ( x* ) < 0
,
( x k , d k , d ) → ( x* , 0, 0)(k → ∞), then, it is easy to
obtain g j ( x k + d k + d k ) ≤ 0 holds.
If j ∈ I* we have
JOURNAL OF COMPUTERS, VOL. 8, NO. 6, JUNE 2013
g j (xk + dk + dk ) = g j (xk + dk ) +∇g j (xk + dk )T d k + O(|| d k ||2 )
=g j (xk + dk ) +∇g j (xk )T dk + O(|| dk |||| d k ||) + O(|| d k ||2 )
1501
− ∑ u kj ∇g j ( x k ) (d k + d )
(21)
j∈Lk
=
=g j (x + d ) +∇g j (x ) d + O(|| d |||| d ||).
k
k
k T
k
k
k
∑ u ∇g
j∈Lk
In addition, from (9) and (10),
∇g j ( x k )T d k = ∇g j ( x k )T d 0k − δ k ,
k
(xk )
(26)
⎞ k
k 2
⎟⎟ d + o(|| d || ),
⎠
k
∇f ( xk )T (d k + d )
so, for τ ∈ (2,3) we have
= −(d k )T Hk d k − ∑ ukj ∇g j ( xk )
g j ( xk + d k + d k )
(27)
j∈Lk
=− || d0k ||τ + g j ( xk ) + ∇g j ( x k )T d0k − δ k + O(|| d k |||| d k ||)
≤ − || d || +O(|| d |||| d ||) ≤ 0.
Hence, the second inequalities of (20) hold for t = 1
and k is sufficiently large.
The next objective is to show the first inequality of (19)
holds. From Taylor expansion and taking into account
Lemma 4.1 and Lemma 4.3, we have
s
T
j
From (25) and (26), we have
+ g j ( x k ) + ∇g j ( x k )T d k ,
τ
k
j
⎛
T
1
+ (d k )T ⎜⎜ ∑ u kj ∇ 2 g j ( x k )
2
⎝ j∈Lk
∇g j ( x k )T d = − || d 0k ||τ − g j ( x k + d k )
k
0
k
T
k
⎛
T ⎞
1
+ (d k )T ⎜⎜ ∑ ukj ∇2 g j ( xk ) ⎟⎟ d k + o(|| d k ||2 )
2
⎝ j∈Lk
⎠
k
(22)
k
k
2
k
j
T
k
1
+ (d k )T ∇ 2 L( x k , u Jkk ) − H k d k + o(|| d k ||2 ).
2
Then, together assumption H 3.1 and H 4.2 as well as
u kj g j ( x k ) ≤ 0, shows that
(
∇f ( x k ) = − H k d 0k − ∑ u kj ∇g j ( x k )
= − H k d k − ∑ u kj ∇g j ( x k ) + o(|| d k ||2 ),
2
j ∈ Lk
On the other hand, from the KKT condition of (8) and
the active set Lk defined by Lemma 4.3 one has
j∈Lk
1
⎛
⎞
+ (d k )T ⎜ ∇ f ( x ) + ∑ u ∇ g ( x ) − H ⎟ d k + o(|| d k ||2 )
2
⎝
⎠
1 k T
=(α − )(d ) H k d k + (1 − α ) ∑ u kj g j ( x k )
2
j∈Lk
j
f ( x k + d k + d k ) − f ( x k ) + α ∇ f ( x k )T d k
1
=∇f ( x k )T (d k + d k ) + (d k )T ∇ 2 f ( x k )T d k
2
− α∇f ( x k )T d k + o(|| d k ||2 ).
Substituting (27) and (24) into (22), it holds that
1
s (α − )(d k )T H k d k + (1 − α ) ∑ u kj g j ( x k )
2
j∈Lk
(23)
)
1
1
s ≤ (α − )a || d k ||2 +o(|| d k ||2 ) ≤ 0. (α ∈ (0, )).
2
2
Hence, the inequality of (19) holds.
j∈Lk
Furthermore, in a way similar to the proof of Theorem
5.2 in [5] and in [19, Theorem 2.3], we may obtain the
following theorem:
So, from (23) and Lemma 4.3, we have
∇f ( x k )T d k
= −(d k )T H k d k − ∑ u kj ∇g j ( x k ) d k + o(|| d k ||2 )
T
(24)
j∈Lk
= − (d k )T H k d k − ∑ u kj ∇g j ( x k ) d 0k + o(|| d k ||2 )
T
Theorem 4.5 Under all above-mentioned assumptions,
the algorithm is superlinearly convergent, i.e., the
sequence {x k } generated by the algorithm satisfies that
j∈Lk
|| x k +1 − x* ||= o(|| x k − x* ||).
k
∇f ( xk )T (d k + d )
= −(d k )T Hk d k − ∑ ukj ∇g j ( xk ) (d k + d ) + o(|| d k ||2 ),
T
k
(25)
j∈Lk
Again, from (21) and Taylor expansion, it is clear that
Proof.
From Lemma 4.1 and Lemma 4.4, we can know that
the sequence {x k } yielded by Algorithm A has the form
of
x k +1 = x k + d k + d
k
k
= x k + d 0k + (d k + d − d 0k )
o(|| d || ) = g j (x + d + d )
2
k
k
k
k
1
=g j (x ) +∇g j (x ) (d + d ) + (d k )T ∇2 g j (xk )T d k + o(|| d k ||2 )
2
where j ∈ Lk , then, we obtain
k
k T
k
k
x k + d 0k + d .
k
k
k
where d = (d k + d − d 0k ) (for k large enough) and
k
|| d ||= O(|| d 0k ||3 ) . Consequently, we can obtain the
result together with Ref. [5] and [19].
V. NUMERICAL RESULTS
© 2013 ACADEMY PUBLISHER
1502
JOURNAL OF COMPUTERS, VOL. 8, NO. 6, JUNE 2013
In this section, we carry out numerical experiments
based on the Algorithm A. The code of the proposed
algorithm is written by using MATLAB 7.0 and utilized
the optimization toolbox. The results show that the
algorithm is effective. During the numerical experiments,
it is chosen at random some parameters as follows:
ε 0 = 0.5, α = 0.25, τ = 2.25, H 0 = I ,
where I is the n × n unit matrix. H k is updated by the
BFGS formula [2].
H k +1 = BFGS ( H k , s k , y k ),
Where,
^
CPU: the total time taken by the process (unit:
millisecond) ;
FV: the final value of the objective function.
TABLE III.
THE APPROXIMATE VALUE OF THE DIRECTION
d 0k
FOR TABLE I
NO.
n, m
|| d 0k ||
HS12
HS43
HS66
HS100
HS113
2, 1
4, 3
3, 8
7, 4
10, 8
7.329773437334 E-09
5.473511535838 E-09
8.327832675386 E-09
8.595133692328 E-09
6.056765632745 E-09
s k = x k +1 − x k , y k = θ y k + (1 − θ ) H k s k ,
^
m
y k = ∇f ( x k +1 ) − ∇f ( x k ) + ∑ u kj (∇g j ( x k +1 ) − ∇g j ( x k )),
j =1
kT
⎧1,
if y s k ≥ 0.2( s k )T H k s k ,
⎪⎪
θ = ⎨ 0.8( s k )T H k s k
, otherwise.
⎪ k T
kT
k
k
⎪⎩ ( s ) H k s − y s
In the implementation, the stopping criterion of Step 2
is changed to “If || d 0k ||≤ 10−8 , STOP.”
TABLE I.
THE DETAIL INFORMATION OF NUMERICAL EXPERIMENTS
NO.
HS12
HS43
HS66
HS100
HS113
n, m
2, 1
4, 3
3, 8
7, 4
10, 8
NT
10
17
14
18
45
CPU
0
10
0
62
50
TABLE II.
THE APPROXIMATE OPTIMAL SOLUTION
NO.
HS12
HS43
HS66
HS100
HS113
x* FOR TABLE I
*
the approximate optimal solution x
(1.999999999995731, 3.000000000011285) T
(0.000000000000000, 1.000000000000000,
2.000000000000000, −1.000000000000000) T
(0.184126482757009, 1.202167866986839,
3.327322301935746) T
(2.330499372903103, 1.951372372923884,
-0.477541392886392, 4.365726233574537,
-0.624486970384889, 1.038131018506466,
1.594226711671913) T
(2.171996371254668, 2.363682973701174,
8.773925738481299, 5.095984487967813,
0.990654764957730, 1.430573978920189,
1.321644208159091, 9.828725807883636,
8.280091670090108, 8.375926663907775) T
This algorithm has been tested on some problems from
Ref.[20], a feasible initial point is either provided or
obtained easily for each problem. The results are
summarized in Table 1 to Tabe 4. The columns of this
table have the following meanings:
No.: the number of the test problem in [20];
n: the number of variables;
m: the number of inequality constraints;
NT: the number of iterations;
© 2013 ACADEMY PUBLISHER
TABLE IV.
THE FINAL VALUE OF THE OBJECTIVE FUNCTION FOR TABLE I
NO.
HS12
HS43
HS66
HS100
HS113
FV
-29.999999999999705
-44.000000000000000
0.518163274181542
6.806300573744022e+002
24.306209068179822
VI. CONCLUSION
This paper proposed an improved feasible sequential
quadratic programming (FSQP) method for nonlinear
programs. As compared with the existing SQP methods
which required solving the QP sub-problem with
inequality constraints in single iteration, in order to obtain
the feasible direction, the method of this paper is only
necessary to solve an equality constrained quadratic
programming sub-problems. Combined the generalized
projection technique, a height-order correction direction
is yielded by explicit formulas, which can avoids Maratos
effect. Furthermore, under some mild assumptions, the
algorithm is globally convergent and its rate of
convergence is one-step super-linearly. Numerical results
reported show that the algorithm in this paper is effective.
ACKNOWLEDGMENT
The author would like to thank the editors, whose
constructive comments led to a considerable revision of
the original paper.
REFERENCES
[1] S. P. Han. “Superlinearly Convergent Variable Metric
Algorithm for General Nonlinear Programming Problems”,
Mathematical Programming, Berlin, vol.11 (1), pp. 263–
282, December 1976.
[2] M. J. D. Powell. “A Fast Algorithm for Nonlinearly
Constrained Optimization Calculations”, In: Waston, G.A.
(ed). Numerical Analysis. Springer, Berlin, pp. 144–157,
1978.
[3] P. T. Boggs, J. W. Tolle. “A Strategy for Global
Convergence in a Sequential Quadratic Programming
Algorithm”, SIAM J. Num. Anal., Philadelphia, vol. 26 (1),
pp. 600–623, June 1989.
[4] E. R. Panier, A. L. Tits. “On Combining Feasibility,
Descent and Superlinear Convergence in Inequality
JOURNAL OF COMPUTERS, VOL. 8, NO. 6, JUNE 2013
[5]
[6]
[7]
[8]
[9]
[10]
[11]
[12]
Constrained Optimization”, Mathematical Programming,
Berlin, vol. 59, pp. 261–276, 1993.
J.F. Binnans, G. Launay. “Sequential quadratic
programming with penalization the displacement”, SIAM J.
Optimization, Philadelphia, vol. 5 (4), pp. 796–812, 1995.
P. Spellucci. “An SQP Method for General Nonlinear
Programs Using Only Equality Constrained Subproblems”,
Mathematical Programming, Berlin, vol. 82 (3), pp. 413–
448, August 1998.
C. T. Lawarence, and A.L.Tits. “A Computationally
Efficient Feasible Sequential Quadratic Programming
Algorithm”, SIAM J.Optim., Philadelphia, vol. 11, pp.
1092–1118, 2001.
L. Qi, Y. F. Yang. “Globally and Superlinearly Convergent
QP-free
Algorithm
for
Nonlinear
Constrained
Optimization”, Journal of Optimization Theory and
Applications, Berlin, vol. 113 (2), pp. 297–323, May 2002.
(7)
J.B.Jian, C.M.Tang. “An SQP Feasible Descent Algorithm
for Nonlinear Inequality Constrained Optimization without
Strict Complementarity”, An International Journal
Computers and Mathematics with application, vol. 49, pp.
223–238, 2005.
Z.J. Luo, Z.B. Zhu, L.R Wang, and A.J Ren. “An SQP
algorithm with equality constrained sub-problems for
general constrained optimization”, International Journal of
Pure and Applied Mathematics, vol. 39 (1): pp. 125-138,
2007.
Richard H. Byrd, Frank E. Curtis, and Jorge Nocedal.
“Infeasibility Detection and SQP Methods for Nonlinear
Optimization”, SIAM J. Optim. Vol. 20 (5), pp. 2281-2299,
2010.
A. F. Izmailov and M. V. Solodov. “A Truncated SQP
Method Based on Inexact Interior-Point Solutions of Sub-
© 2013 ACADEMY PUBLISHER
1503
[13]
[14]
[15]
[16]
[17]
[18]
[19]
[20]
problems”, SIAM J. Optim. Vol. 20 (5), pp. 2584-2613,
2010.
C.G. Shen, W.J. Xue and X.D. Chen. “Global convergence
of a robust filter SQP algorithm”, European Journal of
Operational Research, vol. 206 (1), pp. 34–45, 2010.
Z. B. Zhu, W.D. Zhang, and Z.J. Geng. “A feasible SQP
method for nonlinear programming”, Applied Mathematics
and Computation, vol. 215, pp. 3956–3969, 2010.
Z. J. Luo, G. H. Chen, and L. R. Wang, "A Modified
Sequence Quadratic Programming Method for Nonlinear
Programming, " cso, pp.458-461, 2011 Fourth
International Joint Conference on Computational Sciences
and Optimization, 2011.
C. Gu, D.T. Zhu. “A non-monotone line search
multidimensional filter-SQP method for general nonlinear
programming”, Numerical Algorithms, vol. 56 (4),
pp.537–559, 2011.
C.L. Hao, X.W. Liu. “A trust-region filter-SQP method for
mathematical programs with linear complementarity
constraints”, Journal of Industrial and Management
Optimization, vol. 7 (4), pp. 1041–1055, 2011.
C.L. Hao, X.W. Liu. “Global convergence of an SQP
algorithm for nonlinear optimization with over determined
constraints”, Numerical algebra control and optimization,
vol. 2 (1), pp. 19–29, 2012.
J.B. Jian. “Two extension models of SQP and SSLE
algorithms for optimization and their superlinear and
quadratical convergence”, Applied Mathematics--A
Journal of Chinese Universities, Set. A, vol. 16 (4),
pp.435-444, 2001.
W. Hock, K. Schittkowski, Test examples for nonlinear
programming codes, Lecture Notes in Economics and
Mathematical Systems, vol. 187, Springer, Berlin, 1981.