VARIATIONAL APPROACH TO SECOND-ORDER - IMJ-PRG

c 2012 Society for Industrial and Applied Mathematics
SIAM J. CONTROL OPTIM.
Vol. 50, No. 3, pp. 1139–1173
VARIATIONAL APPROACH TO SECOND-ORDER OPTIMALITY
CONDITIONS FOR CONTROL PROBLEMS WITH PURE STATE
CONSTRAINTS∗
DANIEL HOEHENER†
Abstract. For optimal control problems with set-valued control constraints and pure state
constraints we propose new second-order necessary optimality conditions. In addition to the usual
second-order derivative of the Hamiltonian, these conditions contain extra terms involving secondorder tangents to the set of feasible trajectory-control pairs at the extremal process under consideration. The second-order necessary optimality conditions of the present work are obtained by using
a variational approach. In particular, we present a new second-order variational equation. This
approach allows us to make direct proofs as opposed to the classical way of obtaining second-order
necessary conditions by using an abstract infinite dimensional optimization problem. No convexity
assumptions on the constraints are imposed and optimal controls are required to be merely measurable.
Key words. optimal control, second-order necessary optimality conditions, pure state constraints, variational equation, second-order variational equation, maximum principle, second-order
tangents
AMS subject classifications. 49K15, 49K21, 49K30, 34K27
DOI. 10.1137/110828320
1. Introduction. We consider the optimal control problem of the Bolza form
(P )
1
Minimize
l(t, x(t), u(t))dt
0
over measurable u and solutions x of the control system
ẋ(t) = f (t, x(t), u(t)) a.e. in [0, 1],
(1.1)
x(0) = x0
satisfying the constraints
x(t) ∈ K
∀t ∈ [0, 1]
and
u(t) ∈ U (t) a.e.,
where the maps f : [0, 1] × Rn × Rm → Rn and l : [0, 1] × Rn × Rm → R, the set-valued
map U : [0, 1] Rm , the subset K ⊂ Rn , and the initial state x0 ∈ Rn are given. The
special case when K = Rn will be denoted by (PN C ).
We refer to a measurable function u : [0, 1] → Rm as a control function or simply
control. The set of all control functions u such that u(t) ∈ U (t) a.e. will be denoted by
U. A process (x, u) comprises a control function u ∈ U and an absolutely continuous
function x ∈ W 1,1 ([0, 1]; Rn ) which satisfies the differential equation (1.1). A state
trajectory x is a solution of (1.1) for some control function u. A process (x, u) is
called feasible if x(t) ∈ K for all t ∈ [0, 1]. Under some standard assumptions, the
∗ Received
by the editors March 22, 2011; accepted for publication (in revised form) January
9, 2012; published electronically May 15, 2012. Financial support for this work by the European
Commission (FP7-PEOPLE-2010-ITN, grant 264735-SADCO) is gratefully acknowledged.
http://www.siam.org/journals/sicon/50-3/82832.html
† Combinatoire & Optimisation, Institut de Mathématiques de Jussieu (UMR 7586), Université
Pierre et Marie Curie, 4 place Jussieu, 75252 Paris cedex 05, France ([email protected]).
1139
1140
DANIEL HOEHENER
state trajectory x is uniquely defined by a control function u and the control system
(1.1). Consequently, denoting by M([0, 1]; Rm ) the set of all measurable maps from
[0, 1] to Rm , we can define the cost functional J : M([0, 1]; Rm ) → R by
J(u) :=
0
1
l(t, xu (t), u(t))dt,
where xu ∈ W 1,1 ([0, 1]; Rn ) is the solution of (1.1). For ū ∈ M([0, 1]; Rm ) and
u ∈ L∞ ((0, 1); Rm ), let J (ū)u denote the directional derivative of J at ū in the
direction u. The Hamiltonian H : [0, 1] × Rn × Rn × Rm of problem (P ) is defined by
H(t, x, p, u) = f (t, x, u), p − l(t, x, u).
(1.2)
Let (x̄, ū) be an optimal solution of problem (P ) and assume that the constrained normal maximum principle holds true. That is, there exist p ∈ W 1,1 ([0, 1]; Rn ), a Borel
measurable ν : [0, 1] → Rn , and a positive Radon measure μ, satisfying the conclusion
of Theorem 3.4 (see section 3 below). It follows then that for ψ(t) := [0,t] ν(s)dμ(s)
whenever t ∈ ]0, 1] and for every pair (y, u) ∈ W 1,1 ([0, 1]; Rn ) × L∞ ((0, 1); Rm ) that
is “tangent” to the set of feasible processes of (P ), we have
(1.3) J (ū)u = −
0
1
Hu (t, x̄(t), p(t) + ψ(t), ū(t))u(t)dt −
[0,1]
ν(t), y(t)dμ(t) ≥ 0.
A more precise description of a suitable notion of tangency is given in section 4. A
proof of (1.3) is provided in section 5.
This paper is devoted to second-order necessary optimality conditions. Similar to
classical analysis, where second-order conditions are meaningful only when first-order
terms vanish, our second-order necessary conditions apply to “tangents” u such that
J (ū)u = 0. These second-order conditions are stated in Theorem 3.5 and constitute
the main contribution of this paper.
Second-order optimality conditions for optimal control problems have been studied for almost half a century, but it remains a challenge to formulate them in a very
general context. Earlier results concerning second-order necessary conditions for control problems without state constraints are due to Hestenes [28] and Warga [52, 53].
Then, Gilbert and Bernstein [27] generalized and simplified these results in order to
state second-order necessary conditions for problems with very general constraints
on controls. More precisely, they allowed U to be an arbitrary subset of the set of
measurable maps M([0, 1]; U ), where U is an open subset of Rm .
Let (x̄, ū) ∈ W 1,1 ([0, 1]; Rn ) × M([0, 1]; Rm ) be an extremal process for a control
problem and let C(x̄, ū) denote a set of critical directions associated with it. Ideally,
C(x̄, ū) should be the set of tangents to the set of feasible processes. However, since
in practice one needs to describe the set of critical directions analytically, C(x̄, ū) is
usually assumed to be a subset of the ideal set of critical directions. The choice of the
set C(x̄, ū) is therefore a part of the second-order statements and has an important
impact on their quality. One can distinguish two ways of formulating second-order
necessary conditions. The first common formulation is as follows:
For all (y, u) ∈ C(x̄, ū), there exist multipliers (adjoint state variable,
cost multipliers, etc.) such that some second-order conditions hold
true.
SECOND-ORDER NECESSARY OPTIMALITY CONDITIONS
1141
The second alternative, which in general is stronger, is as follows:
There exist multipliers such that some second-order conditions hold
true ∀(y, u) ∈ C(x̄, ū).
In the first case, multipliers depend on critical directions; in the second case they are
independent from them.
For instance, in [27, Thm. 3.1], multipliers depend on the critical direction which
makes statements weaker. Also, the authors of [27] impose that ∀(y, u) ∈ C(x̄, ū), we
have u ∈ U, where
(1.4)
U ⊂ L∞ ([0, 1]; Rm ),
U is convex,
U ⊂ U − ū,
0 ∈ U.
Consequently, in the case of a nonconvex U, there might be many feasible directions
which are not in C(x̄, ū).
Later on, Maruyama [36] imposed the additional hypothesis that the set U is
a closed, convex subset of L∞ ((0, 1); Rm ) with nonempty interior and obtained the
same results as in [27] but with multipliers independent from the critical directions. In
[54], Zeidan and Zezza derived results similar to those in [36], without the convexity
assumptions, for control sets described by equality and inequality constraints and
piecewise continuous controls (i.e., in P W C([0, 1]; Rm )). More precisely,
U = {u ∈ P W C([0, 1]; Rm ) | g(u(t)) = 0 and h(u(t)) ≤ 0 ∀t ∈ [0, 1]},
where g and h are some given smooth maps.
In [27, 36, 52, 53] the authors deduced the second-order optimality conditions
from an abstract infinite dimensional optimization problem. A different approach was
proposed by Hestenes [28] and followed more recently by Milyutin and Osmolovskii
[38, 39]. There, the authors worked directly in the state-control space. A particularity
of [38, 39] is that the main results are stated for the so-called Pontryagin minimum.
In general, the notion of the Pontryagin minimum is weaker than the notion of strong
local minima, investigated in the present paper. Indeed, every strong local minimum
is a Pontryagin minimum. Further, in [38, 39] the authors restrict their attention
to mixed state-control constraints. Let us underline that the hypotheses needed to
obtain their results, notably the full rank condition (see, for instance, [39, Ch. 3]), do
not hold for pure state constraints.
In Theorem 3.2 we state second-order necessary optimality conditions for the
Bolza problem (PN C ) involving a normal maximum principle (see, for instance, [51]).
Using second-order tangents, we define a set of admissible variations of a reference control which is in general larger than U from (1.4). Our result applies to any first-order
necessary optimality conditions in the form of the constrained maximum principle. In
addition, the use of second-order variations leads to a very geometric and, as we believe, quite direct proof of Theorem 3.2, in comparison to some very technical results
in second-order theory. In [54], the authors used a variational approach that is similar
to ours but they only considered first-order variations of controls and included the
control constraints into the definition of the Hamiltonian. Therefore their optimality
conditions are different from ours. For the sake of completeness, by using techniques
similar to those of [39], we deduce from Theorem 3.2 second-order necessary conditions
for a problem with mixed state-control equality constraints (Theorem 3.3) similar to
those in [38, 39] but with the additional advantage that the optimal control is assumed
to be essentially bounded instead of piecewise continuous as in [38, 39].
Note that there is a rich literature on problems with mixed equality and inequality
constraints; see, for instance, [16, 49] and the references therein. For problems that are
1142
DANIEL HOEHENER
linear in the control see Dmitruk [19, 20, 21, 22]. The goal of this paper is to provide
second-order necessary conditions that are applicable to optimal control problems
under pure state constraints. A comparison of our results with those on problems
with mixed constraints and/or particular dynamics is therefore beyond the scope of
this introduction.
Pure state constraints, which are important in applications, bring some additional
difficulties because the state is not linked to controls by a functional relation. Consider
the case where the state constraint is formulated as an inequality constraint, i.e.
(1.5)
g(x(t)) ≤ 0 ∀t ∈ [0, 1],
where x : [0, 1] → Rn denotes the state trajectory and g : Rn → Rk is a given (smooth)
map. A classical way of dealing with state constraints of this form is to differentiate
the components gi of the map g(x(·)) with respect to time, until an explicit dependence from the control appears, so that the problem can be treated as a problem
with mixed constraints. See [11, 12, 26, 31] for first-order necessary conditions using
this approach. Bonnans and Hermant [6, 7, 8] provided second-order necessary and
sufficient conditions for optimal control problems with state constraints of the form
(1.5). They used second-order necessary conditions due to Kawasaki for an abstract
optimization problem with infinitely many inequality constraints [32, 34]. The major
difficulty in this approach is that these second-order necessary conditions contain, in
the presence of pure state constraints, a so-called curvature term. By a careful analysis of junction points, using an extension of the junction conditions from [31], they
were able to characterize the curvature term in a way that led to no-gap second-order
optimality conditions, i.e., second-order necessary and sufficient conditions that are as
close as in classical analysis (see [8]). Thanks to the no-gap character of the optimality conditions, the second-order necessary conditions are very strong. On the other
hand, these results are proved under quite restrictive assumptions. One of these is
continuity of the optimal control (their assumptions do imply that the optimal control
is continuous). In addition, except in the case of mixed state-control constraints, they
do not allow constraints on the control, and finally, it is required that there exist only
finitely many boundary arcs, which for a constraint of order ≥3 in general is not true;
see, for instance, [37, 47].
Páles and Zeidan addressed the problem of pure state constraints in a series
of papers, see [40, 43, 44] and the references therein. The optimal control problem
considered in [40] is similar to the one of [8]. However, contrary to the strong regularity
conditions in [8], Páles and Zeidan provide necessary conditions for problems where the
map g from (1.5) is only locally Lipschitz continuous. The drawback of their approach
is that, as in [27], the multipliers depend upon critical directions. In [43] these results
were extended to problems with more general pure and mixed state constraints:
g(t, x(t), u(t)) ∈ Q(t) a.e.,
k(t, x(t)) ∈ S(t) ∀t,
where Q and S are set-valued maps with closed, convex values having nonempty interior. Finally, in [44] the second-order conditions were derived for time independent
state constraints S(t) = S ∀t ∈ [0, 1], where S is a closed convex subset of Rn with
nonempty interior, and control constraints are of the form u(t) ∈ U almost everywhere, where U is an arbitrary subset of Rm . Note that in both cases the pure state
constraints are more general than (1.5), but multipliers depend on critical directions.
In addition, to use the time transformation introduced by Dubovitskii and Milyutin
SECOND-ORDER NECESSARY OPTIMALITY CONDITIONS
1143
in [23], the authors impose in [44] differentiability of f with respect to the time variable. This approach has the advantage that their results can be extended to time
optimal problems; see [45]. However, for nonautonomous control systems arising in
applications, it often occurs that the dynamic of the system is less regular in the time
variable and may even be discontinuous.
In mathematical programming, second-order optimality conditions are classically
derived using second-order variations; see [24]. Since these early results, much work
has been done to extend this approach to a very general context, notably constrained
optimization problems in infinite dimensional spaces (see, for instance, [3, 5, 15, 30,
32, 41, 46] and the references therein). A detailed treatment of this topic can be
found in the book by Bonnans and Shapiro [9]. An important application of such
second-order necessary optimality conditions for abstract optimization problems are
second-order necessary optimality conditions for optimal control problems.
In the present work, unlike previous papers on second-order necessary optimality
conditions for control problems with pure state constraints, we do not reformulate
our optimal control problem in an abstract form. Indeed, some of our assumptions,
in particular the low regularity of the optimal control, are not covered by known
abstract general second-order theories. Therefore we provide a direct proof of secondorder necessary optimality conditions. For this aim we use second-order variations
of the state trajectory, which we obtain from second-order control variations and a
second-order variational equation; see section 4. Some of the results of section 4
(Propositions 4.2 and 4.4) are new. Further, in the difference with the approach
involving an abstract optimization problem, we separate proofs of first- and secondorder conditions, which allows us to benefit entirely from the most general first-order
conditions. See, for instance, [51] for an overview of first-order necessary optimality
conditions. In addition, our main result is a straightforward extension of the case when
state constraints are absent. In particular, the presence of state constraints does not
lead to any restrictions on control constraints and multipliers are independent from
critical directions. Finally, we do not assume any regularity of the optimal control
except the usual measurability and also dynamics of the control system are allowed
to be merely measurable in the time variable. The price to pay for dealing with
measurable optimal controls is that our results are stated for the subset A(1) (x̄, ū) of
the set of first-order tangents A(x̄, ū). (See section 3 for the definitions of these sets.)
Finally, normality of the maximum principle is important for second-order conditions,
and even though there exist some sufficient conditions for normality of the constrained
maximum principle with end point constraints (see, for instance, [25]), they are not
applicable in the case of a fixed initial state. For this reason, we do not consider end
point constraints here.
The paper is organized as follows. In section 2 we recall some definitions and
state the main assumptions. In section 3 we present the main results of this paper,
in particular, second-order necessary optimality conditions for problems (PN C ) and
(P ). First- and second-order variational equations for problems with and without
state constraints are provided in section 4. The proofs of the main results are given
in sections 5 and 6.
2. Notation and main assumptions.
2.1. Basic notation. We denote the norm in Rn by |·| and by ·, · the inner
product. B := {x ∈ Rn | |x| ≤ 1} denotes the closed unit ball, B̊ its interior,
i.e., the open unit ball, and S n−1 the unit sphere, i.e., the boundary of B. B(x, r)
denotes a closed ball with radius r > 0 and center x ∈ Rn , i.e., B(x, r) = x + rB.
1144
DANIEL HOEHENER
For an arbitrary set S ⊂ Rn , ∂S denotes its boundary, Int S its interior, and S c its
complement. The convex hull and the closed convex hull of the set S will be denoted
by co S and co S, respectively.
W 1,1 ([0, 1]; Rn ) stands for the space of absolutely continuous maps from [0, 1]
to Rn and M([0, 1]; Rm ) for the space of measurable maps from [0, 1] to Rm . For
1 ≤ p < ∞ (resp., p = ∞), Lp ((0, 1); Rm ) denotes the Lebesgue space; in particular,
for maps u ∈ Lp ((0, 1); Rm ) we have that |u|p is integrable, respectively, that |u|
is essentially bounded. N BV ([0, 1]; Rn ) denotes the space of normalized functions of
bounded variation on [0, 1] with values in Rn , i.e., functions of bounded total variation,
vanishing at 0 and right-continuous on ]0, 1[. For any ψ ∈ N BV ([0, 1]; Rn ), the right
(left) limit of ψ at t ∈ [0, 1[ (resp., t ∈ ]0, 1]) is denoted by ψ(t+) (resp., ψ(t−)). For
properties of the space N BV ([0, 1]; Rn ) see, for instance, [35].
For a continuous map f : [0, 1] → Rn we denote the supremum norm by f ∞ :=
maxt∈[0,1] |f (t)|. The norm of a map u ∈ Lp ((0, 1); Rm ) for 1 ≤ p < ∞ (resp., p = ∞)
is denoted by
1/p
up :=
[0,1]
|u(t)|p dt
,
respectively, u∞ := inf {C ≥ 0 | |u(t)| ≤ C for a.e. t ∈ [0, 1]}. Here and in the rest
of the paper a.e. means with respect to the Lebesgue measure.
Finally, let Y be a normed, finite dimensional vector space and L : Rn → Y a
linear map. Then the norm of L is the operator norm, i.e., L := inf {C ≥ 0 |
LxY ≤ C ∀x ∈ S n−1 }.
Let f : [0, 1]×Rn ×Rm → Rk be a map such that f (t, ·, ·) is differentiable for every
t ∈ [0, 1]. Then we denote by fx (t0 , x0 , u0 ) and fu (t0 , x0 , u0 ) the partial derivative of
f with respect to x, respectively, u, at (t0 , x0 , u0 ). Further, f (t, x0 , u0 ) denotes the
derivative of the map (x, u) → f (t, x, u) evaluated at (x0 , u0 ). Analogously, if f (t, ·, ·)
is twice differentiable for all t ∈ [0, 1], we denote the second-order partial derivatives with respect to x and/or u at (t0 , x0 , u0 ) by fxx (t0 , x0 , u0 ), fxu (t0 , x0 , u0 ), and
fuu (t0 , x0 , u0 ). Further, f (t, x0 , u0 ) denotes the Hessian of the map (x, u) → f (t, x, u)
evaluated at (x0 , u0 ). Second-order derivatives are bilinear maps and for u, v ∈ Rn
we write fxx (t0 , x0 , u0 )uv instead of fxx (t0 , x0 , u0 )(u, v). The same simplification will
be used for fxu and fuu .
When dealing with Lipschitz continuous functions that are not differentiable, we
will use the notion of the Clarke gradient.
Definition 2.1. Consider a function f : Rn → R and a point x ∈ Rn . Assume
that f is Lipschitz continuous on a neighborhood of x. The Clarke gradient of f at x
is the set
∂f (x) := co {ξ ∈ Rn | ∃xi → x, ∇f (xi ) exists and ∇f (xi ) → ξ}.
Let A be a measurable subset of [0, 1]. The characteristic function of A, denoted
by χA : [0, 1] → {0, 1}, is defined by
1 if t ∈ A,
χA (t) :=
0 otherwise.
For an interval I ⊂ R, B(I) denotes the Borel σ-algebra associated with the interval I.
Next, we recall some definitions concerning tangent sets to subsets of Rn . The
distance between a point x ∈ Rn and a subset K ⊂ Rn is defined by dist(x, K) :=
1145
SECOND-ORDER NECESSARY OPTIMALITY CONDITIONS
inf k∈K |x−k|. We denote by TK (x) and CK (x), respectively, the Bouligand and Clarke
tangent cone of K at x. See [2] for their definitions. First- and second-order adjacent
subsets are defined next.
Definition 2.2. Let K be a closed subset of Rn and let x ∈ K. The adjacent
cone to K at x is the set defined by
dist(x + hu, K)
n =0 .
TK (x) := u ∈ R lim
h→0+
h
We recall that a set K is called sleek if the Bouligand and the Clarke tangent
cone to K coincide for all x ∈ K. In this case, also the adjacent tangent cone to K
coincides with the Clarke tangent cone to K for all x ∈ K and the set-valued map
(x) is lower semicontinuous at every point of K. For further properties of
x TK
sleek sets, see, for instance, [2].
Definition 2.3. Let K be a closed subset of Rn . Further, let x ∈ K and u ∈ Rn .
The second-order adjacent subset to K at (x, u) is the set defined by
dist(x + hu + h2 v, K)
(2)
n
=0 .
TK (x, u) := v ∈ R lim
h→0+
h2
(2)
(2)
(x). Furthermore, TK (x, 0) =
Note that TK (x, u) = ∅ implies that u ∈ TK
(2)
TK (x), and finally, if K is convex, then TK (x, u) is convex as well. The set K is
called twice derivable if for all (x, u) the second-order Bouligand and the second-order
adjacent tangent set to K at (x, u) coincide. See, for instance, [2] for more information
about derivable sets. We will also use polar and normal cones.
Definition 2.4. Let K be a closed subset of Rn . The (negative) polar cone to
K is the set defined by
K := {x ∈ Rn | x, k ≤ 0 ∀k ∈ K}.
The normal cone to K at x is the set NK (x) := CK (x) .
For basic properties of tangent sets and normal cones, see, for instance, [2]. Finally, we give a useful regularity property of a subset of Rn .
Definition 2.5. Let K be a closed subset of Rn and let r > 0. We say that K
satisfies the uniform interior sphere property of radius r if for every x ∈ ∂K there
exists yx ∈ Rn such that
B(yx , r) ⊂ K and |yx − x| = r.
The following proposition is an immediate consequence of [13, Prop. 2.2.2].
Proposition 2.6. Let r > 0, y ∈ Rn and x ∈ ∂B(y, r). Then for every u ∈
T∂B(y,r) (x),
dist(x + u, B(y, r)) ≤
|u|2
.
r
2.2. Main assumptions. In this subsection we give a list of the assumptions
that we will use in the remainder of this paper.
Definition 2.7. A process (x̄, ū) is a strong local minimizer if there exists δ > 0
such that (x̄, ū) minimizes J over all feasible processes (x, u) satisfying x̄−x∞ ≤ δ.
Let us consider an admissible process (x̄, ū) for either (P ) or (PN C ). Throughout
the paper we assume the following:
1146
DANIEL HOEHENER
(A1) The set-valued map U is measurable and has closed nonempty images.
(A2) (a) For a.e. t ∈ [0, 1], ∀u ∈ U (t), ∀x ∈ x̄(t) + δB, f (·, x, u) and l(·, x, u) are
measurable and f (t, ·, ·), l(t, ·, ·) are continuously differentiable.
(b) For all R > 0, there exists kR ∈ L1 ((0, 1); R+ ) such that for a.e. t ∈ [0, 1],
f (t, ·, u) and l(t, ·, u) are kR (t)-Lipschitz on RB for all u ∈ U (t).
(c) There exists γ ∈ L1 ((0, 1); R+ ) such that for a.e. t ∈ [0, 1] and all x ∈ Rn ,
|f (t, x, u)| ≤ γ(t)(1 + |x|)
∀u ∈ U (t).
(d) For all R > 0, the mapping t → sup |x|≤R |l(t, x, u)| is integrable on
u∈U(t)∩RB
[0, 1].
For the rest of the paper, we are going to abbreviate (t, x̄(t), ū(t)) by [t]; thus, for
instance, f [t] := f (t, x̄(t), ū(t)).
Some of the results of this paper will require additional assumptions. In order to
state them, let (x̄, ū) be an admissible process of either (P ) or (PN C ) and let δ > 0.
(H1) (a) fu [·] ∈ L1 ((0, 1); R+ ) and ∃a1 ∈ L1 ((0, 1); R+ ) such that for a.e. t ∈
[0, 1],
fu (t, x, u) − fu (t, y, v) ≤ a1 (t)(|x − y| + |u − v|)
∀x, y ∈ x̄(t) + δ B̊, ∀u, v ∈ ū(t) + δ B̊.
(b) For a.e. t ∈ [0, 1] ∀x ∈ x̄(t) + δB, f (t, x, U (t)) is a closed set.
(c) For all t ∈ [0, 1], f (t, ·, ·) is twice differentiable on (x̄(t)+δ B̊)×(ū(t)+δ B̊).
(d) f [·] ∈ L1 ((0, 1); R+ ) and ∃a2 ∈ L1 ((0, 1); R+ ) such that for a.e. t ∈
[0, 1],
f (t, x, u) − f (t, y, v) ≤ a2 (t)(|x − y| + |u − v|)
∀x, y ∈ x̄(t) + δ B̊, ∀u, v ∈ ū(t) + δ B̊
(H2) (a) For all t ∈ [0, 1], f (t, ·, ·) is twice differentiable on (x̄(t)+δ B̊)×(ū(t)+δ B̊).
(b) The mapping b1 : [0, 1] → R+ defined by
b1 (t) := f [t] + f [t]
is integrable on [0, 1].
(c) There exists b2 ∈ L1 ((0, 1); R+ ) such that for a.e. t ∈ [0, 1],
f (t, x, u)−f (t, y, v)+f (t, x, u)−f (t, y, v) ≤ b2 (t)(|x−y|+|u−v|)
∀x, y ∈ x̄(t) + δ B̊, ∀u, v ∈ ū(t) + δ B̊.
3. Main results. In this section we present the main results of the present work.
The proofs rely on the variational equations stated in section 4 and will be given in
section 6.
3.1. Second-order necessary optimality conditions for problem (PN C ).
In this subsection we present second-order necessary optimality conditions for the
problem (PN C ). We start with first-order necessary conditions, usually called the
maximum principle. The Hamiltonian H : [0, 1] × Rn × Rn × Rm → R is defined
by (1.2).
Theorem 3.1. Let (x̄, ū) be a strong local minimizer for problem (PN C ). Assume
(H1)(b) and that l satisfies the same assumptions as f in (H1)(b). Then there exists
p ∈ W 1,1 ([0, 1]; Rn ) such that
SECOND-ORDER NECESSARY OPTIMALITY CONDITIONS
1147
(i) −ṗ(t) = Hx (t, x̄(t), p(t), ū(t)) a.e.,
(ii) p(1) = 0,
(iii) H(t, x̄(t), p(t), ū(t)) = maxu∈U(t) H(t, x̄(t), p(t), u) a.e.
Proof. It is well known that by using an additional state variable, the Bolza
problem can be reduced to the Mayer one. Applying this to (PN C ) we get a problem
for which [51, Prop. 6.4.4] applies. By using this proposition, the statement of the
theorem can be easily deduced.
Let p be the adjoint state variable from the maximum principle. Below we will use
the abbreviation [t] also for (t, x̄(t), p(t), ū(t)). The precise meaning will be given by
the context. We introduce the following linear system, which will play an important
role throughout the paper:
ẏ(t) = fx [t]y(t) + fu [t]u(t) a.e.,
(3.1)
y(0) = 0.
Define A := {t ∈ [0, 1] | ū(t) ∈ ∂U (t), Hu [t] = 0} and
UA := u ∈ L∞ ((0, 1); Rm ) | ∃c > 0, h0 > 0 s.t.
∀h ∈ [0, h0 ], ∀t ∈ A, dist(ū(t) + hu(t), U (t)) ≤ ch2 .
The next theorem was announced in [29]. We will provide its detailed proof in section 6.
Theorem 3.2. Let (x̄, ū) be a strong local minimizer for problem (PN C ). Assume
(H1) and that l fulfills the same assumptions as f in (H1). Let p be as in the maximum
principle of Theorem 3.1. Then ∀u ∈ UA and v ∈ L∞ ((0, 1); Rm ) such that
(u(t)) a.e.
u(t) ∈ TU(t)
(3.2)
either
1
0
Hu [t]u(t)dt < 0 or
(3.3)
0
1
1
0
and
(2)
v(t) ∈ TU(t) (ū(t), u(t)) for a.e. t ∈ A,
Hu [t]u(t)dt = 0 and then the inequality
2Hu [t]v(t) + Hxx [t]y(t)y(t) + 2Hxu [t]y(t)u(t) + Huu [t]u(t)u(t) dt ≤ 0
holds true, where y ∈ W 1,1 ([0, 1]; Rn ) is the solution of the linear system (3.1).
Remark 3.1. Observe that if there exists r > 0 such that U (t) satisfies the
uniform interior sphere property of radius r for almost every t ∈ A, then for every
(ū(t)) and Hu [t]u(t) = 0, we
u ∈ L∞ ((0, 1); Rm ) satisfying for a.e. t ∈ A, u(t) ∈ TU(t)
have that u ∈ UA . Indeed, let t ∈ A. It follows from the maximum principle that
(3.4)
Hu [t]ũ ≤ 0
∀ũ ∈ TU(t)
(ū(t)).
By the uniform interior sphere property there exists yt ∈ Rm such that Bt :=
B(yt , r) ⊂ U (t) and |yt − ū(t)| = r. Since Hu [t] = 0, we deduce from (3.4) that
Hu [t] ∈ TU(t)
(ū(t)) . Then obviously Hu [t] ∈ TB t (ū(t)) . This implies that
TB t (ū(t)) = {λHu [t] | λ ≥ 0}.
Next, since Hu [t]u(t) = 0, we have that u(t) ∈ T∂B
(ū(t)). Furthermore,
t
dist(ū(t) + hu(t), U (t)) ≤ dist(ū(t) + hu(t), Bt ),
and therefore we deduce from Proposition 2.6 that dist(ū(t)+hu(t), U (t)) ≤ 1r u2∞ h2 .
1148
DANIEL HOEHENER
3.2. Second-order necessary optimality conditions for problems with
mixed constraints. Theorem 3.2 can be applied to deduce second-order necessary
conditions for problems with mixed equality constraints. Consider problem (PN C ),
where U (t) = Rm ∀t ∈ [0, 1], and suppose that we have the additional mixed statecontrol equality constraint
(M C)
γ(t, x(t), u(t)) = 0
∀t ∈ [0, 1],
where γ : [0, 1] × Rn × Rm → Rk is given. We denote this problem by (PMC ). Let Q
be an open subset of [0, 1] × Rn × Rm satisfying
Γ := {(t, x, u) ∈ [0, 1] × Rn × Rm | γ(t, x, u) = 0} ⊂ Q.
Before stating the second-order necessary conditions, we introduce the following Hamiltonian:
H(t, x, p, u, b) := p, f (t, x, u) − l(t, x, u) − b, γ(t, x, u).
Analogously to the previous subsection, [t] abbreviates (t, x̄(t), p(t), ū(t), b(t)) as well
as (t, x̄(t), ū(t)). The precise meaning will be given by the context.
Theorem 3.3 (see [29, Thm. 15]). Let (x̄, ū) be a strong local minimizer for
problem (PMC ). Assume the following hypotheses:
(a) f , l, and γ are twice continuously differentiable on Q.
(b) γu satisfies the full rank condition, rank γu (t, x, u) = k ∀(t, x, u) ∈ Γ.
(c) ū ∈ L∞ ((0, 1); Rm ).
Then there exist p ∈ W 1,1 ([0, 1]; Rn ) and an essentially bounded b ∈ L∞ ((0, 1); Rk )
satisfying
(i) −ṗ(t) = Hx (t, x̄(t), p(t), ū(t), b(t)) a.e.,
(ii) p(1) = 0,
(iii) Hu (t, x̄(t), p(t), ū(t), b(t)) = 0 a.e.
and such that the inequality
0
1
Hxx [t]y(t)y(t) + 2Hxu [t]y(t)u(t) + Huu [t]u(t)u(t)dt ≤ 0
holds for every u ∈ L∞ ((0, 1); Rm ) satisfying
γx [t]y(t) + γu [t]u(t) = 0
∀t ∈ [0, 1],
where y ∈ W 1,1 ([0, 1]; Rn ) is the solution of the linear system (3.1).
Remark 3.2. Results of the same nature for problems in the Mayer form can be
found in [38] and [39], where the authors require ū to be piecewise continuous.
Remark 3.3. Assuming that an appropriate full rank condition holds, problems
with mixed inequality constraints can be reduced to problems with mixed equality
constraints; see [50]. Thus, the above theorem can also be used to find second-order
necessary conditions for problems with mixed inequality constraints.
Proof of Theorem 3.3. Similar to [39], the main idea of the proof is to reduce
problem (PMC ) to a problem without the mixed constraints, i.e., a problem to which
Theorem 3.2 can be applied. See [29] for details.
3.3. Second-order necessary optimality conditions for problem (P ). Below we state necessary optimality conditions for problem (P ). We start by presenting
SECOND-ORDER NECESSARY OPTIMALITY CONDITIONS
1149
a normal maximum principle for problem (P ). This result is based on the maximum
principle [51, Thm. 9.5.1] and a normality result due to Bettiol and Frankowska [4].
In this subsection we will make the following assumptions on the set of state
constraints K ⊂ Rn :
r
(SC) (a) There exist g1 , . . . , gr ∈ C 2 (Rn ; R) such that K = i=1 Ki , where for
i = 1, . . . , r, Ki := {x ∈ Rn | gi (x) ≤ 0}.
(b) For all x ∈ ∂K, 0 ∈
/ co {∇gi (x) | i ∈ I(x)}, where I(x) := {i ∈
{1, . . . , r} | gi (x) = 0}.
Remark 3.4. Note that assumption (SC) implies that the sets Ki are of class
1,1
Cloc
. See, for instance, [17, 18] for properties of smooth sets. In particular, it follows
that ∂Ki = gi−1 (0) := {x ∈ Rn | gi (x) = 0} ∀i.
In addition, (SC) implies that the set K is sleek and twice derivable (see [2,
Thm. 4.7.4, Prop. 4.7.5]). Finally, one has the following characterizations of the
tangent and the normal cone:
(x) = {w ∈ Rn | ∇gi (x), w ≤ 0 ∀i ∈ I(x)},
TK
NK (x) = R+ co {∇gi (x) | i ∈ I(x)},
where x ∈ ∂K.
In addition, we require that the set K ⊂ Rn satisfy the following inward pointing
condition:
(IP) For all R > 0, ∃ηR > 0, MR > 0, ρR > 0 such that ∀x ∈ ∂K ∩ RB, ∀z ∈
K ∩ B(x, ηR ), ∀s ∈ [0, 1], ∃uz ∈ U (s) satisfying
|f (s, z, uz )| ≤ MR ,
sup
n∈NK (x)∩S n−1
n, f (s, z, uz ) ≤ −ρR .
The Hamiltonian H is defined by (1.2). We provide the proof of the following theorem
in section 5.
Theorem 3.4. Let x0 ∈ Int K and (x̄, ū) be a strong local minimizer for the
problem (P ). Assume (SC) and (IP). If x̄ is Lipschitz continuous, then there exist
p ∈ W 1,1 ([0, 1]; Rn ), a positive Radon measure μ, and a Borel measurable ν : [0, 1] →
Rn satisfying
ν(t) ∈ NK (x̄(t)) ∩ B,
μ-a.e.,
such that for ψ(t) := [0,t] ν(s)dμ(s) for t ∈ ]0, 1] and q := p + ψ we have
(i) −ṗ(t) = Hx (t, x̄(t), q(t), ū(t)) a.e.,
(ii) q(1) = 0,
(iii) H(t, x̄(t), q(t), ū(t)) = maxu∈U(t) H(t, x̄(t), q(t), u) a.e.,
(iv) supp μ ⊂ {t ∈ [0, 1] | I(x̄(t)) = ∅}.
Furthermore, the following jump conditions hold true:
ψ(t) − ψ(t−) ∈ NK (x̄(t)) ∀t ∈ ]0, 1].
(3.5)
Remark 3.5. It is not difficult to show that there exist positive Radon measures
μi , i = {1, . . . , r}, such that
ψ(t) =
ν(s)dμ(s) =
[0,t]
r i=1
[0,t]
∇gi (x̄(s))dμi (s) ∀t ∈ ]0, 1].
1150
DANIEL HOEHENER
Next we state second-order necessary conditions for problem (P ). First, we introduce some notation. Let (x̄, ū) be an admissible process of problem (P ). We set
H[t] := H(t, x̄(t), q(t), ū(t)), where q = p+ψ is as in Theorem 3.4. Hu [t], Hx [t], Hxx [t],
etc., are defined similarly. The set of tangent directions A(x̄, ū) for problem (P ) at
(x̄, ū) is defined as follows:
A(x̄, ū) := {(y, u) ∈ W 1,1 ([0, 1]; Rn ) × L∞ ((0, 1); Rm ) |
u(t) ∈ TU(t)
(ū(t)) a.e., y is solution of (3.1) s.t. y(t) ∈ TK
(x̄(t)) ∀t ∈ [0, 1]}.
Under all the assumptions of Theorem 3.4, ∀(y, u) ∈ A(x̄, ū) we have
1
0
Hu [t]u(t)dt ≤ 0
and
[0,1]
ν(t), y(t)dμ(t) ≤ 0.
For second-order conditions, we have to use a smaller set of tangents in order to ensure
that the tangent variations to the control remain uniformly bounded:
A(1) (x̄, ū) := (y, u) ∈ A(x̄, ū) ∃c > 0, h0 > 0 s.t. for a.e. t ∈ [0, 1],
dist(ū(t) + hu(t), U (t)) ≤ ch2 ∀h ∈ ]0, h0 ] .
Remark 3.6. Consider the set U ∞ := {u ∈ L∞ ((0, 1); Rm ) | u(t) ∈ U (t) a.e.},
(2)
ū ∈ U ∞ , and u ∈ L∞ ((0, 1); Rm ). If the second-order adjacent set TU ∞ (ū, u) defined
in the L∞ topology is nonempty, then there exist c > 0, h0 > 0 such that for almost
every t ∈ [0, 1],
dist(ū(t) + hu(t), U (t)) ≤ ch2
∀h ∈ [0, h0 ].
Note that the converse statement is not true in general. That is, the above inequality
(2)
does not imply nonemptyness of TU ∞ (ū, u).
(1)
For (y, u) ∈ A (x̄, ū) and v ∈ L∞ ((0, 1); Rm ), consider the following secondorder approximation of the nonlinear control system (1.1):
(3.6)
⎧
ẇ(t) = fx [t]w(t) + fu [t]v(t)
⎪
⎪
⎨
1
1
+ fxx [t]y(t)y(t) + fxu [t]y(t)u(t) + fuu [t]u(t)u(t) a.e.,
⎪
2
2
⎪
⎩
w(0) = 0.
For i ∈ {1, . . . , r}, we set
Ti := {t ∈ [0, 1] | ∃sj → t such that ∀j, gi (x̄(sj )) < 0, ∇gi (x̄(sj )), y(sj ) > 0}.
Then, for t ∈ [0, 1] we define
⎧
⎪
⎪
⎨
ci (t) :=
lim sup
s→t
gi (x̄(s))<0
⎪
(x̄(s)),y(s)>0
∇g
⎪ i
⎩
∇gi (x̄(s)), y(s)2
4|gi (x̄(s))|
if t ∈ Ti ,
0
otherwise.
We say that w ∈ W 1,1 ([0, 1]; Rn ) satisfies (TV) if
SECOND-ORDER NECESSARY OPTIMALITY CONDITIONS
1151
(TV) For all i ∈ {1, . . . , r} and ∀t ∈ [0, 1] such that gi (x̄(t)) = 0 and ∇gi (x̄(t)),
y(t) = 0 we have
1
∇gi (x̄(t)), w(t) + gi (x̄(t))y(t)y(t) + ci (t) ≤ 0.
2
(3.7)
We say that w satisfies (TV) strictly if (3.7) holds with strict inequality.
Remark 3.7. Note that by [34, Thm. 3.2], the condition (TV) is equivalent to the
statement that the map t → g (x̄(t))w(t) + 12 g (x̄(t))y(t)y(t) is in the second-order
adjacent set to the set of nonpositive continuous functions at (g(x̄(·)), g (x̄(·))y(·)).
The set of second-order tangent directions A(2) (x̄, ū, u, y) of problem (P ) at the
process (x̄, ū) and the first-order tangent (y, u) ∈ A(1) (x̄, ū) is defined by
A(2) (x̄, ū, y, u) := (w, v) ∈ W 1,1 ([0, 1]; Rn ) × L∞ ((0, 1); Rm ) (2)
v(t) ∈ TU(t) (ū(t), u(t)) a.e., w is solution of (3.6) s.t. (TV) is satisfied .
We can state now the second-order conditions. Their proof will be given in section 6.
Theorem 3.5. Assume (SC) and (IP), x0 ∈ Int K, and let (x̄, ū) be a strong
local minimizer such that (H2) holds true and x̄ is Lipschitz. Further suppose that l
fulfills the same assumptions as f in (H2). Let q, μ, and ν be as in the maximum
principle of Theorem 3.4 and let (y, u) ∈ A(1) (x̄, ū) be such that
1
Hu [t]u(t)dt = 0 and
ν(t), y(t)dμ(t) = 0.
0
[0,1]
(2)
Let (w̃, ṽ) ∈ A(2) (x̄, ū, y, u) satisfy (TV) strictly and D(t) ⊂ TU(t) (ū(t), u(t)) be convex
sets such that ṽ(t) ∈ D(t) for a.e. t ∈ [0, 1]. Then for every (w, v) ∈ A(2) (x̄, ū, y, u)
with v(t) ∈ D(t) a.e., we have
(3.8)
[0,1]
w(t), ν(t)dμ(t) +
1
+
0
0
1
Hu [t]v(t)dt
1
1
Hxx [t]y(t)y(t) + Hxu [t]y(t)u(t) + Huu [t]u(t)u(t) dt ≤ 0.
2
2
Note that if the set-valued map U (·) has convex images, then one can define
(2)
D(t) = TU(t) (ū(t), u(t)) ∀t.
Second, we would like to underline that the above result is derived for any q, μ,
ν satisfying the constrained maximum principle.
Remark 3.8. Consider (y, u) ∈ A(1) (x̄, ū) and for t ∈ [0, 1] set
B(t) :=
1
1
fxx [t]y(t)y(t) + fxu [t]y(t)u(t) + fuu [t]u(t)u(t).
2
2
The set of second-order active constraints is defined as follows:
I (2) (x̄(t), y(t)) := {i ∈ {1, . . . , r} | gi (x̄(t)) = 0 and ∇gi (x̄(t)), y(t) = 0} .
Observe that every solution w of (3.6) can be represented by w = w1 + w2 , where
w1 ∈ W 1,1 ([0, 1]; Rn ) is the solution of
ẇ1 (t) = fx [t]w1 (t) + B(t) a.e. and w1 (0) = 0,
1152
DANIEL HOEHENER
and w2 ∈ W 1,1 ([0, 1]; Rn ) is the solution of
(3.9)
ẇ2 (t) = fx [t]w2 (t) + fu [t]v(t)
a.e. and w2 (0) = 0.
For t ∈ [0, 1], we define the reachable set at t by
(2)
R(t) := w(t) w solves (3.9), v ∈ L∞ ((0, 1); Rm ) s.t. v(s) ∈ TU(s) (ū(s), u(s)) a.e. .
Consequently, if there exists a unique t ∈ [0, 1] such that I (2) (x̄(t), y(t)) = ∅, then
there exists (w, v) ∈ A(2) (x̄, ū, y, u) satisfying (TV) strictly if and only if
inf
r∈R(t)
max
1
∇gi (x̄(t)), r + w1 (t) + gi (x̄(t))y(t)y(t) + ci (t) < 0.
2
i∈I (2) (x̄(t),y(t))
It would be interesting to study more general conditions for the existence of pairs
(w, v) ∈ A(2) (x̄, ū, y, u) satisfying (TV) strictly. This problem, however, is out of the
scope of this paper.
3.4. Examples. We conclude this section with two examples. In the first example we consider the case without pure state constraints and we compare Theorem 3.2
with other known results. The second example illustrates the application of Theorem 3.5.
Example 1. We consider problem (PN C ), where n = m = 2, x = (x1 , x2 )T ,
u = (u1 , u2 )T , x0 = 0, and f : R4 → R2 , l : R4 → R are defined by f (x1 , x2 , u1 , u2 ) =
(u1 , u1 + u2 )T , l(x1 , x2 , u1 , u2 ) = −x1 x2 + u21 − u2 , respectively. The set-valued map
U : [0, 1] R2 is time independent,
U (t) := {(u1 , u2 ) ∈ R2 | |u1 | ≤ 1, 0 ≤ u2 ≤ 1 and g(u1 , u2 ) ≤ 0} ∀t ∈ [0, 1],
where g : R2 → R is given by g(u1 , u2 ) = −u21 + u2 .
Consider the feasible process (x̄, ū) defined by (x̄(t), ū(t)) = (0, 0) ∀t ∈ [0, 1].
Clearly this process is not optimal. In what follows, we show that one can use the
second-order necessary conditions of Theorem 3.2 in order to prove that (x̄, ū) is not
a strong local minimum. We will also show that the necessary conditions stated in
[27] and [43] do not allow a similar conclusion.
First, note that for the above data all our assumptions are satisfied and the statements of Theorem 3.2 and Theorem 3.5 are equivalent. The Hamiltonian H : R6 → R
is defined by the following:
H(x1 , x2 , p1 , p2 , u1 , u2 ) = p1 u1 + p2 (u1 + u2 ) + x1 x2 − u21 + u2 .
Thus, the adjoint equation of the maximum principle corresponding to (x̄, ū) = 0 is
ṗ(t) = 0 and p(1) = 0.
Consequently, p ≡ 0 and it is not difficult to verify that
sup H(x̄(t), p(t), u) = H(x̄(t), p(t), ū(t)) = 0
u∈U(t)
∀t ∈ [0, 1];
hence the maximum principle holds true.
Consider u(t) = (1, 0)T ∀t ∈ [0, 1]. Then A = [0, 1], u ∈ UA , and v(t) := (0, 1)T ∈
(2)
TU(t) (ū(t), u(t)) ∀t ∈ [0, 1]. The linear system (3.1) for u(t) is
ẏ(t) = (1
1)T
and y(0) = 0.
1153
SECOND-ORDER NECESSARY OPTIMALITY CONDITIONS
Thus y(t) = (t, t)T . The partial derivatives of the Hamiltonian are as follows:
−2 0
0 1
, Hxx [t] =
, and Hxu [t] = 0.
Hu [t] = (0 1), Huu [t] =
0 0
1 0
Then,
0
1
1
0
Hu [t]u(t)dt = 0 and
2Hu [t]v(t) + Hxx [t]y(t)y(t) + 2Hxu y(t)u(t) + Huu [t]u(t)u(t) dt =
1
2t2 dt > 0,
0
which contradicts the second-order conditions (3.3).
On the other hand, observe that U (t) = {(u1 , u2 ) ∈ R2 | g̃(u1 , u2 ) ∈ Q}, where
g̃(u1 , u2 ) = (−u21 + u2
u1 − 1
− u2 )T
− u1 − 1 u2 − 1
and Q = R5− .
However, [43, Thm. 4.1] cannot be applied since the required normality condition is
not satisfied by the data.
We show next that necessary optimality conditions of [27, Thm. 3.1] are satisfied
by (x̄, ū). Applying these conditions to our problem, we find again that p(·) ≡ 0 and
the Hamiltonian H : R6 → R is defined by
H(x1 , x2 , p1 , p2 , u1 , u2 ) = p1 u1 + p2 (u1 + u2 ) − x1 x2 + u21 − u2 .
It is obvious that all u ∈ Ũ, where Ũ is as in (1.4), must satisfy u(t) ∈ R × {0}.
1
Consequently, ∀u ∈ Ũ , 0 Hu [t]u(t)dt = 0, i.e., the first-order conditions are satisfied.
Further, the linearized system considered in [27] is
1 0
u1 (t)
y1 (0)
0
ẏ1 (t)
=
and y(0) =
=
,
ẏ(t) =
ẏ2 (t)
u2 (t)
y2 (0)
1 1
0
where u(·) = (u1 (·), u2 (·))T ∈ U. The second-order conditions derived in [27] are as
follows:
(3.10)
1
1
Hxx [t]y(t)y(t)+ 2Hxu [t]y(t)u(t)+ Huu [t]u(t)u(t)dt =
−2y1 (t)2 + 2u1 (t)2 dt ≥ 0,
0
0
where u1 ∈ L∞ ((0, 1); R) and y1 (t) =
1
that y1 (t)2 ≤ t 0 u1 (s)2 ds. Hence,
0
1
2y1 (t)2 dt ≤
0
1
t
0
u1 (s)ds. By the Hölder inequality we have
tdt ·
1
0
2u1 (t)2 dt ≤
0
1
2u1 (t)2 dt.
Thus (3.10) holds true and consequently the necessary optimality conditions from
[27, Thm. 3.1] are satisfied. Therefore we cannot deduce that the process (x̄, ū) is not
optimal.
Example 2. We consider problem (P ), where again n = m = 2, and we use the
same notation as in the previous example. Let C, M > 0 be given constants and
assume that the dynamics of the control system are as follows:
2Cu1
u2
if t ≤ 12 ,
if t ≤ 12 ,
and f2 (t, x, u) =
f1 (t, x, u) =
(u1 + u2 )x1 + u2 otherwise,
−x2 otherwise,
1154
DANIEL HOEHENER
where f (t, x, u) = (f1 (t, x, u), f2 (t, x, u))T . The cost is given by
−M x1 − 2Cu1
if t ≤ 12 ,
l(t, x, u) =
2
2
2
2
u1 (x1 + x2 ) + u2 − x2 otherwise.
The control constraints are given by the set-valued map U : [0, 1] R2 defined by
U (t) = U = [−1, 1] × [−1, 1] ∀t ∈ [0, 1].
In order to formulate the state constraints we define the map g : R2 → R4 by
T
g(x) = (g1 (x), g2 (x), g3 (x), g4 (x))T = (x1 , −x1 − 2C, −C − x2 , x2 − C) ,
and we set K = {x ∈ R2 | g(x) ≤ 0}. Finally we assume that x0 = (−C, 0)T . It is
easy to see that this data satisfies all the assumptions of Theorem 3.5.
Consider the reference control ū = (ū1 , ū2 )T given by
1 if t ≤ 12 ,
and ū2 (t) ≡ 0.
ū1 (t) =
0 otherwise,
This leads to the following reference trajectory x̄ = (x̄1 , x̄2 )T :
−C + 2Ct if t ≤ 12 ,
and x̄2 (t) ≡ 0.
x̄1 (t) =
0
otherwise,
(x̄, ū) is obviously a feasible process. Next, we are going to show that it is an extremal.
First-order optimality conditions. First, notice that NK (x̄(t)) ∩ B = [0, 1] ×
{0} ∀t ≥ 1/2 and I(x̄(t)) = ∅ ∀t ∈ [0, 1/2[. Thus if we set
1
(0, 0)T if t < 12 ,
ν(t) ≡
, μ = δ{ 1 } , and ψ(t) =
2
0
(1, 0)T otherwise,
then ν(t) ∈ NK (x̄(t)) ∩ B μ-a.e. The corresponding solution p of the adjoint equation
(see Theorem 3.4(i)) for q := p + ψ is given by
1
1
−M t + M
−M t + M
2 − 1 if t ≤ 2 ,
2 − 1 if t < 2 ,
q1 (t) =
p1 (t) =
−1
otherwise,
0
otherwise,
and p2 (t) = q2 (t) ≡ 0. It remains to check the maximum principle. If t < 1/2, then
M
H(t, x̄(t), q(t), u) = 2q1 (t)Cu1 +q2 (t)u2 +M x̄1 (t)+2Cu1 = 2C
− M t u1 +M x̄1 (t).
2
Hence, maxu∈U(t) H(t, x̄(t), q(t), u) = H[t]. If t > 1/2, then
H(t, x̄(t), q(t), u) = −q1 (t)(u1 x̄1 (t) + u2 x̄1 (t) + u2 ) − q2 (t)x̄2 (t)
− u1 (x̄1 (t)2 + x̄2 (t)2 ) − u22 + x̄2 (t)2
= −u22 ≤ 0.
Thus the maximum principle holds true for this choice of p, ν, μ and consequently
(x̄, ū) is an extremal. We show next, using the second-order necessary conditions,
that it is not a strong local minimizer.
SECOND-ORDER NECESSARY OPTIMALITY CONDITIONS
1155
Second-order necessary conditions. We consider the following first-order control
variations:
√
2 4 + M if t ≤ 12 ,
u1 (t) ≡ 0 and u2 (t) =
0
otherwise.
It is obvious that u(t) ∈ TU (ū(t)) a.e. Furthermore, this control variation leads to the
following solution of the linearized system (3.1):
√
2t 4 + M
if t ≤ 12 ,
y1 (t) ≡ 0 and y2 (t) = −t+ 1 √
2
e
4 + M otherwise.
Clearly, y(t) ∈ TK
(x̄(t)) ∀t ∈ [0, 1]. Furthermore, it is easy to see that
[0,1]
ν(t), y(t)dμ(t) = 0
and
0
1
Hu [t]u(t)dt = 0,
and, consequently, the direction (y, u) is critical. Let us now consider the secondorder control variation v(·) and the corresponding solution w(·) of the second-order
approximation (3.6),
T
1
v(t) ≡ − , 0
C
(−2t, 0)T
and w(t) =
(−1, 0)T
if t ≤ 12 ,
otherwise.
(2)
Then v(t) ∈ TU (ū(t), u(t)). The only active constraint is g1 (x) ≤ 0. However, since
y1 (t) ≡ 0, it follows immediately that ci (t) ≡ 0. In addition, w satisfies (TV) strictly
since
1
∇g1 (x̄(t)), w(t) + g1 (x̄(t))y(t)y(t) = w1 (t) = −1 ∀t ∈ [1/2, 1].
2
On the other hand, the second-order inequality (3.8) does not hold because
[0,1]
w(t), ν(t)dμ(t) +
0
1
Hu [t]v(t)dt
1
1
Hxx [t]y(t)y(t) + Hxu [t]y(t)u(t) + Huu [t]u(t)u(t) dt
2
2
0
M
M
1
1
M
M
+ 2+
= 0.
= −1 −
1−
>
2+
−1−
4
2
e
2
2
4
1
+
We conclude that the extremal (x̄, ū) is not a strong local minimizer.
4. Variational equations. In real analysis, second-order necessary conditions
for a local minimum of a twice continuously differentiable function ϕ : Rn → R are
valid for critical points, i.e., points x ∈ Rn , where the first-order necessary condition
ϕ (x) = 0 holds, and are based on the fact that if the critical point x is a local
minimum then for all perturbations δx ∈ Rn with |δx| small enough, the inequality
ϕ(x + δx) − ϕ(x) ≥ 0
holds. The same approach can be used to derive second-order necessary conditions
for optimal control problems. However, there are three questions to be answered:
1156
DANIEL HOEHENER
1. What are the admissible perturbations of controls under control constraints?
2. What are the associated perturbations of trajectories?
3. How do we deal with problems involving state constraints?
The purpose of this section is to answer these questions.
It is natural to investigate perturbations of the control, since they imply corresponding perturbations of the state trajectories. Forgetting the state constraints for a
moment, an intuitive choice for the control perturbations is essentially bounded maps
that are “tangent” to the set of admissible controls. This intuition is confirmed by
the following proposition, which is a slight modification of [2, Thm. 8.5.1].
Proposition 4.1. Consider ū ∈ U and let u ∈ L∞ ((0, 1); Rm ) be such that
(ū(t)) a.e. Then ∀h > 0, ∃uh ∈ L∞ ((0, 1); Rm ) satisfying
u(t) ∈ TU(t)
(i) uh ∞ ≤ 2u∞,
(ii) ū(t) + huh (t) ∈ U (t) a.e.,
a.e.
(iii) uh −→ u, when h → 0+.
We would like to underline that in general the statement of Proposition 4.1 is not
true when we replace the adjacent tangent cone by the (larger) Bouligand tangent
cone.
Proof. The proof is a simple modification of the proof of [2, Thm. 8.5.1] and very
similar to the proof of Proposition 4.2 below and therefore it is omitted.
In order to obtain second-order conditions, we need second-order perturbations.
For second-order tangents we have the following result, similar to Proposition 4.1.
Proposition 4.2. Consider ū ∈ U and let u, v ∈ L∞ ((0, 1); Rm ) be such that
(a) ∃c, h0 > 0, such that ∀h ∈ ]0, h0 ], dist(ū(t) + hu(t), U (t)) ≤ h2 c a.e.,
(2)
(b) v(t) ∈ TU(t) (ū(t), u(t)) a.e.
Then, ∀h ∈ ]0, h0 ], ∃vh ∈ L∞ ((0, 1); Rm ) satisfying
(i) vh ∞ ≤ 2v∞ + c,
(ii) ū(t) + hu(t) + h2 vh (t) ∈ U (t) a.e.,
a.e.
(iii) vh −→ v, when h → 0+.
(2)
(ū(t)), then TU(t) (ū(t), u) = Rm .
Remark 4.1. If U (t) is sleek and u ∈ Int TU(t)
Indeed, for sleek U (t), u ∈ Int TU(t)
(ū(t)) implies by [48, Thm. 2] that ∃ > 0 such
that
ū(t) + [0, ]B(u, ) ⊂ U (t),
from which we deduce readily the statement.
Remark 4.2. It is possible to prove that if for a.e. t ∈ [0, 1] the set U (t) is
twice derivable, then assumption (a) is a sufficient condition for the nonemptyness of
(2)
TU(t) (ū(t), u(t)).
Remark 4.3. It is easy to verify that any pair (u, v) ∈ L∞ ((0, 1); Rm )2 with
(ū(t)) a.e. satisfies the assumptions of Proposition 4.2 and that
u ≡ 0 and v(t) ∈ TU(t)
in this case, the statement of Proposition 4.2 is a consequence of Proposition 4.1.
However, as the following example shows, in general there are other possible choices
than u ≡ 0.
Example. Let U = {(x, y) ∈ R2 | x2 ≤ y}, ū = (ū1 , ū2 ) = (1, 1). Then, u =
(u1 , u2 ) = (1, 2) and v = (v1 , v2 ) = (0, 1) are such that
(2)
v ∈ TU (ū, u) and dist(ū + hu, U ) ≤ h2 .
Proof. Let x → ϕ(x) := x2 . Then ϕ (ū1 ) = 2 and ϕ (ū1 ) = 2. First, we show
that ∀hn → 0+ there exists a sequence vn → v such that ū + hn u + h2n vn ∈ U . By
1157
SECOND-ORDER NECESSARY OPTIMALITY CONDITIONS
the Taylor formula,
1
ϕ(ū1 + hn u1 ) = ϕ(ū1 ) + hn ϕ (ū1 )u1 + h2n ϕ (ū1 )u21 + Rn
2
= ū2 + hn u2 + h2n v2 + Rn ,
where Rn /h2n → 0, when n → ∞. Hence, by defining vn = (0, v2 +
the desired result. On the other hand,
Rn
h2n ),
we obtain
dist(ū + hu, U ) ≤ |ϕ(ū1 + hu1 ) − ū2 − hu2 | ≤ h2 .
This completes the proof.
Proof of Proposition 4.2. Let h ∈ ]0, h0 ]. We define
ah (t) := dist(ū(t) + hu(t) + h2 v(t), U (t)).
Then ah is measurable. Further, there exists a measurable yh such that yh (t) ∈ U (t) ∀t
and such that
ah (t) = |ū(t) + hu(t) + h2 v(t) − yh (t)|.
Similarly there exists a measurable zh such that zh (t) ∈ U (t) ∀t and such that
bh (t) := |ū(t) + hu(t) − zh (t)| = dist(ū(t) + hu(t), U (t)).
It follows from assumption (a) that for almost every t in [0, 1], bh (t) ≤ h2 c. Thus, by
definition of ah , we have obviously
ah (t) ≤ |ū(t) + hu(t) + h2 v(t) − zh (t)|≤ h2 c + h2 |v(t)|
a.e. in [0, 1].
Therefore,
ah (t)
≤ c + v∞
h2
a.e. in [0, 1].
Next, we define vh : [0, 1] → Rm by
vh (t) :=
yh (t) − hu(t) − ū(t)
.
h2
With this definition we find
|v(t) − vh (t)| =
ah (t)
,
h2
∀t ∈ [0, 1].
Therefore, for almost every t ∈ [0, 1], |vh (t)| ≤ 2v∞ + c and consequently vh ∈
L∞ ((0, 1); Rm ). Further, it follows directly from the definition of vh that ū(t) +
(2)
hu(t) + h2 vh (t) ∈ U (t) a.e. Finally, since v(t) ∈ TU(t) (ū(t), u(t)) a.e., ah (t)/h2 → 0
a.e., when h → 0+.
The two previous propositions provide an answer to the first of the three questions
stated at the beginning of this section for the case of unconstrained state trajectories.
In the following, we remain in this particular case and respond to the second question.
First, let us recall the following well-known fact, which can be found, for example,
in [1, pp. 255–256].
1158
DANIEL HOEHENER
Proposition 4.3 (variational equation). Let (x̄, ū) be an admissible process for
(PN C ) and suppose that assumption (H1)(a) is satisfied. Let u ∈ L∞ ((0, 1); Rm ) be
(ū(t)) and let y ∈ W 1,1 ([0, 1]; Rn ) be
such that for almost every t ∈ [0, 1], u(t) ∈ TU(t)
the solution of the linearized control system (3.1) corresponding to u. For all h > 0,
let uh be as in Proposition 4.1 and let xh ∈ W 1,1 ([0, 1]; Rn ) be the solution of
ẋh (t) = f (t, xh (t), ū(t) + huh (t)) a.e.,
xh (0) = x0 .
Then,
xh − x̄ unif
−→ y
h
when h → 0 + .
Remark 4.4. It follows from assumption (A2) and the Picard–Lindelöf theorem
that ∀h > 0, xh exists and is unique.
Remark 4.5. The hypothesis (H1)(a) could be replaced by the following weaker
hypothesis:
∃ku ∈ L1 ((0, 1); R+ ) such that for a.e. t ∈ [0, 1] and ∀x ∈ x̄(t) + δB,
f (t, x, ·) is ku (t)-Lipschitz on ū(t) + δB.
Next, we present a second-order analogue of Proposition 4.3. To the best of our
knowledge, this is a new result dealing with a second-order variational equation.
Proposition 4.4 (second-order variational equation). Let (x̄, ū) be an admissible process for (PN C ) and assume that (H2) is satisfied. Let u, v ∈ L∞ ((0, 1); Rm )
be such that assumptions (a) and (b) of Proposition 4.2 are verified, and let y ∈
W 1,1 ([0, 1]; Rn ) be the solution of the linear system (3.1). For all h ∈ ]0, h0 ], let vh
be as in Proposition 4.2 and let xh ∈ W 1,1 ([0, 1]; Rn ) be the solution of
ẋh (t) = f (t, xh (t), ū(t) + hu(t) + h2 vh (t)) a.e.,
(4.1)
xh (0) = x0 .
Then,
xh − x̄ − hy unif
−→ w
h2
when h → 0+,
where w ∈ W 1,1 ([0, 1]; Rn ) is the solution of the second-order approximation (3.6).
Proof. We define zh ∈ W 1,1 ([0, 1]; Rn ) by
zh (t) = x̄(t) + hy(t) + h2 w(t)
and Rh : [0, 1] → Rn by
(4.2)
Rh (t) = żh (t) − f (t, x̄(t) + hy(t) + h2 w(t), ū(t) + hu(t) + h2 vh (t)).
Step 1. We prove that ∀h > 0 small enough, the map Rh (·)/h2 is integrably
bounded by a mapping independent from h.
In order to simplify the notation, we define
f [t, h, τ ] := f (t, x̄(t) + τ y(t) + τ 2 w(t), ū(t) + τ u(t) + τ 2 vh (t)).
fx [t, h, τ ], fu [t, h, τ ], fxx [t, h, τ ], fxu [t, h, τ ], and fuu [t, h, τ ] are defined in a similar
way.
SECOND-ORDER NECESSARY OPTIMALITY CONDITIONS
1159
Before we start the actual estimation of Rh (·), observe that
d
f [t, h, τ ] = fx [t, h, τ ](y(t) + 2τ w(t)) + fu [t, h, τ ](u(t) + 2τ vh (t)).
dτ
Therefore,
h
fx [t, h, s] (y(t) + 2sw(t)) + fu [t, h, s] (u(t) + 2svh (t)) ds = f [t, h, h] − f [t].
(4.3)
0
Also,
d
fx [t, h, τ ]y(t) = fxx [t, h, τ ]y(t) (y(t) + 2τ w(t)) + fxu [t, h, τ ]y(t) (u(t) + 2τ vh (t)) ,
dτ
d
fu [t, h, τ ]u(t) = fuu [t, h, τ ]u(t) (u(t) + 2τ vh (t)) + fux [t, h, τ ]u(t) (y(t) + 2τ w(t)) .
dτ
Thus,
s
fxx [t, h, τ ]y(t) (y(t) + 2τ w(t)) + fxu [t, h, τ ]y(t) (u(t) + 2τ vh (t)) dτ
0
(4.4)
= fx [t, h, s]y(t) − fx [t]y(t),
s
0
fuu [t, h, τ ]u(t) (u(t) + 2τ vh (t)) + fux [t, h, τ ]u(t) (y(t) + 2τ w(t)) dτ
= fu [t, h, s]u(t) − fu [t]u(t).
From (4.3) and (4.4) we deduce that
Rh (t) = h2 ẇ(t) −
h
−
0
0
0
0
h
−
s
0
h
fx [t, h, s]2sw(t) + fu [t, h, s]2svh (t) ds
fxx [t, h, τ ]y(t) (y(t) + 2τ w(t)) + fxu [t, h, τ ]y(t) (u(t) + 2τ vh (t)) dτ ds
s
fuu [t, h, τ ]u(t) (u(t) + 2τ vh (t)) + fux [t, h, τ ]u(t) (y(t) + 2τ w(t)) dτ ds.
Using this, assumption (H2), the fact that max{w∞ , u∞ , 2v∞ + c} < ∞, and
the fact that vh ∞ ≤ 2v∞ + c ∀h > 0 small enough, we deduce that for some
C > 0 independent from t and ∀h > 0 small enough,
|Rh (t)| ≤ h2 |ẇ(t)| + h2 C(b1 (t) + b2 (t)),
and consequently we have shown that ∀h > 0 small enough, |Rh (·)|/h2 is integrably
bounded by a function that is independent from h.
Step 2. We prove that
Rh (t) = h2 fu [t](v(t) − vh (t)) + o(t, h2 ) a.e.,
where |o(t, h2 )|/h2 → 0 when h → 0+.
Indeed, it suffices to consider the second-order Taylor development of
f (t, x̄(t) + hy(t) + h2 w(t), ū(t) + hu(t) + h2 vh (t)).
1160
DANIEL HOEHENER
Then the statement follows readily from the definitions of y and w. By Proposition 4.2
a.e.
we know that vh −→ v when h → 0+ and therefore,
|Rh (t)|
−−−−→ 0
h→0+
h2
(4.5)
a.e.
Step 3. We show that |xh (t) − zh (t)|/h2 converges uniformly to 0.
Using assumption (A2), it is not difficult to show that there exist r > 0 and
kr ∈ L1 ((0, 1); R+ ) such that
|f (t, xh (t), ū(t) + hu(t) + h2 vh (t)) − f (t, zh (t), ū(t) + hu(t) + h2 vh (t))|
≤ kr (t)|xh (t) − zh (t)|
∀h > 0 small enough for a.e. t ∈ [0, 1]. Then, since by Step 1 Rh (·) is integrable, we
deduce from (4.2) that
t
|f (t, xh (s), ū(s) + hu(s) + h2 vh (s)) − żh (t)|ds
|xh (t) − zh (t)| ≤
0
≤
t
0
kr (s)|xh (s) − zh (s)| + |Rh (s)|ds.
Hence, by the Gronwall lemma, ∃M > 0, independent from h, such that
t
|xh (t) − zh (t)| ≤ M
|Rh (s)|ds ∀t ∈ [0, 1].
0
Using this and (4.5), it follows from the first step and the Lebesgue dominated convergence theorem that
1
|Rh (t)|
|xh (t) − zh (t)|
dt ≥ lim sup
0 = lim M
2
h→0+
h→0+
h
h2
t∈[0,1]
0
xh (t) − x̄(t) − hy(t)
= lim sup − w(t) .
2
h→0+
h
t∈[0,1]
Consequently (xh − x̄ − hy)/h2 converges to w uniformly when h → 0+.
The above result answers the second question. In the state constrained case, we
are going to show that the trajectories xh constructed as in Proposition 4.4 remain in
the set K if K satisfies (SC) and if we impose that the solution w of the second-order
approximation (3.6) satisfies (TV) strictly.
Proposition 4.5. Let (x̄, ū) be a feasible process for (P ). Assume that (H2) and
(SC) are satisfied. Let u, v, y, w, and xh be as in Proposition 4.4. In addition assume
that
(4.6)
(x̄(t))
y(t) ∈ TK
∀t ∈ [0, 1]
and that ∀i ∈ {1, . . . , r} and ∀t ∈ [0, 1] such that
gi (x̄(t)) = 0
and
∇gi (x̄(t)), y(t) = 0,
we have
(4.7)
1
∇gi (x̄(t)), w(t) + gi (x̄(t))y(t)y(t) + ci (t) < 0,
2
SECOND-ORDER NECESSARY OPTIMALITY CONDITIONS
1161
where ci (·) is defined as in section 3.3. Then there exists η > 0 such that ∀h ∈ ]0, η],
xh (t) ∈ K ∀t ∈ [0, 1].
Proposition 4.5 can be proved using relation (4.13) below and [42, Lemma 3.3]
(see also [33] and [41]). For the sake of completeness we provide its direct proof.
Proof. It suffices to prove that ∀i ∈ {1, . . . , r}, there exists η > 0 such that
∀h ∈ ]0, η],
gi (xh (t)) ≤ 0 ∀t ∈ [0, 1].
(4.8)
Therefore we fix an arbitrary i ∈ {1, . . . , r} and prove that (4.8) is satisfied.
We start by stating some implications of the assumptions and introducing some
unif
notation. First, note that since xh −→ x̄, when h → 0+, for some constant C > 0,
C := (x̄∞ + C) B xh (t) ∀t ∈ [0, 1]
for all small h > 0. Since C is compact and gi ∈ C 2 (Rn ; R), gi is Lipschitz on C with
Lipschitz constant Lgi .
Define BT i := {t ∈ [0, 1] | gi (x̄(t)) = 0}. By Remark 3.4 and (4.6), ∀t ∈ BT i ,
∇gi (x̄(t)), y(t) ≤ 0.
(4.9)
As in the proof of Proposition 4.4, we define
zh (t) = x̄(t) + hy(t) + h2 w(t)
(4.10)
∀t ∈ [0, 1].
unif
Obviously, zh −→ x̄ when h → 0+. Hence, we have that zh (t) ∈ C ∀t ∈ [0, 1] and for
all small h > 0. Therefore, the following inequality holds:
(4.11)
gi (xh (t)) ≤ gi (zh (t)) + Lgi zh − xh ∞ .
Also, by Proposition 4.4,
(4.12)
zh − xh ∞ /h2 → 0 when h → 0 + .
Thus, using the Taylor formula and (4.10)–(4.12), we obtain that ∀h > 0 small enough
and ∀t ∈ [0, 1],
(4.13)
gi (xh (t)) ≤ gi (x̄(t)) + h∇gi (x̄(t)), y(t)
1 2
+h
g (x̄(t))y(t)y(t) + ∇gi (x̄(t)), w(t) + o(h2 ),
2 i
where o(h2 )/h2 → 0, when h → 0+ and o(h2 ) is independent from t.
Our first goal is to show that (4.8) holds on (BT i + 1 B̊) ∩ [0, 1] for some 1 > 0.
We consider t ∈ N := {s ∈ BT i | ∇gi (x̄(s)), y(s) = 0}. We claim that there exist
t > 0 and ht > 0 such that ∀h ∈ ]0, ht ],
(4.14)
gi (xh (s)) ≤ 0 ∀s ∈ ]t − t , t + t [ ∩ [0, 1].
Indeed, assume to the contrary that the claim is false. Then there exist hn > 0
converging to 0 and a sequence S := {sn }n∈N converging to t when n → ∞ such that
(4.15)
gi (xhn (sn )) > 0 ∀n.
1162
DANIEL HOEHENER
We distinguish two cases. First assume that the sequence S has a subsequence
S̃ := {s̃n } with corresponding subsequence {h̃n } such that ∀s ∈ S̃,
gi (x̄(s)) < 0 and ∇gi (x̄(s)), y(s) > 0.
(4.16)
We show that this leads to a contradiction with (4.15). For all s ∈ S̃ and all h > 0
we have
1 gi (x̄(s))y(s)y(s) + ∇gi (x̄(s)), w(s) + o(h2 )
gi (x̄(s)) + h∇gi (x̄(s)), y(s) + h2
2
1 o(h2 )
∇gi (x̄(s)), y(s)2
gi (x̄(s))y(s)y(s) + ∇gi (x̄(s)), w(s) −
+
= h2
2
4gi (x̄(s))
h2
2
h∇gi (x̄(s)), y(s)
.
+ gi (x̄(s)) 1 +
2gi (x̄(s))
≤0
Thus, it follows from (4.13) that ∀s ∈ S˜ and ∀h > 0 small enough,
gi (xh (s))
1 o(h2 )
∇gi (x̄(s)), y(s)2
gi (x̄(s))y(s)y(s) + ∇gi (x̄(s)), w(s) +
+
≤ h2
.
2
4|gi (x̄(s))|
h2
Then, by (4.7), there exists δ̃ > 0 such that for some ñ0 ∈ N,
∀n ≥ ñ0
gi (xh (s̃n )) ≤ h2 −δ̃ + o(h2 )/h2
∀h > 0 small enough. Let ñ1 ∈ N be such that o(h̃2n )/h̃2n ≤ δ̃/2 ∀n ≥ ñ1 . Then,
gi (xh̃n (s̃n )) ≤ −h̃2n
δ̃
≤0
2
∀n ≥ max{ñ0 , ñ1 },
which contradicts (4.15).
Consider now the case when (4.16) is not satisfied for s = sn and all large n.
Then there exists a subsequence Ŝ := {ŝn } such that ∀s ∈ Ŝ either gi (x̄(s)) = 0 and
thus by (4.9), ∇gi (x̄(s)), y(s) ≤ 0 or gi (x̄(s)) < 0 and ∇gi (x̄(s)), y(s) ≤ 0. In both
cases it is obvious that ∀s ∈ Ŝ,
gi (x̄(s)) + h∇gi (x̄(s)), y(s) ≤ 0
∀h > 0.
Hence by (4.13) ∀h > 0 small enough,
gi (xh (s)) ≤ h
2
1 o(h2 )
gi (x̄(s))y(s)y(s) + ∇gi (x̄(s)), w(s) +
2
h2
∀s ∈ Ŝ.
Finally, using (4.7), continuity, and the fact that ci (t) ≥ 0, we find that there exists
δ̂ > 0 such that for some n̂0 ∈ N,
gi (xh (ŝn )) ≤ h2 −δ̂ + o(h2 )/h2
∀n ≥ n̂0
SECOND-ORDER NECESSARY OPTIMALITY CONDITIONS
1163
∀h > 0 small enough. Analogously to the previous case, we denote the subsequence
corresponding to Ŝ by {ĥn }. Thus, choosing n̂1 > 0 such that o(ĥ2n )/ĥ2n ≤ δ̂/2 ∀n ≥
n̂1 , we obtain that
gi (xĥn (ŝn )) ≤ −ĥ2n
δ̂
≤0
2
∀n ≥ max{n̂0 , n̂1 },
which also contradicts (4.15), and consequently we have established that (4.14) holds.
Let us now consider t ∈ BT i \ N . By the definition of N and (4.9), there exists
δt > 0 such that
∇gi (x̄(t)), y(t) ≤ −δt ,
and therefore there exists t > 0 such that ∀s ∈ ]t − t , t + t [ ∩ [0, 1],
∇gi (x̄(s)), y(s) ≤ −
δt
.
2
Then, by (4.13),
gi (xh (s)) ≤ −h
δt
+ o(h)
2
∀s ∈ ]t − t , t + t [ ∩ [0, 1],
where o(h)/h → 0 when h → 0+ and o(h) is independent from s. Consequently, we
can find ht > 0 such that ∀h ∈ ]0, ht ],
(4.17)
gi (xh (s)) ≤ −
hδt
≤0
4
∀s ∈ ]t − t , t + t [ ∩ [0, 1].
Next we show that there exist 1 > 0, h1 > 0 such that (4.8) holds on (BT i +
1 B̊) ∩ [0, 1] ∀h ∈ ]0, h1 ]. Thanks to the compactness of BT i , (4.14), and (4.17), we
know that there exists a finite family T := {t1 , . . . , tq } such that
BT i ⊂
q t
t
]tk − k , tk + k [ ∩ [0, 1]
2
2
k=1
and such that ∀k ∈ {1, . . . , q},
(4.18)
gi (xh (s)) ≤ 0
tk
2
> 0 and h1 := mink htk > 0. We claim that ∀h ∈ ]0, h1 ],
gi (xh (t)) ≤ 0 ∀t ∈ BT i + 1 B̊ ∩ [0, 1].
Then we set 1 := mink
(4.19)
∀s ∈ ]tk − tk , tk + tk [, h ∈ ]0, htk ].
Indeed, let t ∈ (BT i + 1 B̊) ∩ [0, 1]. Then there exists tk ∈ T such that |t − tk | ≤
1 + tk /2 ≤ tk . Then the claim follows from (4.18), since htk ≥ h1 .
It remains to prove that (4.8) holds on M := [0, 1] \ (BT i + 1 B̊). Since M is
compact, there exists δ > 0 such that
gi (x̄(t)) ≤ −δ ∀t ∈ M.
Then, using the uniform convergence of xh to x̄, when h → 0+, we deduce the
existence of h2 > 0 such that ∀h ∈ ]0, h2 ],
xh − x̄∞ ≤
δ
.
Lgi
1164
DANIEL HOEHENER
Consequently, ∀h ∈ ]0, h2 ],
(4.20)
gi (xh (t)) ≤ gi (x̄(t)) + Lgi xh − x̄∞ ≤ 0
∀t ∈ M.
Using (4.19) and (4.20), it follows for η := min{h1 , h2 } that ∀h ∈ ]0, η], (4.8)
holds.
5. First-order necessary conditions. In this section we provide a proof of
Theorem 3.4. Further, we will use this theorem to deduce the relation (1.3). Finally,
we state a condition similar to (1.3) for problem (PN C ).
We start with the proof of Theorem 3.4. All notation and definitions are as in
section 3.3.
Proof of Theorem 3.4. We start by proving a (not necessarily normal) maximum
principle and then deduce normality from a theorem due to Bettiol and Frankowska
[4].
Let us define g : Rn → R by
g(x) =
max
i∈{1,...,r}
gi (x)
∀x ∈ Rn .
With this definition, g is locally Lipschitz. For λ ≥ 0, let us define the unnormalized
Hamiltonian Hλ : [0, 1] × Rn × Rn × Rm → R as follows:
Hλ (t, x, p, u) = p, f (t, x, u) − λl(t, x, u).
Using a standard argument, we deduce from [51, Thm. 9.5.1] that there exist p ∈
W 1,1 ([0, 1]; Rn ), λ ≥ 0, a positive Radon measure μ̃, and a Borel measurable function
ν̃ : [0, 1] → Rn satisfying
(5.1)
ν̃(t) ∈ ∂g(x̄(t)),
μ̃-a.e.,
such that
(i) (p, μ̃, λ) = 0,
(ii) supp μ̃ ⊂ {t | g(x̄(t)) = 0},
(iii) −ṗ(t) = Hxλ (t, x̄(t), q̃(t), ū(t)) a.e.,
(iv) q̃(1) = 0,
(v) Hλ (t, x̄(t), q̃(t), ū(t)) = maxu∈U(t) Hλ (t, x̄(t), q̃(t), u) a.e.,
where q̃(t) = p(t) + ψ̃(t) with
ν̃(s)dμ̃(s) for t ∈ [0, 1[,
ψ̃(t) := [0,t[
ν̃(s)dμ̃(s) for t = 1.
[0,1]
Next, we show that (iii)–(v) remain true if we replace q̃(t) by q(t) = p(t) + ψ(t), where
ψ ∈ N BV ([0, 1]; Rn ) and
ν̃(s)dμ̃(s) ∀t ∈ ]0, 1].
(5.2)
ψ(t) =
[0,t]
Indeed, since ν̃ is bounded and μ̃ is a finite measure, ψ̃ is a function of bounded
variation. This implies that ψ̃ has a countable number of discontinuities. Thus we
can define
⎧
ψ̃(t)
if ψ̃ is continuous at t ∈ ]0, 1[,
⎪
⎪
⎨
ψ̃(t+) if ψ̃ is discontinuous at t ∈ ]0, 1[,
ψ(t) :=
0
if t = 0,
⎪
⎪
⎩
ψ̃(1)
if t = 1.
SECOND-ORDER NECESSARY OPTIMALITY CONDITIONS
1165
Note that with this definition ψ satisfies (5.2) and is of bounded variation. In addition,
it is right continuous and vanishes at zero. Hence ψ ∈ N BV ([0, 1]; Rn ) and ψ̃(t) =
ψ(t) a.e., because they differ only at a countable number of points, i.e., on a set of
zero Lebesgue measure. Therefore, if we define q(t) = p(t) + ψ(t) as above, then also
q̃(t) = q(t) a.e. and consequently (iii)–(v) are also satisfied for q.
Let us now show that there exist a Borel measurable ν : [0, 1] → Rn and a positive
Radon measure μ such that
(5.3) ν(t) ∈ NK (x̄(t)) ∩ B, μ-a.e.
and
ψ(t) =
ν(s)dμ(s) ∀t ∈ ]0, 1].
[0,t]
First, we show that (5.1) implies that
ν̃(t) ∈ NK (x̄(t)),
(5.4)
μ̃-a.e.
Indeed, it follows directly from the maximum rule [14, Prop. 2.3.12] that
∂g(x̄(t)) = co {∇gi (x̄(t)) | gi (x̄(t)) = g(x̄(t))}
∀t ∈ [0, 1].
For t ∈ [0, 1] such that g(x̄(t)) = 0, this condition can be rewritten
∂g(x̄(t)) = co {∇gi (x̄(t)) | i ∈ I(x̄(t))} ⊂ NK (x̄(t)).
The above set inclusion is a consequence of Remark 3.4. Thus, since supp μ̃ ⊂ {t ∈
[0, 1] | g(x̄(t)) = 0}, (5.4) holds true. Finally, because there exists M > 0 such that
ν̃(t) ≤ M μ̃-a.e., we define
∀t ∈ [0, 1], ν(t) =
Then ∀t ∈ ]0, 1],
ν̃(t)
,
M
and ∀S ∈ B([0, 1]), μ(S) = M μ̃(S).
ν(s)dμ(s) =
[0,t]
ν̃(s)dμ̃(s) = ψ(t),
[0,t]
and we have
ν(t) ∈ NK (x̄(t)) ∩ B,
μ-a.e.
Next, let us prove that ψ satisfies the jump conditions (3.5). By [25, Prop. 2.5], we
know that if the set-valued map t NK (x̄(t)) has a closed graph, then ψ satisfies
the jump conditions. Let us therefore prove that t NK (x̄(t)) has a closed graph.
Note that since K is sleek, the set-valued map CK (·) is lower semicontinuous and
(·) = CK (·). Now consider a sequence (tj , yj ) → (t, y) such that yj ∈ NK (x̄(tj )) ∀j.
TK
We have to show that y ∈ NK (x̄(t)).
(x̄(t)). Since TK
(·) is lower semicontinuous,
For this, consider an arbitrary v ∈ TK
for every neighborhood V of v there exists η > 0 such that ∀x ∈ K ∩ B(x̄(t), η),
TK
(x ) ∩ V = ∅.
In particular, it follows that yj , w ≤ 0 for j such that x̄(tj ) ∈ B(x̄(t), η),
yj ∈ NK (x̄(tj )), and w ∈ TK
(x̄(tj )) ∩ V . Since the neighborhood V can be chosen arbitrarily small, by continuity of the scalar product we find y, v ≤ 0.
(x̄(t)), we have established
Finally, because the above is true for arbitrary v ∈ TK
that y ∈ NK (x̄(t)) and hence NK (x̄(·)) has a closed graph.
It remains to show that λ can be chosen equal to 1. This follows directly from [4,
Cor. 3.6] if we can show that the following regularity condition is verified:
1166
DANIEL HOEHENER
For any x ∈ ∂K and > 0 there exists ηx > 0 such that ∀y, z ∈
K ∩ B(x, ηx ),
sup
n∈NK (y)∩S n−1
n, z − y ≤ |z − y|.
To prove this, let x ∈ ∂K and > 0 be fixed. Then ∀i = 1, . . . , r, by continuity of
∇gi (·), there exists an ηxi > 0 such that ∀x ∈ B(x, ηxi ) we have
|∇gi (x ) − ∇gi (x)| ≤
(5.5)
.
2
Define ηx := mini ηxi . Then, consider an arbitrary pair y, z ∈ K ∩ B(x, ηx ). We can
assume that y ∈ ∂K; otherwise NK (y) ∩ S n−1 = ∅ and the supremum equals, by
convention, −∞. Further, y = z; otherwise the statement is trivial. Fix i ∈ I(y). As
gi (y) = 0 we find
gi (z) − gi (y) ≤ 0.
(5.6)
Then, by the mean-value theorem, there exists ξ ∈ B(x, ηx ) such that
gi (z) − gi (y) = ∇gi (ξ), z − y.
(5.7)
It follows from the definition of ηx and (5.5)–(5.7) that
∇gi (y), z − y ≤ ∇gi (y) − ∇gi (ξ), z − y ≤ |z − y|.
This allows us to conclude, since NK (y) = R+ co {∇gi (y) | i ∈ I(y)} and 0 ∈
/
co {∇gi (y) | i ∈ I(y)}. The proof is complete.
The remainder of this section is dedicated to statement (1.3) from the introduction. Let ū ∈ M([0, 1]; Rm ) and u ∈ L∞ ((0, 1); Rm ). The directional derivative of
J(·) at ū in the direction u is defined by
J(ū + hu) − J(ū)
.
h→0+
h
J (ū)u := lim
Further, the set of tangent directions A(x̄, ū) for problem (P ) at (x̄, ū) is defined as
in section 3.3. We have the following corollary of Theorem 3.4.
Corollary 5.1. Let x0 ∈ Int K and let (x̄, ū) be a strong local minimizer
for problem (P ). Suppose that (SC) and (IP) are satisfied, f satisfies (H1)(a), and
l satisfies the same assumptions as f in (H1)(a). If x̄ is Lipschitz continuous and
p, μ, ν, ψ are as in Theorem 3.4, then ∀(y, u) ∈ A(x̄, ū),
1
Hu (t, x̄(t), p(t) + ψ(t), ū(t))u(t)dt −
ν(t), y(t)dμ(t) ≥ 0.
J (ū)u = −
0
[0,1]
Proof. It suffices to show that
1
Hu (t, x̄(t), p(t) + ψ(t), ū(t))u(t)dt −
J (ū)u = −
0
[0,1]
ν(t), y(t)dμ(t).
a.e.
For h > 0, let uh be such that {uh } is bounded in L∞ ((0, 1); Rm ) and uh −→ u. By
Proposition 4.1 such sequence uh does exist. Then, by Proposition 4.3, the solution
xh ∈ W 1,1 ([0, 1]; Rn ) of
ẋh (t) = f (t, xh (t), ū(t) + huh (t)) a.e.,
xh (0) = x0
SECOND-ORDER NECESSARY OPTIMALITY CONDITIONS
1167
satisfies
yh :=
xh − x̄ unif
−→ y
h
when h → 0 + .
Hence, using the above definitions and the Lipschitz continuity of f and l, we have
that ∀(y, u) ∈ A(x̄, ū),
J (ū)u = lim
(5.8)
h→0+
0
1
l(t, x̄(t) + hyh (t), ū(t) + huh (t)) − l[t] + R(h, t)
dt,
h
where for almost every t ∈ [0, 1], R(h, t)/h → 0 when h → 0+. It is not difficult to
show that the integrand on the right-hand side of (5.8) is bounded by an integrable
function that is independent of h. Thus, the Lebesgue dominated convergence theorem
applies and it remains to show that
1
l(t, x̄(t) + hyh (t), ū(t) + huh (t)) − l[t]
dt
h→0+
h
1
=−
Hu (t, x̄(t), p(t) + ψ(t), ū(t))u(t)dt −
lim
0
0
[0,1]
ν(t), y(t)dμ(t).
Indeed, this can be proved using the definition of the Hamiltonian, the Taylor formula,
and integration by parts. See the beginning of the proof of Theorem 3.5 in section 6
for details.
It is obvious that the same line of argument leads to the following statement for
the case of problem (PN C ).
Corollary 5.2. Let (x̄, ū) be a strong local minimizer for problem (PN C ). Assume that (H1)(a)–(b) are satisfied and that l satisfies the same assumptions as f in
(A1), (A2), and (H1)(a)–(b). Let p ∈ W 1,1 ([0, 1]; Rn ) be as in the maximum principle.
Then ∀u ∈ L∞ ((0, 1); Rm ) satisfying u(t) ∈ TU(t)
(ū(t)) a.e.,
J (ū)u = −
0
1
Hu (t, x̄(t), p(t), ū(t))u(t)dt ≥ 0.
6. Second-order necessary conditions. In this section, we provide proofs of
Theorems 3.2 and 3.5. We start with the proof of Theorem 3.2. We use the same
notation as in section 3.1. In addition, we introduce the critical set CN C (x̄, ū) for
problem (PN C ) at the process (x̄, ū) which is defined as follows:
CN C (x̄, ū) :=
(y, u, v) ∈ W 1,1 ([0, 1]; Rn ) × UA × L∞ ((0, 1); Rm ) (3.1) and (3.2) are satisfied and
0
1
Hu [t]u(t)dt = 0 .
The critical set has the property that if (y, u, v) ∈ CN C (x̄, ū) and λ > 0, then
(λy, λu, λ2 v) ∈ CN C (x̄, ū).
Proof of Theorem 3.2. Since u(t) ∈ TU(t)
(ū(t)) a.e., by Corollary 5.2 we know
1
that 0 Hu [t]u(t)dt ≤ 0. Thus, all we have to prove is that the inequality (3.3) holds
∀(y, u, v) ∈ CN C (x̄, ū). Fix (y, u, v) ∈ CN C (x̄, ū).
1168
DANIEL HOEHENER
First step. Let χA denote the characteristic function of A. We prove that ∀h > 0,
a.e.
a.e.
there exist uh , vh ∈ L∞ ((0, 1); Rm ) such that uh −→ u, vh −→ χA v when h → 0+
and such that ∀h > 0 small enough,
(6.1)
(6.2)
ū(t) + huh (t) + h2 vh (t) ∈ U (t)
uh ∞ ≤ 2u∞ ,
(6.3)
vh ∞ ≤ 2v∞ + c.
a.e.,
(ū(t)) a.e., there exists a sequence ûh ∈
By Proposition 4.1, since u(t) ∈ TU(t)
L∞ ((0, 1); Rm ), converging a.e. to u, when h → 0+ and such that ∀h > 0,
ū(t) + hûh (t) ∈ U (t) for a.e. t ∈ [0, 1],
ûh ∞ ≤ 2u∞.
With this, one defines uh as follows:
u(t)
uh (t) :=
ûh (t)
if t ∈ A,
otherwise.
(2)
Further, since 0 ∈ TU(t) (ū(t), 0) a.e. and u ∈ UA , the pair (χA u, χA v) satisfies
all the assumptions of Proposition 4.2. Hence, there exists a sequence of essentially
bounded maps vh such that ∀h > 0 small enough
a.e.
vh −→ χA v,
vh ∞ ≤ 2v∞ + c,
ū(t) + hu(t) + h2 vh (t) ∈ U (t)
for a.e. t ∈ A.
Hence, vh , uh satisfy (6.1)–(6.3).
Second step. We prove that ∀h > 0, the solution xh ∈ W 1,1 ([0, 1]; Rn ) of
ẋh (t) = f (t, xh (t), ū(t) + huh (t) + h2 vh (t)) a.e.,
xh (0) = x0 ,
satisfies
xh − x̄ unif
−→ y
h
(6.4)
when h → 0 + .
To show this, we set ũh := uh + hvh . Then, ∀h > 0 small enough, the sequence
{ũh }h>0 satisfies the hypotheses of Proposition 4.3. Hence, we can deduce (6.4).
Third step. We show that ∀h > 0,
(6.5) J(ū + huh + h2 vh ) − J(ū)
1
Hx [t]hyh (t) + H[t] − H(t, xh (t), p(t), ū(t) + huh (t) + h2 vh (t)) dt,
=
0
where yh := (xh − x̄)/h. This is a direct consequence of integration by parts of
1
p(t), f (t, xh (t), ū(t) + huh (t) + h2 vh (t)) − f [t]dt,
0
and the maximum principle (Theorem 3.1).
1169
SECOND-ORDER NECESSARY OPTIMALITY CONDITIONS
unif
Last step. Since xh −→ x̄, when h → 0+ and x̄ is a strong local minimizer, we
deduce from (6.5) that
(6.6)
0≤
1
0
Hx [t]hyh (t) + H[t] − H(t, xh (t), p(t), ū(t) + huh (t) + h2 vh (t)) dt
∀h > 0 small enough. Then we use the Taylor formula and the results of the second
and the third steps in order to deduce (3.3) from (6.6). This is very similar to what
is done in the proof of Theorem 3.5 below and therefore we omit the details of this
step.
Next we are going to prove Theorem 3.5. The proof is based on the first-order conditions from Theorem 3.4 and the second-order variational equation, Proposition 4.5.
We use the same notation as in section 3.3. In addition, we define the set of critical
directions C(x̄, ū) of problem (P ) at the process (x̄, ū) by
C(x̄, ū) :=
(y, u) ∈ A(1) (x̄, ū) 0
1
Hu [t]u(t)dt +
[0,1]
ν(t), y(t)dμ(t) = 0 ,
where μ and ν are as in the maximum principle (Theorem 3.4). Finally, let us recall
the following integration by parts formula, which follows from [10, Thm. 6]: If f ∈
W 1,1 ([0, 1]; Rn ) and ϕ ∈ N BV ([0, 1]; Rn ), then
[0,1]
f (t)dϕ(t) = f (1), ϕ(1) −
0
1
ϕ(s), f (s)ds.
Proof of Theorem 3.5. Fix (y, u) ∈ C(x̄, ū). Note that it follows from the maximum
principle that
Hu [t]u(t) = 0
(6.7)
a.e.
and
ν(t), y(t) = 0
μ-a.e.
Let (w̃, ṽ) ∈ A(2) (x̄, ū, y, u) be such that w̃ satisfies (TV) strictly. First we show
that (3.8) holds for (w̃, ṽ). By Propositions 4.4 and 4.5, for h > 0, there exist w̃h ∈
unif
a.e.
W 1,1 ([0, 1]; Rn ) and ṽh ∈ L∞ ((0, 1); Rm ) such that w̃h −→ w̃ and ṽh −→ ṽ when
h → 0+ and ∀h > 0 small enough,
(x̃h , ũh ) := (x̄ + hy + h2 w̃h , ū + hu + h2 ṽh )
is an admissible process of (P ). Since (x̄, ū) is a strong local minimizer, ∀h > 0 small
enough, the inequality
J(ũh ) − J(ū) ≥ 0
holds true. Using the very definition of the Hamiltonian, this can be reformulated as
follows:
1
(6.8)
0
H[t] − H(t, x̃h (t), q(t), ũh (t)) + q(t), f (t, x̃h (t), ũh (t)) − f [t] dt ≥ 0.
1170
DANIEL HOEHENER
Integrating by parts, we find
1
p(t), f (t, x̃h (t), ũh (t)) − f [t]dt
0
= p(1), hy(1) + h2 w̃h (1) +
(6.9)
0
1
1
0
ψ(t), f (t, x̃h (t), ũh (t)) − f [t]dt
= ψ(1), hy(1) + h2 w̃h (1) −
Hx [t] hy(t) + h2 w̃h (t) dt,
[0,1]
ν(t), hy(t) + h2 w̃h (t)dμ(t).
Then, since p(1) + ψ(1) = q(1) = 0, we deduce from (6.7)–(6.9) that
(6.10)
0
1
H[t] + Hx [t] hy(t) + h2 w̃h (t) − H(t, x̃h (t), q(t), ũh (t)) dt
−
ν(t), h2 w̃h (t)dμ(t) ≥ 0
[0,1]
∀h > 0 small enough.
Next, we are going to show that the map
1
(H(t, x̃h (t), q(t), ũh (t)) − H[t] − Hx [t]hy(t))
h2
is integrably bounded by a map which does not depend from h. In order to do this,
we use the same principle as in the proof of Proposition 4.4. Set
H[t, τ, h] := H(t, x̄(t) + τ y(t) + τ 2 w̃h (t), q(t), ū(t) + τ u(t) + τ 2 ṽh (t)).
Hx [t, τ, h], Hu [t, τ, h], Hxx [t, τ, h], etc., are defined in a similar way. This leads to the
following:
1
(H(t, x̃h (t), q(t), ũh (t)) − H[t] − Hx [t]hy(t))
h2
h
1
= 2
(Hx [t, s, h]2sw̃h (t) + Hu [t, s, h]2sṽh (t)) ds
h 0
h s
1
+ 2
Hxx [t, τ, h]y(t) (y(t) + 2τ w̃h (t)) + Hxu [t, τ, h]y(t) (u(t) + 2τ ṽh (t)) dτ ds
h 0 0
h s
1
+ 2
Huu [t, τ, h]u(t) (u(t) + 2τ ṽh (t)) + Hux [t, τ, h]u(t) (y(t) + 2τ w̃h (t)) dτ ds.
h 0 0
Then, using the assumptions, it is not difficult to see that the map on the right-hand
side is integrably bounded by a function independent from h. Furthermore, by the
Taylor formula we have
H[t, h, h] − H[t] − Hx [t] hy(t) + h2 w̃h (t)
1
= Hu [t]h2 ṽh (t) + Hxx [t]h2 y(t)y(t) + Hxu [t]h2 y(t)u(t)
(6.11)
2
1
+ Huu [t]h2 u(t)u(t) + Rh (t),
2
1171
SECOND-ORDER NECESSARY OPTIMALITY CONDITIONS
where Rh (t)/h2 → 0 when h → 0+. Therefore, since w̃h converges uniformly to w̃
when h → 0+, we can apply Lebesgue’s dominated convergence Theorem to (6.10):
1
1 lim 2 H[t] + Hx [t] hy(t) + h2 w̃h (t) − H[t, h, h] dt
0 h→0+ h
−
lim ν(t), w̃h (t)dμ(t)
(6.12)
1
h→0+ h2
= lim
0
[0,1] h→0+
1
H[t] + Hx [t] hy(t) + h2 w̃h (t) − H[t, h, h] dt
−
2
[0,1]
ν(t), h w̃h (t)dμ(t)
≥ 0.
Consequently, (3.8) can be deduced from (6.11) and (6.12).
(2)
Now let D(t) ⊂ TU(t) (ū(t), u(t)) be convex subsets such that ṽ(t) ∈ D(t) for
almost every t ∈ [0, 1] and consider (w, v) ∈ A(2) (x̄, ū, x, u) with v(t) ∈ D(t) a.e. For
λ ∈ ]0, 1[, we define
wλ := λw̃ + (1 − λ)w
and v λ := λṽ + (1 − λ)v.
It is obvious that (wλ , v λ ) ∈ A(2) (x̄, ū, y, u) and wλ satisfies (TV) strictly. Hence,
∀λ ∈ ]0, 1[,
(6.13)
[0,1]
wλ (t), ν(t)dμ(t) +
+
0
1
0
1
Hu [t]v λ (t)dt
1
1
Hxx [t]y(t)y(t) + Hxu [t]y(t)u(t) + Huu [t]u(t)u(t) dt ≤ 0.
2
2
unif
a.e.
Finally, since obviously wλ −→ w and v λ −→ v when λ → 0 and {v λ }λ is bounded in
L∞ ((0, 1); Rm ), we deduce (3.8) from (6.13) by taking the limit when λ →
0+.
Acknowledgments. The author is grateful to H. Frankowska for providing very
helpful advice and to anonymous referees for their constructive comments.
REFERENCES
[1] M. Athans and P. L. Falb, Optimal Control, An Introduction to the Theory and its Applications, McGraw-Hill, New York, 1966.
[2] J. P. Aubin and H. Frankowska, Set-Valued Analysis, Systems Control Found. Appl.,
Birkhäuser Boston, Boston, MA, 1990.
[3] A. Ben-Tal, Second order and related extremality conditions in nonlinear programming, J.
Optim. Theory Appl., 31 (1980), pp. 143–165.
[4] P. Bettiol and H. Frankowska, Normality of the maximum principle for nonconvex constrained Bolza problems, J. Differential Equations, 243 (2007), pp. 256–269.
[5] J. F. Bonnans, R. Cominetti, and A. Shapiro, Second order optimality conditions based on
parabolic second order tangent sets, SIAM J. Optim., 9 (1999), pp. 466–492.
[6] J. F. Bonnans and A. Hermant, No-gap second-order optimality conditions for optimal control problems with a single state constraint and control, Math. Program., 117 (2009),
pp. 21–50.
[7] J. F. Bonnans and A. Hermant, Revisiting the analysis of optimal control problems with
several state constraints, Control Cybernet., 38 (2009), pp. 1021–1052.
1172
DANIEL HOEHENER
[8] J. F. Bonnans and A. Hermant, Second-order analysis for optimal control problems with
pure state constraints and mixed control-state constraints, Ann. Inst. H. Poincaré Anal.
Non Linéaire, 26 (2009), pp. 561–598.
[9] J. F. Bonnans and A. Shapiro, Perturbation Analysis of Optimization Problems, Springer
Ser. Oper. Res., Springer, New York, 2000.
[10] H. E. Bray, Elementary properties of the Stieltjes integral, Ann. of Math. (2), 20 (1919),
pp. 177–186.
[11] A. E. Bryson, W. F. Denham, and S. E. Dreyfus, Optimal programming problems with
inequality constraints I: Necessary conditions for extremal solutions, AIAA J., 1 (1963),
pp. 2544–2550.
[12] A. E. Bryson and Y.-C. Ho, Applied Optimal Control, Taylor & Francis, Bristol, PA, 1975.
[13] P. Cannarsa and C. Sinestrari, Semiconcave Functions, Hamilton-Jacobi Equations and
Optimal Control, Birkhäuser Boston, Boston, MA, 2004.
[14] F. H. Clarke, Optimization and Nonsmooth Analysis, Classics in Appl. Math. 5, SIAM,
Philadelphia, 1990.
[15] R. Cominetti, Metric regularity, tangent sets and second order optimality conditions, Appl.
Math. Optim., 21 (1990), pp. 265–287.
[16] M. R. de Pinho and J. F. Rosenblueth, Mixed constraints in optimal control: An implicit
function approach, IMA J. Math. Control Inform., 24 (2007), pp. 197–218.
[17] M. C. Delfour and J.-P. Zolésio, Shape analysis via oriented distance functions, J. Funct.
Anal., 123 (1994), pp. 129–201.
[18] M. C. Delfour and J.-P. Zolésio, Shapes and Geometries: Analysis, Differential Calculus
and Optimization, SIAM, Philadelphia, 2001.
[19] A. V. Dmitruk, Quadratic problems for a Pontryagin minimum in an optimum control problem
linear in the control I: A decoding theorem, Izv. Math., 28 (1987), pp. 275–303.
[20] A. V. Dmitruk, Second order optimality conditions for singular extremals, in Computational
Optimal Control, R. Bulirsch and D. Kraft, eds., Birkhäuser, Basel, 1994, pp. 71–81.
[21] A. V. Dmitruk, Quadratic order conditions of a local minimum for abnormal extremals, Nonlinear Anal., 30 (1997), pp. 2439–2448.
[22] A. V. Dmitruk, Jacobi type conditions for singular extremals, Control Cybernet., 37 (2008),
pp. 285–306.
[23] A. Y. Dubovitskii and A. A. Milyutin, Extremum problems with constraints, Dokl. Akad.
Nauk, 149 (1963), pp. 759–762.
[24] A. V. Fiacco and G. P. McCormick, Nonlinear Programming: Sequential Unconstrained
Minimization Techniques, Classics in Appl. Math. 4, SIAM, Philadelphia, 1990.
[25] H. Frankowska, Regularity of minimizers and of adjoint states in optimal control under state
constraints, J. Convex Anal., 13 (2006), pp. 299–328.
[26] R. V. Gamkrelidze, Optimal processes with bounded phase coordinates, Izv. Math., 24 (1960),
pp. 315–356.
[27] E. G. Gilbert and D. S. Bernstein, Second-order necessary conditions in optimal control:
Accessory-problem results without normality conditions, J. Optim. Theory Appl., 41 (1983),
pp. 75–106.
[28] M. Hestenes, Calculus of Variations and Optimal Control, John Wiley, New York, 1964.
[29] D. Hoehener, Second-order optimality conditions for a Bolza problem with mixed constraints,
in Proceedings of the 18th IFAC World Congress, S. Bittanti, A. Cenedese, and S. Zampieri,
eds., 2011, pp. 2594–2599.
[30] A. D. Ioffe, Necessary and sufficient conditions for a local minimum 3: Second order conditions and augmented duality, SIAM J. Control Optim., 17 (1979), pp. 266–288.
[31] D. H. Jacobson and M. M. Lele, New necessary conditions of optimality for control problems
with state-variable inequality constraints, J. Math. Anal. Appl., 35 (1971), pp. 255–284.
[32] H. Kawasaki, An envelope-like effect on infinitely many inequality constraints on second-order
necessary conditions for minimization problems, Math. Program., 41 (1988), pp. 73–96.
[33] H. Kawasaki, The upper and lower second order directional derivatives of a sup-type function,
Math. Program., 41 (1988), pp. 327–339.
[34] H. Kawasaki, Second order necessary optimality conditions for minimizing a sup-type function,
Math. Program., 49 (1991), pp. 213–229.
[35] D. G. Luenberger, Optimization by Vector Space Methods, John Wiley, New York, 1969.
[36] Y. Maruyama, Second-order necessary conditions for nonlinear optimization problems in Banach spaces and their application to an optimal control problem, Math. Oper. Res., 15
(1990), pp. 467–482.
[37] A. A. Milyutin, On a certain family of optimal control problems with phase constraint, J.
Math. Sci. (New York), 100 (2000), pp. 2564–2571.
SECOND-ORDER NECESSARY OPTIMALITY CONDITIONS
1173
[38] A. A. Milyutin and N. P. Osmolovskii, Calculus of Variations and Optimal Control, American Mathematical Society, Providence, RI, 1998.
[39] N. P. Osmolovskii, Quadratic extremality conditions for broken extremals in the general problem of the calculus of variations, J. Math. Sci. (New York), 123 (2004), pp. 3987–4122.
[40] Z. Páles and V. Zeidan, First- and second-order necessary conditions for control problems
with constraints, Trans. Amer. Math. Soc., 346 (1994), pp. 421–453.
[41] Z. Páles and V. Zeidan, Nonsmooth optimum problems with constraints, SIAM J. Control
Optim., 32 (1994), pp. 1476–1502.
[42] Z. Páles and V. Zeidan, Optimum problems with certain lower semicontinuous set-valued
constraints, SIAM J. Optim., 8 (1998), pp. 707–727.
[43] Z. Páles and V. Zeidan, Optimal control problems with set-valued control and state constraints, SIAM J. Optim., 14 (2003), pp. 334–358.
[44] Z. Páles and V. Zeidan, Strong local optimality conditions for state constrained control problems, J. Global Optim., 28 (2004), pp. 363–377.
[45] Z. Páles and V. Zeidan, First- and second-order optimality conditions for a strong local
minimum in control problems with pure state constraints, Nonlinear Anal., 67 (2007),
pp. 2506–2526.
[46] J.-P. Penot, Optimality conditions in mathematical programming and composite optimization,
Math. Program., 67 (1994), pp. 225–245.
[47] H. M. Robbins, Junction phenomena for optimal control with state-variable inequality constraints of third order, J. Optim. Theory Appl., 31 (1980), pp. 85–99.
[48] R. T. Rockafellar, Clarke’s tangent cones and the boundaries of closed sets in Rn , Nonlinear
Anal., 3 (1979), pp. 145–154.
[49] G. Stefani and P. Zezza, Optimality conditions for a constrained control problem, SIAM J.
Control Optim., 34 (1996), pp. 635–659.
[50] F. A. Valentine, The problem of Lagrange with differential inequalities as added side conditions, in Contributions to the Calculus of Variations 1933-37, University of Chicago Press,
1937, pp. 407–448.
[51] R. B. Vinter, Optimal Control, Birkhäuser, Boston, MA, 2000.
[52] J. Warga, A second-order Lagrangian condition for restricted control problems, J. Optim.
Theory Appl., 24 (1978), pp. 475–483.
[53] J. Warga, A second-order condition that strengthens Pontryagin’s maximum principle, J.
Differential Equations, 28 (1979), pp. 284–307.
[54] V. Zeidan and P. Zezza, The conjugate point condition for smooth control sets, J. Math.
Anal. Appl., 132 (1988), pp. 572–589.