PDF

Weak Dynamic Programming for Generalized
State Constraints
βˆ—
Bruno Bouchard
Marcel Nutz
†
October 19, 2012
Abstract
We provide a dynamic programming principle for stochastic optimal
control problems with expectation constraints. A weak formulation,
using test functions and a probabilistic relaxation of the constraint,
avoids restrictions related to a measurable selection but still implies
the Hamilton-Jacobi-Bellman equation in the viscosity sense. We treat
open state constraints as a special case of expectation constraints and
prove a comparison theorem to obtain the equation for closed state
constraints.
Keywords weak dynamic programming, state constraint, expectation constraint,
Hamilton-Jacobi-Bellman equation, viscosity solution, comparison theorem
AMS 2000 Subject Classications 93E20, 49L20, 49L25, 35K55
1 Introduction
We study the problem of stochastic optimal control under state constraints.
In the most classical case, this is the problem of maximizing an expected
reward, subject to the constraint that the controlled state process has to
remain in a given subset
π’ͺ
of the state space.
There is a rich literature
on the associated partial dierential equations (PDEs), going back to [13,
14, 9, 16, 17] in the rst order case and [12, 11, 10] in the second order
case. The connection between the control problem and the equation is given
by the dynamic programming principle. However, in the stochastic case, it
is frequent practice to make this connection only formally, and in fact, we
are not aware of a generally applicable, rigorous technique in the literature.
Of course, there are specic situations where it is indeed possible to avoid
proving the state-constrained dynamic programming principle; in particular,
penalization arguments can be useful to reduce to the unconstrained case
(e.g., [11]). We refer to [6, 18] for further background.
βˆ—
†
[email protected]
[email protected].
CEREMADE, Université Paris Dauphine and CREST-ENSAE,
Dept. of Mathematics, Columbia University, New York,
Research supported by Swiss National Science Foundation Grant PDFM2-120424/1. The
authors thank the two referees for their helpful comments.
1
Generally speaking, it is dicult to prove the dynamic programming
principle when the regularity of the value function is not known a priori, due
to certain measurable selection problems. It was observed in [3] that, in the
unconstrained case, these diculties can be avoided by a weak formulation
of the dynamic programming principle where the value function is replaced
by a test function. This formulation, which is tailored to the derivation of
the PDE in the sense of viscosity solutions, avoids the measurable selection
and uses only a simple covering argument. It turns out that the latter does
not extend directly to the case with state constraints. Essentially, the reason
is that if
𝜈
π‘₯ ∈ π’ͺi.e.,
π‘₯ remains in π’ͺthen 𝜈 may fail
β€²
condition π‘₯ ∈ π’ͺ .
is some admissible control for the initial condition
𝜈
the controlled state process 𝑋𝑑,π‘₯ started at
to have this property for a nearby initial
However, if π’ͺ is open and mild continuity assumptions are satised, then
𝜈
𝑋𝑑,π‘₯
β€² will violate the state constraint with at most small probability when
π‘₯β€² is close to π‘₯. This observation leads us to consider optimization problems
with constraints in probability, and more generally expectation constraints
of the form
𝜈 (𝑇 ))] ≀ π‘š for given π‘š ∈ ℝ.
𝐸[𝑔(𝑋𝑑,π‘₯
We shall see that, following
the idea of [2], such problems are amenable to dynamic programming if the
constraint level
π‘š
is formulated dynamically via an auxiliary family of mar-
tingales. A key insight of the present paper is that relaxing the level
π‘š
by
a small constant allows to prove a weak dynamic programming principle for
general expectation constraint problems (Theorem 2.4), while the PDE can
be derived despite the relaxation.
We shall then obtain the dynamic pro-
gramming principle for the classical state constraint problem (Theorem 3.1)
by passing to a limit
π‘š ↓ 0, with a suitable choice of 𝑔
and certain regularity
assumptions. Of course, expectation constraints are of independent interest
and use; e.g., in Mathematical Finance, where one considers the problem of
𝜈 (𝑇 )] under
𝐸[𝑋0,π‘₯
given π‘š, 𝑝 > 0.
maximizing expected terminal wealth
𝜈 (𝑇 )
𝐸[([𝑋0,π‘₯
βˆ’
π‘₯]βˆ’ )𝑝 ]
≀ π‘š,
for some
the loss constraint
We exemplify the use of these results in the setting of controlled diusions
and show how the PDEs for expectation constraints and state constraints
can be derived (Theorems 4.2 and 4.6). For the latter case, we introduce an
appropriate continuity condition at the boundary, under which we prove a
comparison theorem. While the above concerned an open set
apply directly to the closed domain
that the value function for
π’ͺ
π’ͺ,
π’ͺ and does not
we show via the comparison result
coincides with the one for
π’ͺ,
under certain
conditions.
The remainder of the paper is organized as follows.
In Section 2 we
introduce an abstract setup for dynamic programming under expectation
constraints and prove the corresponding relaxed weak dynamic programming
principle. In Section 3 we deduce the dynamic programming principle for the
state constraint
π’ͺ.
We specialize to the case of controlled diusions in Sec-
tion 4, where we study the Hamilton-Jacobi-Bellman PDEs for expectation
and state constraints. Appendix A provides the comparison theorem.
2
Throughout this paper, (in)equalities between random variables are in
the almost sure sense and relations between processes are up to evanescence,
unless otherwise stated.
2 Dynamic Programming Principle for Expectation
Constraints
𝑇 ∈ (0, ∞) and a probability space (Ξ©, β„±, 𝑃 ) equipped
𝔽 = (ℱ𝑑 )π‘‘βˆˆ[0,𝑇 ] . For each 𝑑 ∈ [0, 𝑇 ], we are given a set 𝒰𝑑
whose elements are seen as controls at time 𝑑. Given a separable metric space
𝑆 , the state space, we denote by S := [0, 𝑇 ] × π‘† the time-augmented state
space. For each (𝑑, π‘₯) ∈ S and 𝜈 ∈ 𝒰𝑑 , we are given a càdlàg adapted process
𝜈 = {𝑋 𝜈 (𝑠), 𝑠 ∈ [𝑑, 𝑇 ]} with values in 𝑆 , the controlled state process.
𝑋𝑑,π‘₯
𝑑,π‘₯
Finally, we are given two measurable functions 𝑓, 𝑔 : 𝑆 β†’ ℝ. We assume
We x a time horizon
with a ltration
that
𝜈
𝐸[βˆ£π‘“ (𝑋𝑑,π‘₯
(𝑇 ))∣] < ∞
so that the
𝜈
𝐸[βˆ£π‘”(𝑋𝑑,π‘₯
(𝑇 ))∣] < ∞
and
for all
𝜈 ∈ 𝒰𝑑 ,
(2.1)
reward and constraint functions
𝜈
𝐹 (𝑑, π‘₯; 𝜈) := 𝐸[𝑓 (𝑋𝑑,π‘₯
(𝑇 ))],
𝜈
𝐺(𝑑, π‘₯; 𝜈) := 𝐸[𝑔(𝑋𝑑,π‘₯
(𝑇 ))]
are well dened. We also introduce the
𝑉 (𝑑, π‘₯, π‘š) :=
sup
value function
𝐹 (𝑑, π‘₯; 𝜈),
(𝑑, π‘₯, π‘š) ∈ SΜ‚,
(2.2)
πœˆβˆˆπ’° (𝑑,π‘₯,π‘š)
where
SΜ‚ := S × β„ ≑ [0, 𝑇 ] × π‘† × β„ and
{
}
𝒰(𝑑, π‘₯, π‘š) := 𝜈 ∈ 𝒰𝑑 : 𝐺(𝑑, π‘₯; 𝜈) ≀ π‘š
is the set of controls admissible at constraint level
π‘š.
Here
(2.3)
sup βˆ… := βˆ’βˆž.
The following observation is the heart of our approach: Given a control
𝜈
admissible at level
π‘š
at the point
π‘₯, relaxing the level π‘š will make 𝜈 adπ‘₯. This will be crucial for the covering
missible in an entire neighborhood of
arguments used below. We use the acronym u.s.c. (l.s.c.) to indicate upper
(lower) semicontinuity.
Lemma 2.1.
Let (𝑑, π‘₯, π‘š) ∈ SΜ‚, 𝜈 ∈ 𝒰(𝑑, π‘₯, π‘š) and assume that the function
π‘₯β€² β†’
7 𝐺(𝑑, π‘₯β€² ; 𝜈) is u.s.c. at π‘₯. For each 𝛿 > 0 there exists a neighborhood 𝐡
of π‘₯ ∈ 𝑆 such that 𝜈 ∈ 𝒰(𝑑, π‘₯β€² , π‘šβ€² + 𝛿) for all π‘₯β€² ∈ 𝐡 and all π‘šβ€² β‰₯ π‘š.
Proof.
𝒰(𝑑, π‘₯, π‘š). In view
of the upper semicontinuity, there exists a neighborhood 𝐡 of π‘₯ such that
𝐺(𝑑, π‘₯β€² ; 𝜈) ≀ π‘š + 𝛿 ≀ π‘šβ€² + 𝛿 for π‘₯β€² ∈ 𝐡 ; that is, 𝜈 ∈ 𝒰(𝑑, π‘₯β€² , π‘šβ€² + 𝛿).
We have
𝐺(𝑑, π‘₯; 𝜈) ≀ π‘š
by the denition (2.3) of
3
A control problem with an expectation constraint of the form (2.3) is
not amenable to dynamic programming if we just consider a xed level
π‘š.
Extending an idea from [2, 1], we shall see that this changes if the constraint is formulated dynamically by using auxiliary martingales.
end, we consider for each
𝑀 = {𝑀 (𝑠), 𝑠 ∈ [𝑑, 𝑇 ]}
𝑑 ∈ [0, 𝑇 ]
ℳ𝑑,0 of
𝑀 (𝑑) = 0. We
a family
with initial value
ℳ𝑑,π‘š := {π‘š + 𝑀 : 𝑀 ∈ ℳ𝑑,0 },
We assume that, for all
there exists
(𝑑, π‘₯) ∈ S
π‘€π‘‘πœˆ [π‘₯] ∈ ℳ𝑑,π‘š
where, necessarily,
and
also introduce
π‘š ∈ ℝ.
𝜈 ∈ 𝒰𝑑 ,
such that
𝜈 (𝑇 ))].
π‘š = 𝐸[𝑔(𝑋𝑑,π‘₯
To this
càdlàg martingales
𝜈 (𝑇 )),
π‘€π‘‘πœˆ [π‘₯](𝑇 ) = 𝑔(𝑋𝑑,π‘₯
In particular, given
𝜈 ∈ 𝒰𝑑 ,
(2.4)
the set
𝜈
β„³+
𝑑,π‘š,π‘₯ (𝜈) := {𝑀 ∈ ℳ𝑑,π‘š : 𝑀 (𝑇 ) β‰₯ 𝑔(𝑋𝑑,π‘₯ (𝑇 ))}
is nonempty for any
𝜈 (𝑇 ))].
π‘š β‰₯ 𝐸[𝑔(𝑋𝑑,π‘₯
More precisely, we have the follow-
ing characterization.
Lemma 2.2.
Proof.
Let (𝑑, π‘₯) ∈ S and π‘š ∈ ℝ. Then
{
}
𝒰(𝑑, π‘₯, π‘š) = 𝜈 ∈ 𝒰𝑑 : β„³+
𝑑,π‘š,π‘₯ (𝜈) βˆ•= βˆ… .
𝜈 ∈ 𝒰𝑑 . If there exists some 𝑀 ∈ β„³+
𝑑,π‘š,π‘₯ (𝜈), then taking expec𝜈
tations yields 𝐸[𝑔(𝑋𝑑,π‘₯ (𝑇 ))] ≀ 𝐸[𝑀 (𝑇 )] = π‘š and hence 𝜈 ∈ 𝒰(𝑑, π‘₯, π‘š).
β€²
𝜈
Conversely, let 𝜈 ∈ 𝒰(𝑑, π‘₯, π‘š); i.e., we have π‘š := 𝐸[𝑔(𝑋𝑑,π‘₯ (𝑇 ))] ≀ π‘š.
+
𝜈
𝜈
β€²
With 𝑀𝑑 [π‘₯] as in (2.4), 𝑀 := 𝑀𝑑 [π‘₯]+π‘šβˆ’π‘š is an element of ℳ𝑑,π‘š,π‘₯ (𝜈).
Let
𝑑 ∈ [0, 𝑇 ] an auxiliary
𝜈 and 𝑀 are 𝔽𝑑 -adapted for
𝔽𝑑 = (ℱ𝑠𝑑 )π‘ βˆˆ[0,𝑇 ] of 𝔽 such that 𝑋𝑑,π‘₯
𝑑 the set of
all π‘₯ ∈ 𝑆 , 𝜈 ∈ 𝒰𝑑 and 𝑀 ∈ ℳ𝑑,0 . Moreover, we denote by 𝒯
𝑑
𝔽 -stopping times with values in [𝑑, 𝑇 ].
It will be useful in the following to consider for each
subltration
The following assumption corresponds to one direction of the dynamic
programming principle; cf. Theorem 2.4(i) below.
Assumption A.
𝜏 ∈ 𝒯𝑑
and
𝑃 -a.e.
(𝑑, π‘₯, π‘š) ∈ SΜ‚, 𝜈 ∈ 𝒰(𝑑, π‘₯, π‘š), 𝑀 ∈ β„³+
𝑑,π‘š,π‘₯ (𝜈),
𝜈
πœ” ∈ Ξ©, there exists πœˆπœ” ∈ 𝒰(𝜏 (πœ”), 𝑋𝑑,π‘₯ (𝜏 )(πœ”), 𝑀 (𝜏 )(πœ”))
For all
such that
]
(
)
[
𝜈
𝜈
(𝜏 )(πœ”); πœˆπœ” .
𝐸 𝑓 (𝑋𝑑,π‘₯
(𝑇 ))β„±πœ (πœ”) ≀ 𝐹 𝜏 (πœ”), 𝑋𝑑,π‘₯
(2.5)
Next, we state two variants of the assumptions for the converse direction
of the dynamic programming principle; we shall comment on the dierences
in Remark 2.5 below.
In the rst variant, the intermediate time is deter-
ministic; this will be enough to cover stopping times with countably many
values in Theorem 2.4(ii) below. We recall the notation
4
π‘€π‘‘πœˆ [π‘₯]
from (2.4).
Assumption B. Let (𝑑, π‘₯) ∈ S, 𝜈 ∈ 𝒰𝑑 , 𝑠 ∈ [𝑑, 𝑇 ], 𝜈¯ ∈ 𝒰𝑠 and Ξ“ ∈ ℱ𝑠𝑑 .
(B1) There exists a control 𝜈
˜ ∈ 𝒰𝑑 such that
𝜈˜
𝜈
𝑋𝑑,π‘₯
(β‹…) = 𝑋𝑑,π‘₯
(β‹…)
on
[𝑑, 𝑇 ] × (Ξ© βˆ– Ξ“);
(2.6)
𝜈˜
𝑋𝑑,π‘₯
(β‹…)
on
[𝑠, 𝑇 ] × Ξ“;
(2.7)
on
Ξ“.
(2.8)
𝜈¯
(β‹…)
𝑋𝑠,𝑋
𝜈
𝑑,π‘₯ (𝑠)
=
[
]
𝜈˜
𝜈
𝐸 𝑓 (𝑋𝑑,π‘₯
(𝑇 ))ℱ𝑠 β‰₯ 𝐹 (𝑠, 𝑋𝑑,π‘₯
(𝑠); 𝜈¯)
𝜈˜ is
(𝑠, Ξ“).
The control
B2)
(
and
𝜈¯
Let
𝑀 ∈ ℳ𝑑,0 .
on
denoted by
𝜈 βŠ—(𝑠,Ξ“) 𝜈¯
and called a
There exists a process
concatenation
¯ = {𝑀
¯ (π‘Ÿ), π‘Ÿ ∈ [𝑠, 𝑇 ]}
𝑀
of
𝜈
such
that
(
)
¯ (β‹…)(πœ”) = 𝑀 𝜈¯ [𝑋 𝜈 (𝑠)(πœ”)](β‹…) (πœ”)
𝑀
𝑠
𝑑,π‘₯
on
[𝑠, 𝑇 ] 𝑃 -a.s.
and
(
[
] )
¯ βˆ’π‘€
¯ (𝑠) + 𝑀 (𝑠) 1Ξ“ ∈ ℳ𝑑,0 .
𝑀 1[𝑑,𝑠) + 1[𝑠,𝑇 ] 𝑀 1Ξ©βˆ–Ξ“ + 𝑀
B3)
(
π‘š ∈ ℝ and 𝑀 ∈ β„³+
𝑑,π‘š,π‘₯ (𝜈).
𝜈 (𝑠)(πœ”), 𝑀 (𝑠)(πœ”)).
πœˆπœ” ∈ 𝒰(𝑠, 𝑋𝑑,π‘₯
Let
For
𝑃 -a.e. πœ” ∈ Ξ©, there exist a control
In the second variant, the intermediate time is a stopping time and we
have an additional assumption about the structure of the sets
𝒰𝑠 .
This
variant corresponds to Theorem 2.4(ii') below.
Assumption B'. Let (𝑑, π‘₯) ∈ S, 𝜈 ∈ 𝒰𝑑 , 𝜏 ∈ 𝒯 𝑑 , Ξ“ ∈ β„±πœπ‘‘ and 𝜈¯ ∈ 𝒰βˆ₯𝜏 βˆ₯𝐿∞ .
β€²
(B0') 𝒰𝑠 βŠ‡ 𝒰𝑠′ for all 0 ≀ 𝑠 ≀ 𝑠 ≀ 𝑇 .
(B1') There exists a control 𝜈
˜ ∈ 𝒰𝑑 , denoted by 𝜈 βŠ—(𝜏,Ξ“) 𝜈¯, such that
𝜈˜
𝜈
𝑋𝑑,π‘₯
(β‹…) = 𝑋𝑑,π‘₯
(β‹…)
on
[𝑑, 𝑇 ] × (Ξ© βˆ– Ξ“);
𝜈˜
𝑋𝑑,π‘₯
(β‹…)
on
[𝜏, 𝑇 ] × Ξ“;
on
Ξ“.
𝜈¯
π‘‹πœ,𝑋
(β‹…)
𝜈
𝑑,π‘₯ (𝜏 )
=
[
]
𝜈˜
𝜈
𝐸 𝑓 (𝑋𝑑,π‘₯
(𝑇 ))β„±πœ β‰₯ 𝐹 (𝜏, 𝑋𝑑,π‘₯
(𝜏 ); 𝜈¯)
B2')
(
Let
𝑀 ∈ ℳ𝑑,0 .
There exists a process
(2.9)
(2.10)
¯ = {𝑀
¯ (π‘Ÿ), π‘Ÿ ∈ [𝜏, 𝑇 ]}
𝑀
such
that
(
)
¯ (β‹…)(πœ”) = 𝑀 𝜈¯ [𝑋 𝜈 (𝜏 )(πœ”)](β‹…) (πœ”)
𝑀
𝑑,π‘₯
𝜏 (πœ”)
on
[𝜏, 𝑇 ] 𝑃 -a.s.
and
(
[
] )
¯ βˆ’π‘€
¯ (𝜏 ) + 𝑀 (𝜏 ) 1Ξ“ ∈ ℳ𝑑,0 .
𝑀 1[𝑑,𝜏 ) + 1[𝜏,𝑇 ] 𝑀 1Ξ©βˆ–Ξ“ + 𝑀
B3')
(
π‘š ∈ ℝ and 𝑀 ∈ β„³+
𝑑,π‘š,π‘₯ (𝜈). For 𝑃 -a.e. πœ” ∈ Ξ©, there exist a control
𝜈
πœˆπœ” ∈ 𝒰(𝜏 (πœ”), 𝑋𝑑,π‘₯ (𝜏 )(πœ”), 𝑀 (𝜏 )(πœ”)).
Let
5
Remark 2.3.
(i) Assumption B' implies Assumption B. Moreover, As-
sumption A implies (B3').
D := {(𝑑, π‘₯, π‘š) ∈ SΜ‚ : 𝒰(𝑑, π‘₯, π‘š) βˆ•= βˆ…}
(ii) Let
denote the natural domain
of our optimization problem. Then (B3') can be stated as follows: for
𝜈 ∈ 𝒰(𝑑, π‘₯, π‘š) and 𝑀 ∈ β„³+
𝑑,π‘š,π‘₯ (𝜈),
(
)
𝜈
𝜏, 𝑋𝑑,π‘₯
(𝜏 ), 𝑀 (𝜏 ) ∈ D 𝑃 -a.s.
any
𝜏
Under (B3), the same holds if
for all
𝜏 ∈ 𝒯 𝑑.
(2.11)
takes countably many values.
We
remark that the invariance property (2.11) corresponds to one direction
in the geometric dynamic programming of [15].
(iii) We note that (2.9) (and similarly (2.7)) states in particular that
( 𝜈¯
(π‘Ÿ, πœ”) 7β†’ π‘‹πœ,𝑋
𝜈
)
(π‘Ÿ)
𝑑,π‘₯ (𝜏 )
(
(πœ”) := π‘‹πœπœˆ¯(πœ”),𝑋 𝜈
)
(π‘Ÿ)
𝑑,π‘₯ (𝜏 )(πœ”)
(πœ”)
[𝑑, 𝑇 ] × Ξ“. Of
𝜈¯ .
(𝑑, π‘₯) 7β†’ 𝑋𝑑,π‘₯
is a well-dened adapted process (up to evanescence) on
course, this is an implicit measurability condition on
(iv) For an illustration of (B1'), let us assume that the controls are predictable processes. In this case, one can often take
(
)
𝜈 βŠ—(𝜏,Ξ“) 𝜈¯ := 𝜈1[0,𝜏 ] + 1(𝜏,𝑇 ] 𝜈¯1Ξ“ + 𝜈1Ξ©βˆ–Ξ“
(2.12)
𝜈 βŠ—(𝜏,Ξ“) 𝜈¯ ∈ 𝒰𝑑 is called stability under concatenation. The idea is that we use the control 𝜈 up to time 𝜏 ; after time
𝜏 , we continue using 𝜈 in the event Ξ© βˆ– Ξ“ while we switch to the control
𝜈¯ in the event Ξ“. Of course, one can omit the set Ξ“ ∈ β„±πœπ‘‘ by observing
and the condition that
that
𝜈 βŠ—(𝜏,Ξ“) 𝜈¯ = 𝜈1[0,𝜏 β€² ] + 𝜈¯1(𝜏 β€² ,𝑇 ]
for the
𝔽𝑑 -stopping
time
𝜏 β€² := 𝜏 1Ξ“ + 𝑇 1Ξ©βˆ–Ξ“ .
We can now state our weak dynamic programming principle for the
stochastic control problem (2.2) with expectation constraint.
The formu-
lation is weak in two ways; namely, the value function is replaced by a
test function and the constraint level
constant
𝛿>0
π‘š
is relaxed by an arbitrarily small
(cf. (2.15) below). The exibility of choosing the set
pearing below, will be used in Section 3. We recall the set
D
π’Ÿ
ap-
introduced in
Remark 2.3(ii).
Theorem 2.4.
Let (𝑑, π‘₯, π‘š) ∈ SΜ‚, 𝜈 ∈ 𝒰(𝑑, π‘₯, π‘š) and 𝑀 ∈ β„³+
𝑑,π‘š,π‘₯ (𝜈). Let
𝑑
𝜈
𝜏 ∈ 𝒯 and let π’Ÿ βŠ† SΜ‚ be a set such that (𝜏, 𝑋𝑑,π‘₯ (𝜏 ), 𝑀 (𝜏 )) ∈ π’Ÿ holds 𝑃 -a.s.
(i) Let Assumption A hold true and let πœ‘ : SΜ‚ [β†’ [βˆ’βˆž, ∞] be a measurable
]
𝜈 (𝜏 ), 𝑀 (𝜏 ))βˆ’ < ∞
function such that 𝑉 ≀ πœ‘ on π’Ÿ. Then 𝐸 πœ‘(𝜏, 𝑋𝑑,π‘₯
and
]
[
𝜈
(𝜏 ), 𝑀 (𝜏 )) .
(2.13)
𝐹 (𝑑, π‘₯; 𝜈) ≀ 𝐸 πœ‘(𝜏, 𝑋𝑑,π‘₯
6
(ii) Let 𝛿 > 0, let Assumption B hold true and assume that 𝜏 takes countably many values (𝑑𝑖 )𝑖β‰₯1 . Let πœ‘ : SΜ‚ β†’ [βˆ’βˆž, ∞) be a measurable function such that 𝑉 β‰₯ πœ‘ on π’Ÿ. Assume that for any xed 𝜈¯ ∈ 𝒰𝑑𝑖 ,
⎫
(π‘₯β€² , π‘šβ€² ) 7β†’ πœ‘(𝑑𝑖 , π‘₯β€² , π‘šβ€² ) is u.s.c. 
⎬
π‘₯β€² 7β†’ 𝐹 (𝑑𝑖 , π‘₯β€² ; 𝜈¯) is l.s.c.
on π’Ÿπ‘–
(2.14)

⎭
β€²
β€²
π‘₯ 7β†’ 𝐺(𝑑𝑖 , π‘₯ ; 𝜈¯) is u.s.c.
for all 𝑖 β‰₯ 1, where π’Ÿπ‘– := {(π‘₯β€² , π‘šβ€² ) : (𝑑𝑖 , π‘₯β€² , π‘šβ€² ) ∈ π’Ÿ} βŠ† 𝑆 × β„. Then
[
]
𝜈
𝑉 (𝑑, π‘₯, π‘š + 𝛿) β‰₯ 𝐸 πœ‘(𝜏, 𝑋𝑑,π‘₯
(𝜏 ), 𝑀 (𝜏 )) .
(2.15)
(ii') Let 𝛿 > 0 and let Assumption B' hold true. Let πœ‘ : SΜ‚ β†’ [βˆ’βˆž, ∞) be a
measurable function such that 𝑉 β‰₯ πœ‘ on π’Ÿ. Assume that π’Ÿ ∩ D βŠ† SΜ‚
is 𝜎 -compact and that for any xed 𝜈¯ ∈ 𝒰𝑑0 , 𝑑0 ∈ [𝑑, 𝑇 ],
⎫
(𝑑′ , π‘₯β€² , π‘šβ€² ) 7β†’ πœ‘(𝑑′ , π‘₯β€² , π‘šβ€² ) is u.s.c. 
⎬
(𝑑′ , π‘₯β€² ) 7β†’ 𝐹 (𝑑′ , π‘₯β€² ; 𝜈¯) is l.s.c.
on π’Ÿ ∩ {𝑑′ ≀ 𝑑0 }. (2.16)

(𝑑′ , π‘₯β€² ) 7β†’ 𝐺(𝑑′ , π‘₯β€² ; 𝜈¯) is u.s.c. ⎭
Then
(2.15)
holds true.
π‘Œ
The following convention is used on the right hand side of (2.15): if
is
any random variable,
we set
𝐸[π‘Œ ] := βˆ’βˆž
𝐸[π‘Œ + ] = 𝐸[π‘Œ βˆ’ ] = ∞.
if
We note that Assumption (B0') ensures that the expressions
𝐺(𝑑′ , π‘₯β€² ; 𝜈¯) in (2.16) are well dened for 𝑑′
Remark 2.5.
(2.17)
𝐹 (𝑑′ , π‘₯β€² ; 𝜈¯)
and
≀ 𝑑0 .
The dierence between parts (ii) and (ii') of the theorem
[0, 𝑇 ] × π‘† × β„
𝑆 × β„ and hence no
stems from the fact that in the proof of (ii') we consider
as the state space while for (ii) it suces to consider
assumptions on the time variable are necessary. Regarding applications, (ii)
is obviously the better choice if stopping times with countably many values
(and in particular deterministic times) are sucient.
There is a number of cases where the extension to a general stopping
time
𝜏
can be accomplished a posteriori by approximation, in particular if
one has a priori estimates for the value function so that one can restrict to
test functions with specic growth properties. Assume for illustration that
𝑓
is bounded from above, then so is
functions
πœ‘
hurt to assume that
let
(πœπ‘› )
𝑉
and one will be interested only in test
which are bounded from above; moreover, it will typically not
πœ‘
is u.s.c. (or even continuous) in all variables.
Now
be a sequence of stopping time taking nitely many (e.g., dyadic)
values such that
πœπ‘› ↓ 𝜏 𝑃 -a.s.
Applying (2.15) to each
7
πœπ‘›
and using Fatou's
𝜈
𝑋𝑑,π‘₯
time 𝜏 .
lemma as well as the right-continuity of the paths of
nd that (2.15) also holds for the general stopping
and
𝑀,
we then
On the other hand, it is not always possible to pass to the limit as above
and then (ii') is necessary to treat general stopping times. The compactness
provided that 𝑆 is 𝜎 -compact; e.g., 𝑆 = ℝ𝑑 .
Then, the typical way to apply (ii') is to take (𝑑, π‘₯, π‘š) ∈ int D and let π’Ÿ be
a open or closed neighborhood of (𝑑, π‘₯, π‘š) such that π’Ÿ βŠ† D.
assumption should be reasonable
Proof of Theorem 2.4. (i) Fix (𝑑, π‘₯, π‘š) ∈ SΜ‚. With πœˆπœ” as in (2.5), the denition (2.2) of 𝑉 and 𝑉 ≀ πœ‘ on π’Ÿ imply that
[
]
(
)
𝜈
𝜈
𝐸 𝑓 (𝑋𝑑,π‘₯
(𝑇 ))βˆ£β„±πœ (πœ”) ≀ 𝐹 𝜏 (πœ”), 𝑋𝑑,π‘₯
(𝜏 )(πœ”); πœˆπœ”
(
)
𝜈
≀ 𝑉 𝜏 (πœ”), 𝑋𝑑,π‘₯
(𝜏 )(πœ”), 𝑀 (𝜏 )(πœ”)
(
)
𝜈
≀ πœ‘ 𝜏 (πœ”), 𝑋𝑑,π‘₯
(𝜏 )(πœ”), 𝑀 (𝜏 )(πœ”) 𝑃 -a.s.
After noting that the left hand side is integrable by (2.1), the result follows
by taking expectations.
(𝑑, π‘₯, π‘š) ∈ SΜ‚ and let πœ€ > 0.
1. Countable Cover. Like π’Ÿ, the set π’Ÿ ∩ D has the property
(
)
𝜈
𝜏, 𝑋𝑑,π‘₯
(𝜏 ), 𝑀 (𝜏 ) ∈ π’Ÿ ∩ D 𝑃 -a.s.;
(ii) Fix
that
(2.18)
𝜏 takes countably many values. Replacing π’Ÿ
π’Ÿ ∩ D if necessary, we may therefore assume that π’Ÿ βŠ† D. Using also
that 𝑉 β‰₯ πœ‘ on π’Ÿ , the denition (2.2) of 𝑉 shows that there exists for each
(𝑠, 𝑦, 𝑛) ∈ π’Ÿ some 𝜈 𝑠,𝑦,𝑛 ∈ 𝒰(𝑠, 𝑦, 𝑛) satisfying
this follows from (2.11) since
by
𝐹 (𝑠, 𝑦; 𝜈 𝑠,𝑦,𝑛 ) β‰₯ πœ‘(𝑠, 𝑦, 𝑛) βˆ’ πœ€.
Fix one of the points
𝑑𝑖
(2.19)
(𝑦, 𝑛) ∈ π’Ÿπ‘– ,
in time. For each
the semicontinuity
assumptions (2.14) and Lemma 2.1 imply that there exists a neighborhood
(𝑦, 𝑛) ∈ 𝐡 𝑖 (𝑦, 𝑛) βŠ† 𝑆ˆ := 𝑆 × β„
(of size depending on
⎫

𝜈 𝑑𝑖 ,𝑦,𝑛 ∈ 𝒰(𝑑𝑖 , 𝑦 β€² , 𝑛′ + 𝛿)
⎬
β€²
β€²
πœ‘(𝑑𝑖 , 𝑦 , 𝑛 ) ≀ πœ‘(𝑑𝑖 , 𝑦, 𝑛) + πœ€

𝐹 (𝑑𝑖 , 𝑦 β€² ; 𝜈 𝑑𝑖 ,𝑦,𝑛 ) β‰₯ 𝐹 (𝑑𝑖 , 𝑦; 𝜈 𝑑𝑖 ,𝑦,𝑛 ) βˆ’ πœ€ ⎭
for all
𝑦, 𝑛, 𝑖, πœ€, 𝛿 )
such that
(𝑦 β€² , 𝑛′ ) ∈ 𝐡 𝑖 (𝑦, 𝑛) ∩ π’Ÿπ‘– .
(2.20)
Here the rst inequality may read
βˆ’βˆž ≀
βŠ† 𝑆ˆ is metric
Λ†.
topology on 𝑆
βˆ’βˆž. We note that π’Ÿπ‘–
separable for the subspace topology relative to the product
Λ† forms an open cover
{𝐡 𝑖 (𝑦, 𝑛) ∩ π’Ÿπ‘– : (𝑦, 𝑛) ∈ 𝑆}
𝑖
𝑖
𝑖
Λ†
of π’Ÿ , there exists a sequence (𝑦𝑗 , 𝑛𝑗 )𝑗β‰₯1 in 𝑆 such that {𝐡 (𝑦𝑗 , 𝑛𝑗 ) ∩ π’Ÿ }𝑗β‰₯1
𝑖
𝑖
𝑑
,𝑦
,𝑛
𝑖
𝑖
is a countable subcover of π’Ÿ . We set πœˆπ‘— := 𝜈 𝑖 𝑗 𝑗 and 𝐡𝑗 := 𝐡 (𝑦𝑗 , 𝑛𝑗 ), so
Therefore, since the family
that
π’Ÿπ‘– βŠ†
βˆͺ
𝑗β‰₯1
8
𝐡𝑗𝑖 .
(2.21)
We can now dene, for
𝑖
still being xed, a measurable partition
(𝐴𝑖𝑗 )𝑗β‰₯1
of
βˆͺ𝑗β‰₯1 𝐡𝑗𝑖 by
𝐴𝑖1 := 𝐡1𝑖 ,
Since
𝐴𝑖𝑗 βŠ† 𝐡𝑗𝑖 ,
𝑖
𝐴𝑖𝑗+1 := 𝐡𝑗+1
βˆ– (𝐡1𝑖 βˆͺ β‹… β‹… β‹… βˆͺ 𝐡𝑗𝑖 ),
𝑗 β‰₯ 1.
the inequalities (2.19) and (2.20) yield that
𝐹 (𝑑𝑖 , 𝑦 β€² ; πœˆπ‘—π‘– ) β‰₯ πœ‘(𝑑𝑖 , 𝑦 β€² , 𝑛′ ) βˆ’ 3πœ€
for all
(𝑦 β€² , 𝑛′ ) ∈ 𝐴𝑖𝑗 ∩ π’Ÿπ‘– .
(2.22)
2. Concatenation. Fix an integer π‘˜ β‰₯ 1; we now focus on (𝑑𝑖 )1β‰€π‘–β‰€π‘˜ . We
may assume that 𝑑1 < 𝑑2 < β‹… β‹… β‹… < π‘‘π‘˜ , by eliminating and relabeling some of
𝑑
the 𝑑𝑖 . We dene the β„±πœ -measurable sets
βˆͺ
{
}
𝜈
Γ𝑖𝑗 := 𝜏 = 𝑑𝑖 and (𝑋𝑑,π‘₯
(𝑑𝑖 ), 𝑀 (𝑑𝑖 )) ∈ 𝐴𝑖𝑗 ∈ ℱ𝑑𝑖 and Ξ“(π‘˜) :=
Γ𝑖𝑗 .
1≀𝑖,π‘—β‰€π‘˜
β€²
𝑑𝑖 are distinct and 𝐴𝑖𝑗 ∩ 𝐴𝑖𝑗 β€² = βˆ… for 𝑗 βˆ•= 𝑗 β€² , we have Γ𝑖𝑗 ∩ Γ𝑖𝑗 β€² = βˆ…
(𝑖, 𝑗) βˆ•= (𝑖′ , 𝑗 β€² ). We can then consider the successive concatenation
Since the
for
𝜈(π‘˜) := 𝜈 βŠ—(𝑑1 ,Ξ“11 ) 𝜈11 βŠ—(𝑑1 ,Ξ“12 ) 𝜈21 β‹… β‹… β‹… βŠ—(𝑑1 ,Ξ“1 ) πœˆπ‘˜1 βŠ—(𝑑2 ,Ξ“21 ) 𝜈12 β‹… β‹… β‹… βŠ—(π‘‘π‘˜ ,Ξ“π‘˜ ) πœˆπ‘˜π‘˜ ,
π‘˜
π‘˜
𝜈 π‘Ž βŠ— 𝜈 𝑏 βŠ— 𝜈 𝑐 := (𝜈 π‘Ž βŠ— 𝜈 𝑏 ) βŠ— 𝜈 𝑐 . It
follows from an iterated application of (B1) that 𝜈(π‘˜) is well dened and in
particular 𝜈(π‘˜) ∈ 𝒰𝑑 . (To understand the meaning of 𝜈(π‘˜), it may be useful
which is to be read from the left with
to note that in the example considered in (2.12), we would have
)
(
βˆ‘
𝜈(π‘˜) = 𝜈1[0,𝜏 ] + 1(𝜏,𝑇 ] 𝜈1Ξ©βˆ–Ξ“(π‘˜) +
πœˆπ‘—π‘– 1Γ𝑖 ;
𝑗
1≀𝑖,π‘—β‰€π‘˜
πœˆπ‘—π‘–
i.e., at time
𝜏,
we switch to
𝜈(π‘˜)
𝑋𝑑,π‘₯
on
Ξ© βˆ– Ξ“(π‘˜) ∈ β„±πœ
𝜈
= 𝑋𝑑,π‘₯
on
Γ𝑖𝑗 .)
We note that (2.6) implies that
and hence
]
]
[
[
𝜈(π‘˜)
𝜈
𝐸 𝑓 (𝑋𝑑,π‘₯ (𝑇 ))β„±πœ = 𝐸 𝑓 (𝑋𝑑,π‘₯
(𝑇 ))β„±πœ
Moreover, the fact that
𝜏 = 𝑑𝑖
on
on
Ξ© βˆ– Ξ“(π‘˜).
Γ𝑖𝑗 , repeated application of
(2.23)
(2.6), and (2.8)
show that
]
[
𝜈(π‘˜)
𝜈
𝐸 𝑓 (𝑋𝑑,π‘₯ (𝑇 ))β„±πœ β‰₯ 𝐹 (𝜏, 𝑋𝑑,π‘₯
(𝜏 ); πœˆπ‘—π‘– )
on
Γ𝑖𝑗 ,
for
1 ≀ 𝑖, 𝑗 ≀ π‘˜,
and we deduce via (2.22) that
βˆ‘
]
[
𝜈(π‘˜)
𝜈
𝐸 𝑓 (𝑋𝑑,π‘₯ (𝑇 ))β„±πœ 1Ξ“(π‘˜) β‰₯
𝐹 (𝜏, 𝑋𝑑,π‘₯
(𝜏 ); πœˆπ‘—π‘– )1Γ𝑖
𝑗
1≀𝑖,π‘—β‰€π‘˜
β‰₯
βˆ‘ (
)
𝜈
πœ‘(𝜏, 𝑋𝑑,π‘₯
(𝜏 ), 𝑀 (𝜏 )) βˆ’ 3πœ€ 1Γ𝑖
𝑗
1≀𝑖,π‘—β‰€π‘˜
𝜈
(𝜏 ), 𝑀 (𝜏 ))1Ξ“(π‘˜) βˆ’ 3πœ€.
β‰₯ πœ‘(𝜏, 𝑋𝑑,π‘₯
9
(2.24)
3. Admissibility.
Next, we show that
there exists, for each pair
such that
1 ≀ 𝑖, 𝑗 ≀ π‘˜ ,
𝜈(π‘˜) ∈ 𝒰(𝑑, π‘₯, π‘š + 𝛿). By (B2)
𝑀𝑗𝑖 = {𝑀𝑗𝑖 (𝑒), 𝑒 ∈ [𝑑𝑖 , 𝑇 ]}
a process
( πœˆπ‘–
)
𝜈
𝑀𝑗𝑖 (β‹…)(πœ”) = 𝑀𝑑𝑖𝑗 [𝑋𝑑,π‘₯
(𝑑𝑖 )(πœ”)](β‹…) (πœ”)
on
Γ𝑖𝑗
(2.25)
and such that
𝑀 (π‘˜) := (𝑀 + 𝛿)1[𝑑,𝜏 )
)
(
βˆ‘ [
]
𝑖
𝑖
+ 1[𝜏,𝑇 ] (𝑀 + 𝛿)1Ξ©βˆ–Ξ“(π‘˜) +
𝑀𝑗 βˆ’ 𝑀𝑗 (𝑑𝑖 ) + 𝑀 (𝑑𝑖 ) + 𝛿 1Γ𝑖
𝑗
1≀𝑖,π‘—β‰€π‘˜
ℳ𝑑,π‘š+𝛿 . We note that 𝑀 (π‘˜) (𝑇 ) β‰₯ 𝑀𝑗𝑖 (𝑇 ) on Γ𝑖𝑗 since
𝑀𝑗𝑖 (𝑑𝑖 ) ≀ 𝑀 (𝑑𝑖 ) + 𝛿 on Γ𝑖𝑗 as a result of the rst condition in (2.20). Hence,
is an element of
using (2.7) and (2.25), we have
( πœˆπ‘–
𝜈(π‘˜)
𝑔(𝑋𝑑,π‘₯ (𝑇 )) = 𝑔 𝑋𝑑𝑖𝑗,𝑋 𝜈
)
(𝑇 )
𝑑,π‘₯ (𝑑𝑖 )
This holds for all
π‘€βˆˆ
1 ≀ 𝑖, 𝑗 ≀ π‘˜ .
≀ 𝑀𝑗𝑖 (𝑇 ) ≀ 𝑀 (π‘˜) (𝑇 )
Γ𝑖𝑗 .
on
(2.26)
On the other hand, using (2.6) and that
β„³+
𝑑,π‘š,π‘₯ (𝜈), we have that
𝜈(π‘˜)
𝜈
𝑔(𝑋𝑑,π‘₯ (𝑇 )) = 𝑔(𝑋𝑑,π‘₯
(𝑇 )) ≀ 𝑀 (𝑇 ) ≀ 𝑀 (𝑇 ) + 𝛿 = 𝑀 (π‘˜) (𝑇 )
on
Ξ© βˆ– Ξ“(π‘˜).
(2.27)
𝜈(π‘˜)
𝑔(𝑋𝑑,π‘₯ (𝑇 )) ≀ 𝑀 (π‘˜) (𝑇 )
𝜈(π‘˜) ∈ 𝒰(𝑑, π‘₯, π‘š + 𝛿).
Combining (2.26) and (2.27), we have
Lemma 2.2 yields that
on
Ξ©
and so
4. πœ€-Optimality. We may assume that either the positive or the negative
𝜈
part of πœ‘(𝜏, 𝑋𝑑,π‘₯ (𝜏 ), 𝑀 (𝜏 )) is integrable, as otherwise our claim (2.15) is
trivial by (2.17). Using the denition (2.2) of 𝑉 as well as (2.23) and (2.24),
we have that
[
]
𝜈(π‘˜)
𝑉 (𝑑, π‘₯, π‘š + 𝛿) β‰₯ 𝐸 𝑓 (𝑋𝑑,π‘₯ (𝑇 ))
]]
[ [
𝜈(π‘˜)
= 𝐸 𝐸 𝑓 (𝑋𝑑,π‘₯ (𝑇 ))β„±πœ
[
]
[
]
𝜈
𝜈
β‰₯ 𝐸 πœ‘(𝜏, 𝑋𝑑,π‘₯
(𝜏 ), 𝑀 (𝜏 ))1Ξ“(π‘˜) βˆ’ 3πœ€ + 𝐸 𝑓 (𝑋𝑑,π‘₯
(𝑇 ))1Ξ©βˆ–Ξ“(π‘˜)
for every
π‘˜ β‰₯ 1.
Letting
π‘˜ β†’ ∞,
we have
Ξ“(π‘˜) ↑ Ξ© 𝑃 -a.s.
by (2.18)
and (2.21); therefore,
[
]
𝜈
𝐸 𝑓 (𝑋𝑑,π‘₯
(𝑇 ))1Ξ©βˆ–Ξ“(π‘˜) β†’ 0
by dominated convergence and (2.1). Moreover, monotone convergence yields
[
]
[
]
𝜈
𝜈
𝐸 πœ‘(𝜏, 𝑋𝑑,π‘₯
(𝜏 ), 𝑀 (𝜏 ))1Ξ“(π‘˜) β†’ 𝐸 πœ‘(𝜏, 𝑋𝑑,π‘₯
(𝜏 ), 𝑀 (𝜏 )) ;
to see this, consider separately the cases where the positive or the negative
part of
𝜈 (𝜏 ), 𝑀 (𝜏 ))
πœ‘(𝜏, 𝑋𝑑,π‘₯
are integrable. Hence we have shown that
𝜈
𝑉 (𝑑, π‘₯, π‘š + 𝛿) β‰₯ 𝐸[πœ‘(𝜏, 𝑋𝑑,π‘₯
(𝜏 ), 𝑀 (𝜏 ))] βˆ’ 3πœ€.
10
As
πœ€>0
was arbitrary, this completes the proof of (ii).
(𝑑, π‘₯, π‘š) ∈ SΜ‚ and let πœ€ > 0. In contrast to the proof of (ii), we
SΜ‚ rather than 𝑆 × β„. By (2.11), we may again assume
that π’Ÿ ∩ D = π’Ÿ . Since 𝑉 β‰₯ πœ‘ on π’Ÿ , the denition (2.2) of 𝑉 shows that
𝑠,𝑦,𝑛 ∈ 𝒰(𝑠, 𝑦, 𝑛) satisfying
there exists for each (𝑠, 𝑦, 𝑛) ∈ π’Ÿ some 𝜈
(ii') Fix
shall cover a subset of
𝐹 (𝑠, 𝑦; 𝜈 𝑠,𝑦,𝑛 ) β‰₯ πœ‘(𝑠, 𝑦, 𝑛) βˆ’ πœ€.
Given
(𝑠, 𝑦, 𝑛) ∈ π’Ÿ,
(2.28)
the semicontinuity assumptions (2.16) and a variant
of Lemma 2.1 (including the time variable) imply that there exists a set
𝐡(𝑠, 𝑦, 𝑛) βŠ† SΜ‚
of the form
(
)
𝐡(𝑠, 𝑦, 𝑛) = (𝑠 βˆ’ π‘Ÿ, 𝑠] ∩ [0, 𝑇 ] × π΅π‘Ÿ (𝑦, 𝑛),
where
π‘Ÿ>0
and
π΅π‘Ÿ (𝑦, 𝑛)
𝑆 × β„+ ,
is an open ball in
⎫

⎬
𝜈 𝑠,𝑦,𝑛 ∈ 𝒰(𝑠′ , 𝑦 β€² , 𝑛′ + 𝛿)
πœ‘(𝑠′ , 𝑦 β€² , 𝑛′ ) ≀ πœ‘(𝑠, 𝑦, 𝑛) + πœ€

𝐹 (𝑠′ , 𝑦 β€² ; 𝜈 𝑠,𝑦,𝑛 ) β‰₯ 𝐹 (𝑠, 𝑦; 𝜈 𝑠,𝑦,𝑛 ) βˆ’ πœ€ ⎭
such that
(𝑠′ , 𝑦 β€² , 𝑛′ ) ∈ 𝐡(𝑠, 𝑦, 𝑛) ∩ π’Ÿ.
for all
(2.29)
Note that we have exploited Assumption (B0'), which forces us to use the
half-closed interval for the time variable. As a result,
the product topology on
and
[0, 𝑇 ]
[0, 𝑇 ]×𝑆 ×ℝ if 𝑆
and
𝐡(𝑠, 𝑦, 𝑛)
is open for
ℝ are given the usual topology
is given the topology generated by the half-closed intervals (the
upper limit topology). Let us denote the latter topological space by
[0, 𝑇 ]βˆ— .
βˆ—
β€²
Like the Sorgenfrey line, [0, 𝑇 ] is a Lindelöf space. Let π’Ÿ be the canoniβ€²
cal projection of π’Ÿ βŠ† SΜ‚ ≑ [0, 𝑇 ]×𝑆×ℝ to 𝑆×ℝ. Then π’Ÿ is again 𝜎 -compact.
𝜎 -compact space
π’Ÿ
βˆ—
was 𝜎 -compact for the original topology on SΜ‚ and the topology of [0, 𝑇 ] is
βˆ—
β€²
ner than the original one, we still have that π’Ÿ βŠ† [0, 𝑇 ] × π’Ÿ is a countable
union of closed subsets and therefore π’Ÿ is Lindelöf also for the new topology.
By the Lindelöf property, the cover {𝐡(𝑠, 𝑦, 𝑛) ∩ π’Ÿ : (𝑠, 𝑦, 𝑛) ∈ SΜ‚} of π’Ÿ
𝑠 ,𝑦 ,𝑛
admits a countable subcover {𝐡(𝑠𝑗 , 𝑦𝑗 , 𝑛𝑗 ) ∩ π’Ÿ}𝑗β‰₯1 . We set πœˆπ‘— := 𝜈 𝑗 𝑗 𝑗
and 𝐡𝑗 := 𝐡(𝑠𝑗 , 𝑦𝑗 , 𝑛𝑗 ), then π’Ÿ βŠ† βˆͺ𝑗β‰₯1 𝐡𝑗 and
It is easy to see that the product of a Lindelöf space with a
is again Lindelöf; in particular,
𝐴1 := 𝐡1 ,
[0, 𝑇 ]βˆ— × π’Ÿβ€²
is Lindelöf. Moreover, since
𝐴𝑗+1 := 𝐡𝑗+1 βˆ– (𝐡1 βˆͺ β‹… β‹… β‹… βˆͺ 𝐡𝑗 ),
denes a measurable partition of
βˆͺ𝑗β‰₯1 𝐡𝑗 .
Since
𝑗β‰₯1
𝐴𝑗 βŠ† 𝐡 𝑗 ,
the inequali-
ties (2.28) and (2.29) yield that
𝐹 (𝑠′ , 𝑦 β€² ; πœˆπ‘— ) β‰₯ πœ‘(𝑠′ , 𝑦 β€² , 𝑛′ ) βˆ’ 3πœ€
Similarly as above, we rst x
π‘˜ β‰₯ 1,
for all
(𝑠′ , 𝑦 β€² , 𝑛′ ) ∈ 𝐴𝑗 ∩ π’Ÿ.
dene the
{
}
𝜈
Γ𝑗 := (𝜏, 𝑋𝑑,π‘₯
(𝜏 ), 𝑀 (𝜏 )) ∈ 𝐴𝑗
and
β„±πœπ‘‘ -measurable sets
βˆͺ
Ξ“(π‘˜) :=
Γ𝑗
1β‰€π‘—β‰€π‘˜
11
and set
( ((
)
)
)
𝜈(π‘˜) := β‹… β‹… β‹… 𝜈 βŠ—(𝜏,Ξ“1 ) 𝜈1 βŠ—(𝜏,Ξ“2 ) 𝜈2 β‹… β‹… β‹… βŠ—(𝜏,Ξ“π‘˜ ) πœˆπ‘˜ .
]
[
𝜈(π‘˜)
𝜈
𝐸 𝑓 (𝑋𝑑,π‘₯ (𝑇 ))β„±πœ β‰₯ 𝐹 (𝜏, 𝑋𝑑,π‘₯
(𝜏 ); πœˆπ‘— )
we use that
𝜏 ≀ 𝑠𝑗
𝜏˜ := 𝜏 ∧ 𝑠𝑗
satisfying
on
on
Γ𝑗
for
To check that
1 ≀ 𝑗 ≀ π‘˜,
Γ𝑗 , so that we can apply (2.10) with the stopping time
βˆ₯˜
𝜏 βˆ₯𝐿∞ ≀ 𝑠𝑗 and thus πœˆπ‘— ∈ 𝒰𝑠𝑗 βŠ† 𝒰βˆ₯˜𝜏 βˆ₯𝐿∞ ; c.f. (B0').
The rest of the proof is analogous to the above.
Remark 2.6.
𝜎 -compactness in Theorem 2.4(ii') was
used only to ensure the Lindelöf property of π’Ÿβˆ©D for the topology introduced
The assumption on
in the proof. Therefore, any other assumption ensuring this will do as well.
Let us record a slight generalization of Theorem 2.4(ii),(ii') to the case
of controls which are not necessarily admissible at the given point
(𝑑, π‘₯, π‘š).
The intuition for this result is that the dynamic programming principle holds
as before if we use such controls for a suciently short time (as formalized by
condition (2.30) below) and then switch to admissible ones. More precisely,
the proof also exploits the relaxation which is anyway present in (2.15). We
use the notation of Theorem 2.4.
Corollary 2.7. Let the assumptions of Theorem 2.4(ii) hold true except
that 𝜈 ∈ 𝒰𝑑 and 𝑀 ∈ ℳ𝑑,π‘š are not necessarily in 𝒰(𝑑, π‘₯, π‘š) and β„³+
𝑑,π‘š,π‘₯ (𝜈),
respectively. In addition, assume that
(
)
𝜈
𝜏, 𝑋𝑑,π‘₯
(𝜏 ), 𝑀 (𝜏 ) ∈ D 𝑃 -a.s.
(2.30)
Then the conclusion (2.15) of Theorem 2.4(ii) still holds true. Moreover, the
same generalization holds true for Theorem 2.4(ii').
Proof.
Let us inspect the proof of Theorem 2.4(ii).
Using directly (2.30)
rather than appealing to (2.11), the construction of the covering in Step 1
remains unchanged and the same is true for the concatenation in Step 2. In
Step 3, we proceed as above up to and including (2.26). Note that (2.27)
no longer holds as it used the assumption that
𝜈(π‘˜)
view of (2.26) and 𝑋𝑑,π‘₯ (𝑇 )
=
𝜈 (𝑇 ) on
𝑋𝑑,π‘₯
𝑀 ∈ β„³+
𝑑,π‘š,π‘₯ (𝜈).
Ξ© βˆ– Ξ“(π‘˜),
However, in
we still have
[
]
[
]
[
]
𝜈(π‘˜)
𝜈
𝐸 𝑔(𝑋𝑑,π‘₯ (𝑇 )) ≀ 𝐸 𝑀 (π‘˜) (𝑇 )1Ξ“(π‘˜) + 𝐸 𝑔(𝑋𝑑,π‘₯
(𝑇 ))1Ξ©βˆ–Ξ“(π‘˜) .
𝜈 (𝑇 )) is integrable by (2.1) and Ξ“(π‘˜) ↑ Ξ©, the latter expectation
𝑔(𝑋𝑑,π‘₯
is bounded by 𝛿 for large π‘˜ . Moreover, Ξ“(π‘˜) ∈ β„±πœ , the martingale property
(π‘˜) , and the fact that 𝑀 (π‘˜) = 𝑀 + 𝛿 on [0, 𝜏 ] yield that
of 𝑀
[
]
[
]
[
]
𝐸 𝑀 (π‘˜) (𝑇 )1Ξ“(π‘˜) = 𝐸 𝑀 (π‘˜) (𝜏 )1Ξ“(π‘˜) = 𝐸 (𝑀 (𝜏 ) + 𝛿)1Ξ“(π‘˜) .
Since
Since
𝐸[𝑀 (𝜏 )] = π‘š,
the right hand side is dominated by
Together, we conclude that
𝜈(π‘˜)
𝐸[𝑔(𝑋𝑑,π‘₯ (𝑇 ))] ≀ π‘š + 3𝛿;
12
π‘š + 2𝛿
for large
π‘˜.
i.e.,
𝜈(π‘˜) ∈ 𝒰(𝑑, π‘₯, π‘š + 3𝛿)
π‘˜ . Step 4 of the previous proof then
𝜈 (𝑇 )) is integrable by (2.1)), except that
𝑓 (𝑋𝑑,π‘₯
𝜈(π‘˜) now results in
for all large
applies as before (recall that
the changed admissibility of
𝜈
𝑉 (𝑑, π‘₯, π‘š + 3𝛿) β‰₯ 𝐸[πœ‘(𝜏, 𝑋𝑑,π‘₯
(𝜏 ), 𝑀 (𝜏 ))].
However, since
𝛿 > 0 was arbitrary, this is the same as (2.15).
The argument
to extend Theorem 2.4(ii') is analogous.
While we shall see that the relaxation
𝛿>0
in (2.15) is harmless for the
derivation of the Hamilton-Jacobi-Bellman equation, it is nevertheless inter-
𝛿
esting to know when the
in
π‘š.
can be omitted; i.e., when
𝑉
is right continuous
The following sucient condition will be used when we consider state
constraints.
Lemma 2.8.
such that
Let (𝑑, π‘₯, π‘š) ∈ D. For each 𝛿 > 0 there is 𝜈 𝛿 ∈ 𝒰(𝑑, π‘₯, π‘š + 𝛿)
𝐹 (𝑑, π‘₯; 𝜈 𝛿 ) β‰₯ 𝛿 βˆ’1 ∧ 𝑉 (𝑑, π‘₯, π‘š + 𝛿) βˆ’ 𝛿.
Let 𝛿0 > 0 and assume that for all 0 < 𝛿 ≀ 𝛿0 there exists πœˆΛœπ›Ώ ∈ 𝒰(𝑑, π‘₯, π‘š)
such that
{ πœˆπ›Ώ
}
πœˆΛœπ›Ώ
lim 𝑃 𝑋𝑑,π‘₯
(𝑇 ) βˆ•= 𝑋𝑑,π‘₯
(𝑇 ) = 0
𝛿↓0
and such that the set
{[
]+
}
πœˆπ›Ώ
πœˆΛœπ›Ώ
𝑓 (𝑋𝑑,π‘₯
(𝑇 )) βˆ’ 𝑓 (𝑋𝑑,π‘₯
(𝑇 )) : 0 < 𝛿 ≀ 𝛿0 βŠ† 𝐿1 (𝑃 )
(2.31)
is uniformly integrable. Then π‘šβ€² 7β†’ 𝑉 (𝑑, π‘₯, π‘šβ€² ) is right continuous at π‘š.
Proof.
π‘šβ€² 7β†’ 𝒰(𝑑, π‘₯, π‘šβ€² ) is increasing, we have that (𝑑, π‘₯, π‘š) ∈ D
𝛿
implies 𝒰(𝑑, π‘₯, π‘š + 𝛿) βˆ•= βˆ… for all 𝛿 β‰₯ 0. Hence 𝜈 exists; of course, the
βˆ’1 is necessary only if 𝑉 (𝑑, π‘₯, π‘š + 𝛿) = ∞. Moreover, the
truncation at 𝛿
β€²
β€²
monotonicity of π‘š 7β†’ 𝑉 (𝑑, π‘₯, π‘š ) implies that the right limit 𝑉 (𝑑, π‘₯, π‘š+)
exists and that 𝑉 (𝑑, π‘₯, π‘š+) β‰₯ 𝑉 (𝑑, π‘₯, π‘š); it remains to prove the opposite
𝛿
πœˆπ›Ώ
πœˆΛœπ›Ώ
inequality. Let 0 < 𝛿 ≀ 𝛿0 and set 𝐴 := {𝑋𝑑,π‘₯ (𝑇 ) βˆ•= 𝑋𝑑,π‘₯ (𝑇 )}, then
Since
𝑉 (𝑑, π‘₯, π‘š) β‰₯ 𝐹 (𝑑, π‘₯, πœˆΛœπ›Ώ )
[
(
)]
πœˆπ›Ώ
πœˆΛœπ›Ώ
= 𝐹 (𝑑, π‘₯, 𝜈 𝛿 ) βˆ’ 𝐸 1𝐴𝛿 𝑓 (𝑋𝑑,π‘₯
(𝑇 )) βˆ’ 𝑓 (𝑋𝑑,π‘₯
(𝑇 ))
[
(
)+ ]
πœˆπ›Ώ
πœˆΛœπ›Ώ
β‰₯ 𝛿 βˆ’1 ∧ 𝑉 (𝑑, π‘₯, π‘š + 𝛿) βˆ’ 𝛿 βˆ’ 𝐸 1𝐴𝛿 𝑓 (𝑋𝑑,π‘₯
(𝑇 )) βˆ’ 𝑓 (𝑋𝑑,π‘₯
(𝑇 ))
.
Letting
𝛿 ↓ 0,
Remark 2.9.
we deduce by (2.31) that
𝑉 (𝑑, π‘₯, π‘š) β‰₯ 𝑉 (𝑑, π‘₯, π‘š+).
The integrability assumption (2.31) is clearly satised if
𝑓
is
bounded. In the general case, it may be useful to consider the value function
for a truncated function
𝑓
in a rst step.
13
Remark 2.10.
Our results can be generalized to a setting with multiple
𝑁 ∈ β„• and π‘š ∈ ℝ𝑁 , let
{
𝒰(𝑑, π‘₯, π‘š) := 𝜈 ∈ 𝒰𝑑 : 𝐺𝑖 (𝑑, π‘₯; 𝜈) ≀ π‘šπ‘–
constraints. Given
for
}
𝑖 = 1, . . . , 𝑁 ,
𝜈 (𝑇 ))] for some measurable function 𝑔 𝑖 . In
𝐺𝑖 (𝑑, π‘₯; 𝜈) := 𝐸[𝑔 𝑖 (𝑋𝑑,π‘₯
this case, ℳ𝑑,0 is dened as the set of càdlàg 𝑁 -dimensional martingales
𝑀 = {𝑀 (𝑠), 𝑠 ∈ [𝑑, 𝑇 ]} with initial value 𝑀 (𝑑) = 0, and
where
π‘š ∈ ℝ𝑁 .
ℳ𝑑,π‘š := {π‘š + 𝑀 : 𝑀 ∈ ℳ𝑑,0 },
This generalization, which has been considered in [4] within the framework
of stochastic target problems with controlled loss, also allows to impose constraints at nitely many intermediate times
0 ≀ 𝑇1 ≀ 𝑇2 ≀ β‹… β‹… β‹… ≀ 𝑇 .
Indeed,
we can increase the dimension of the state process and add the components
𝜈 (β‹… ∧ 𝑇 ).
𝑋𝑑,π‘₯
𝑗
3 Application to State Constraints
π’ͺ βŠ† 𝑆 := ℝ𝑑
We consider an open set
and study the stochastic control
problem under the constraint that the state process has to stay in
π’ͺ.
Namely,
we consider the value function
𝑉¯ (𝑑, π‘₯) :=
(𝑑, π‘₯) ∈ S,
sup 𝐹 (𝑑, π‘₯; 𝜈),
(3.1)
πœˆβˆˆπ’°¯(𝑑,π‘₯)
where
{
𝜈
¯ π‘₯) := 𝜈 ∈ 𝒰𝑑 : 𝑋𝑑,π‘₯
𝒰(𝑑,
(𝑠) ∈ π’ͺ
for all
}
𝑠 ∈ [𝑑, 𝑇 ], 𝑃 -a.s. .
In the following discussion we assume that, for all
𝜈
𝑋𝑑,π‘₯
(𝑑, π‘₯) 7β†’
(𝑑, π‘₯) ∈ S
and
𝜈 ∈ 𝒰𝑑 ,
has continuous paths;
(3.2)
𝜈 (π‘Ÿ) is continuous in probability, uniformly in
𝑋𝑑,π‘₯
¯ π‘₯) βˆ•= βˆ…
𝒰(𝑑,
for
π‘Ÿ;
(𝑑, π‘₯) ∈ [0, 𝑇 ] × π’ͺ.
(3.3)
(3.4)
(𝑑𝑛 , π‘₯𝑛 ) β†’ (𝑑, π‘₯) implies
( 𝜈
)
𝜈
sup 𝑑 𝑋𝑑𝑛 ,π‘₯𝑛 (π‘Ÿ), 𝑋𝑑,π‘₯
(π‘Ÿ) β†’ 0 in probability,
Explicitly, the condition (3.3) means that
π‘Ÿβˆˆ[0,𝑇 ]
where we set
𝜈 (π‘Ÿ) := π‘₯
𝑋𝑑,π‘₯
for
π‘Ÿ < 𝑑, 𝑑(β‹…, β‹…)
𝜈 ∈ 𝒰𝑑𝑛
and it is implicitly assumed that
denotes the Euclidean metric,
for all
𝑛.
We shall augment
the state process so that the state constraint becomes a special case of an
expectation constraint.
𝑑(π‘₯) :=
inf{𝑑(π‘₯, π‘₯β€² )
:
π‘₯β€²
To this end, we introduce the distance function
∈ 𝑆 βˆ– π’ͺ}
for
π‘₯βˆˆπ‘†
𝜈
𝜈
π‘Œπ‘‘,π‘₯,𝑦
(𝑠) := 𝑦 ∧ inf 𝑑(𝑋𝑑,π‘₯
(π‘Ÿ)),
π‘Ÿβˆˆ[𝑑,𝑠]
14
and the auxiliary process
𝑠 ∈ [𝑑, 𝑇 ],
𝑦 ∈ [0, ∞).
(3.5)
𝜈 (π‘Ÿ)(πœ”), π‘Ÿ ∈ [𝑑, 𝑇 ]} βŠ† 𝑆 is compact; therefore,
{𝑋𝑑,π‘₯
distance to 𝑆 βˆ– π’ͺ whenever it is contained in π’ͺ :
By (3.2), each trajectory
it has strictly positive
𝜈
{𝑋𝑑,π‘₯
(π‘Ÿ)(πœ”), π‘Ÿ ∈ [𝑑, 𝑇 ]} βŠ† π’ͺ
if and only if
𝜈
π‘Œπ‘‘,π‘₯,1
(𝑇 )(πœ”) > 0.
We consider the augmented state process
( 𝜈
)
𝜈
𝜈
¯ 𝑑,π‘₯,𝑦
𝑋
(β‹…) := 𝑋𝑑,π‘₯
(β‹…), π‘Œπ‘‘,π‘₯,𝑦
(β‹…)
on the state space
𝑆 × [0, ∞),
then
𝜈
𝜈
¯ 𝑑,π‘₯,𝑦
𝐸[𝑔(𝑋
(𝑇 ))] = 𝑃 {π‘Œπ‘‘,π‘₯,𝑦
(𝑇 ) ≀ 0}
for (π‘₯, 𝑦) ∈ 𝑆 × [0, ∞). Now the
¯ 𝜈 (𝑇 ))] ≀ 0 and therefore
𝐸[𝑔(𝑋
𝑑,π‘₯,1
¯ π‘₯) = 𝒰(𝑑, π‘₯, 1, 0)
𝒰(𝑑,
of course, the value
by setting
𝑔(π‘₯, 𝑦) := 1(βˆ’βˆž,0] (𝑦)
state constraint may be expressed as
and
𝑉¯ (𝑑, π‘₯) = 𝑉 (𝑑, π‘₯, 1, 0);
1 may be replaced by any number 𝑦 > 0.
Here and in the
sequel, we use the notation from the previous section applied to the controlled
¯
𝑋
𝑆 × [0, ∞); i.e., we tacitly replace the variable π‘₯ by (π‘₯, 𝑦)
𝒰(𝑑, π‘₯, 𝑦, π‘š) of admissible controls and the associated value
𝑉 (𝑑, π‘₯, 𝑦, π‘š).
state process
on
to dene the set
function
One direction of the dynamic programming principle will be a consequence of the following counterpart of Assumption A.
Assumption
there exists
¯ π‘₯), 𝜏 ∈ 𝒯 𝑑 and 𝑃 -a.e. πœ” ∈ Ξ©,
AΜ„. For all (𝑑, π‘₯) ∈ S, 𝜈 ∈ 𝒰(𝑑,
𝜈
¯
πœˆπœ” ∈ 𝒰(𝜏 (πœ”), 𝑋𝑑,π‘₯ (𝜏 )(πœ”)) such that
]
[
(
)
𝜈
𝜈
𝐸 𝑓 (𝑋𝑑,π‘₯
(𝑇 ))β„±πœ (πœ”) ≀ 𝐹 𝜏 (πœ”), 𝑋𝑑,π‘₯
(𝜏 )(πœ”); πœˆπœ” .
The more dicult direction of the dynamic programming principle will
be inferred from Theorem 2.4 under a right-continuity condition; we shall
exemplify in the subsequent section how to verify this condition.
Theorem 3.1.
¯ π‘₯)} βŠ† 𝒯 𝑑 .
Consider (𝑑, π‘₯) ∈ S and a family {𝜏 𝜈 , 𝜈 ∈ 𝒰(𝑑,
(i) Let Assumption 𝐴¯ hold true and let πœ™[: S β†’ [βˆ’βˆž, ∞] be
] a measurable
𝜈 (𝜏 𝜈 ))βˆ’ < ∞ for all
function such that 𝑉¯ ≀ πœ™. Then 𝐸 πœ™(𝜏 𝜈 , 𝑋𝑑,π‘₯
¯ π‘₯) and
𝜈 ∈ 𝒰(𝑑,
[
]
𝜈
𝑉¯ (𝑑, π‘₯) ≀ sup 𝐸 πœ™(𝜏 𝜈 , 𝑋𝑑,π‘₯
(𝜏 𝜈 )) .
πœˆβˆˆπ’°¯(𝑑,π‘₯)
¯ on 𝑆 × [0, ∞) and
(ii) Let Assumption B' hold true for the state process 𝑋
let (3.2)(3.4) hold true. Moreover, assume that
𝑉 (𝑑, π‘₯, 1, 0) = 𝑉 (𝑑, π‘₯, 1, 0+)
15
(3.6)
and that (𝑠′ , π‘₯β€² ) 7β†’ 𝐹 (𝑠′ , π‘₯β€² ; 𝜈) is l.s.c. on [0, 𝑑0 ] × π’ͺ for all 𝑑0 ∈ [𝑑, 𝑇 ]
and 𝜈 ∈ 𝒰𝑑0 . Then
[
]
𝜈
𝑉¯ (𝑑, π‘₯) β‰₯ sup 𝐸 πœ™(𝜏 𝜈 , 𝑋𝑑,π‘₯
(𝜏 𝜈 ))
(3.7)
πœˆβˆˆπ’°¯(𝑑,π‘₯)
for any u.s.c. function πœ™ : S β†’ [βˆ’βˆž, ∞) such that 𝑉¯ β‰₯ πœ™.
Proof.
¯ π‘₯) βˆ•= βˆ… as otherwise 𝑉¯ (𝑑, π‘₯) = βˆ’βˆž. As
𝒰(𝑑,
𝜈
in the proof of (2.13), we obtain that 𝐹 (𝑑, π‘₯; 𝜈) ≀ 𝐸[πœ‘(𝜏, 𝑋𝑑,π‘₯ (𝜏 ))] for all
¯ π‘₯). The claim follows by taking supremum over 𝜈 .
𝜈 ∈ 𝒰(𝑑,
¯ π‘₯) βˆ•= βˆ… as otherwise the right hand
(ii) Again, we may assume that 𝒰(𝑑,
side of (3.7) equals βˆ’βˆž. We set
(i) We may assume that
π’Ÿ := [𝑑, 𝑇 ] × π’ͺ × (0, ∞) × {0}.
¯ π‘₯), then 𝑔(𝑋
¯ 𝜈 (𝑇 )) = 0 and hence the constant martingale
𝜈 ∈ 𝒰(𝑑,
𝑑,π‘₯,1
𝜈
𝑀 := 0 is contained in β„³+
𝑑,0,(π‘₯,1) (𝜈). Moreover, π‘Œπ‘‘,π‘₯,1 (𝑠) > 0 and hence
¯ 𝜈 (𝑠), 𝑀 (𝑠)) ∈ π’Ÿ for all 𝑠 ∈ [𝑑, 𝑇 ], 𝑃 -a.s. Furthermore, if we dene
(𝑠, 𝑋
𝑑,π‘₯,1
{
πœ™(𝑑′ , π‘₯β€² ), (𝑑′ , π‘₯β€² , 𝑦 β€² , π‘šβ€² ) ∈ π’Ÿ
πœ‘(𝑑′ , π‘₯β€² , 𝑦 β€² , π‘šβ€² ) :=
βˆ’βˆž,
otherwise,
If
then
πœ‘ is u.s.c. on π’Ÿ.
To see that the third semicontinuity condition in (2.16)
is also satised, note that for any
and
(𝑑′ , π‘₯β€² , 𝑦 β€² ), (𝑑′′ , π‘₯β€²β€² , 𝑦 β€²β€² ) ∈ [0, 𝑑0 ] × π‘† × [0, ∞)
𝜈 ∈ 𝒰𝑑0 ,
βˆ£π‘Œπ‘‘πœˆβ€² ,π‘₯β€² ,𝑦′ (𝑇 )βˆ’π‘Œπ‘‘πœˆβ€²β€² ,π‘₯β€²β€² ,𝑦′′ (𝑇 )∣
≀ βˆ£π‘¦ β€² βˆ’ 𝑦 β€²β€² ∣ + inf 𝑑(π‘‹π‘‘πœˆβ€² ,π‘₯β€² (π‘Ÿ)) βˆ’ inf 𝑑(π‘‹π‘‘πœˆβ€²β€² ,π‘₯β€²β€² (π‘Ÿ))
π‘Ÿβˆˆ[𝑑′ ,𝑇 ]
π‘Ÿβˆˆ[𝑑′′ ,𝑇 ]
)
(
≀ βˆ£π‘¦ β€² βˆ’ 𝑦 β€²β€² ∣ + sup 𝑑 π‘‹π‘‘πœˆβ€² ,π‘₯β€² (π‘Ÿ), π‘‹π‘‘πœˆβ€²β€² ,π‘₯β€²β€² (π‘Ÿ) .
π‘Ÿβˆˆ[0,𝑇 ]
Hence (3.3) implies that
As
(βˆ’βˆž, 0] βŠ‚ ℝ
(𝑑′ , π‘₯β€² , 𝑦 β€² ) 7β†’ π‘Œπ‘‘πœˆβ€² ,π‘₯β€² ,𝑦′ (𝑇 ) is continuous in probability.
is closed, we conclude by the Portmanteau theorem that
{
}
(𝑑′ , π‘₯β€² , 𝑦 β€² ) 7β†’ 𝑃 π‘Œπ‘‘πœˆβ€² ,π‘₯β€² ,𝑦′ (𝑇 ) ∈ (∞, 0] ≑ 𝐺(𝑑′ , π‘₯β€² , 𝑦 β€² ; 𝜈)
as required. Since any open subset of a Euclidean space is
since (3.4) implies
𝑀 =0
π’Ÿ ∩ D = π’Ÿ,
= 𝑉 (𝑑, π‘₯, 1, 0+)
[ (
)]
𝜈
𝜈
β‰₯ 𝐸 πœ‘ 𝜏 𝜈 , 𝑋𝑑,π‘₯
(𝜏 𝜈 ), π‘Œπ‘‘,π‘₯,1
(𝜏 𝜈 ), 0
𝜈
= 𝐸[πœ™(𝜏 𝜈 , 𝑋𝑑,π‘₯
(𝜏 𝜈 ))].
¯ π‘₯)
𝜈 ∈ 𝒰(𝑑,
𝜎 -compact
and
we can use (3.6) and Theorem 2.4(ii') with
to obtain that
𝑉¯ (𝑑, π‘₯) = 𝑉 (𝑑, π‘₯, 1, 0)
As
is u.s.c.
was arbitrary, the result follows.
16
Remark 3.2.
Similarly as in Theorem 2.4(ii), there is also a version of The-
orem 3.1(ii) for stopping times taking countably many values. In this case,
Assumption B replaces Assumption B', all conditions on the time variable
are superuous, and one can consider a general separable metric space
𝑆.
4 Application to Controlled Diusions
In this section, we show how the weak formulation of the dynamic programming principle applies in the context of controlled Brownian stochastic differential equations and how it allows to derive the Hamilton-Jacobi-Bellman
PDEs for the value functions associated to optimal control problems with
expectation or state constraints. As the main purpose of this section is to illustrate the use of Theorems 2.4 and 3.1, we shall choose a fairly simple setup
allowing to explain the main points without too many distractions. Given
the generality of those theorems, extensions such as singular control, mixed
control/stopping problems, etc. do not present any particular diculty.
4.1 Setup for Controlled Diusions
From now on, we take
𝑆 = ℝ𝑑
Ξ© = 𝐢([0, 𝑇 ]; ℝ𝑑 )
measure on Ξ©, and π‘Š
and let
be the space
𝑃 the Wiener
the canonical
π‘Šπ‘‘ (πœ”) = πœ”π‘‘ . Let 𝔽 = (ℱ𝑑 )π‘‘βˆˆ[0,𝑇 ] be the augmentation of the ltration
generated by π‘Š ; without loss of generality, β„± = ℱ𝑇 . For 𝑑 ∈ [0, 𝑇 ], the
𝑑
𝑑
auxiliary ltration 𝔽 = (ℱ𝑠 )π‘ βˆˆ[0,𝑇 ] is chosen to be the augmentation of
𝜎(π‘Šπ‘Ÿ βˆ’ π‘Šπ‘‘ , 𝑑 ≀ π‘Ÿ ≀ 𝑠); in particular, 𝔽𝑑 is independent of ℱ𝑑 .
𝑑
Consider a closed set π‘ˆ βŠ† ℝ and let 𝒰 be the set of all π‘ˆ -valued preβˆ«π‘‡
2
dictable processes 𝜈 satisfying 𝐸[
0 βˆ£πœˆπ‘‘ ∣ 𝑑𝑑] < ∞. Then we set
{
}
𝒰𝑑 = 𝜈 ∈ 𝒰 : 𝜈 is 𝔽𝑑 -predictable .
of continuous paths,
process
This choice will be convenient to verify Assumption B'. We remark that the
𝔽𝑑 -predictable controls entails no loss of generality, in the sense
alternative choice 𝒰𝑑 = 𝒰 would result in the same value function.
restriction to
that the
Indeed, this follows from a well known randomization argument (see, e.g.,
[3, Remark 5.2]).
Let
𝕄𝑑
denote the set of
𝑑×𝑑
matrices. Given two Lipschitz continuous
functions
πœ‡ : ℝ𝑑 × π‘ˆ β†’ ℝ𝑑 ,
and
𝜎 : ℝ𝑑 × π‘ˆ β†’ 𝕄𝑑
𝜈 (β‹…) as the unique strong solution
(𝑑, π‘₯, 𝜈) ∈ [0, 𝑇 ] × β„π‘‘ × π’° , we dene 𝑋𝑑,π‘₯
of the stochastic dierential equation (SDE)
∫
𝑋(𝑠) = π‘₯ +
𝑠
𝑠
∫
πœ‡(𝑋(π‘Ÿ), πœˆπ‘Ÿ ) π‘‘π‘Ÿ +
𝑑
𝜎(𝑋(π‘Ÿ), πœˆπ‘Ÿ ) π‘‘π‘Šπ‘Ÿ ,
𝑑
17
𝑑 ≀ 𝑠 ≀ 𝑇,
(4.1)
where we set
𝜈 ∈ 𝒰,
𝜈 (π‘Ÿ) = π‘₯
𝑋𝑑,π‘₯
for
π‘Ÿ ≀ 𝑑.
As
𝜈 (𝑇 )
𝑋𝑑,π‘₯
is square integrable for any
(2.1) is satised whenever
𝑓
and
𝑔
have quadratic growth,
(4.2)
which we assume from now on. In addition, we also impose that
{
𝑓
𝑔
𝑓 βˆ’ has subquadratic growth,
+ has subquadratic growth,
and 𝑔
is l.s.c. and
is u.s.c.
(4.3)
β„Ž : ℝ𝑑 β†’ ℝ is said to have subquadratic growth if β„Ž(π‘₯)/∣π‘₯∣2 β†’ 0 as
∣π‘₯∣ β†’ ∞. This will be used to obtain the semicontinuity properties (2.16).
Furthermore, we take ℳ𝑑,0 to be the family of all càdlàg martingales
𝑑
which start at 0 and are adapted to 𝔽 . By the independence of the increments of π‘Š , we see that 𝑀 ∈ ℳ𝑑,0 is then also a martingale in the ltration
𝜈 (𝑇 ) ∈ 𝐿2 (β„± 𝑑 , 𝑃 )
𝔽𝑑 and that π‘€π‘Ÿ = 0 for π‘Ÿ ≀ 𝑑. For 𝜈 ∈ 𝒰𝑑 , we have 𝑋𝑑,π‘₯
𝑇
𝜈
𝜈
𝑑
and hence (2.4) is satised with 𝑀𝑑 [π‘₯](β‹…) = 𝐸[𝑋𝑑,π‘₯ (𝑇 )βˆ£β„±β‹… ]. It will be useful
to express the martingales as stochastic integrals. Let π’œπ‘‘ denote the set of
βˆ«π‘‡
ℝ𝑑 -valued 𝔽𝑑 -predictable processes 𝛼 such that 0 βˆ£π›Όπ‘‘ ∣2 𝑑𝑑 < ∞ 𝑃 -a.s. and
such that
∫ β‹…
𝛼
𝑀𝑑,0 (β‹…) :=
π›Όπ‘ βŠ€ π‘‘π‘Šπ‘ 
where
𝑑
⊀
is a martingale ( denotes transposition). Then the Brownian representation
theorem yields that
{ 𝛼
}
ℳ𝑑,0 = 𝑀𝑑,0
: 𝛼 ∈ π’œπ‘‘ .
In the following, we also use the notation
𝛼 := π‘š + 𝑀 𝛼 .
𝑀𝑑,π‘š
𝑑,0
Lemma 4.1.
In the above setup, Assumptions A, 𝐴¯, B, B' are satised and
𝐹 and 𝐺 satisfy (2.16).
Proof.
Assumption (B0') is immediate from the denition of
𝒰𝑑 .
We dene
the concatenation of controls by (2.12).
The validity of Assumptions A,
AΜ„
and (B1') follows from the uniqueness
πœˆπœ” in Assumption A
πœˆπœ” (πœ” β€² ) := 𝜈(πœ” βŠ—πœ πœ” β€² ), πœ” β€² ∈ Ξ©, where the concatenated path
and the ow property of (4.1); in particular, the control
can be dened by
πœ” βŠ—πœ πœ” β€²
is given by
(
)
(πœ” βŠ—πœ πœ” β€² )π‘Ÿ := πœ”π‘Ÿ 1[0,𝜏 (πœ”)] (π‘Ÿ) + πœ”π‘Ÿβ€² βˆ’ πœ”πœβ€² (πœ”) + πœ”πœ (πœ”) 1(𝜏 (πœ”),𝑇 ] (π‘Ÿ).
While we refer to [3, Proposition 5.4] for a detailed proof, we emphasize
that the choice of
𝒰𝑠
is crucial for the validity of (2.10): in the notation of
Assumption B', (2.10) essentially requires that
Let
𝑑, 𝜏, 𝜈, 𝜈¯
𝜈¯
be independent of
β„±πœ .
be as in Assumption B', we show that (B2') holds. Let
be a càdlàg version of
πœˆΛ†
¯ (π‘Ÿ) := 𝐸[𝑔(𝑋𝑑,π‘₯
𝑀
(𝑇 ))βˆ£β„±π‘Ÿ ], π‘Ÿ ∈ [0, 𝑇 ],
18
where
πœˆΛ† := 𝜈1[0,𝜏 ] + 1(𝜏,𝑇 ] 𝜈¯.
¯
𝑀
By the same argument as in [3, Proposition 5.4], we deduce from the uniqueness and the ow property of (4.1) and the fact that
𝜈¯
is independent of
β„±πœ
that
πœˆΛ†
𝜈¯
𝐸[𝑔(𝑋𝑑,π‘₯
(𝑇 ))βˆ£β„±π‘Ÿ ] = 𝐸[𝑔(π‘‹πœ,𝑋
𝜈
𝑑,π‘₯ (𝜏 )
𝜈
(𝑇 ))βˆ£β„±π‘Ÿ ] = π‘€πœπœˆ¯ [𝑋𝑑,π‘₯
(𝜏 )](π‘Ÿ)
on
[𝜏, 𝑇 ].
¯ = 𝑀 𝜈¯ [𝑋 𝜈 (𝜏 )] on [𝜏, 𝑇 ]. The last assertion of (B2') is clear by the
𝑀
𝜏
𝑑,π‘₯
denition of ℳ𝑑,0 . As already mentioned in Remark 2.3, Assumption (B3')
Hence
follows from Assumption A and Assumption B follows from Assumption B'.
𝐹 satises (2.16); i.e., that 𝐹 is l.s.c. For xed 𝜈 ∈ 𝒰 ,
𝐿2 -continuous. Hence the semicontinuity from (4.3)
𝜈
+
and Fatou's lemma yield that (𝑑, π‘₯) 7β†’ 𝐸[𝑓 (𝑋𝑑,π‘₯ (𝑇 )) ] is l.s.c. By the sub𝜈
βˆ’ : (𝑑, π‘₯) ∈ 𝐡} is
quadratic growth from (4.3), we have that {𝑓 (𝑋𝑑,π‘₯ (𝑇 ))
uniformly integrable whenever 𝐡 βŠ‚ S is bounded, hence the semicontinuity
𝜈
βˆ’
of 𝑓 also yields that (𝑑, π‘₯) 7β†’ 𝐸[𝑓 (𝑋𝑑,π‘₯ (𝑇 )) ] is u.s.c. As a result, 𝐹 is l.s.c.
The same arguments show that 𝐺 also satises (2.16).
Next, we check that
𝜈 (𝑇 )
(𝑑, π‘₯) 7β†’ 𝑋𝑑,π‘₯
is
4.2 PDE for Expectation Constraints
In this section, we show how to deduce the Hamilton-Jacobi-Bellman PDE
for the optimal control problem (2.2) with expectation constraint from the
weak dynamic programming principle stated in Theorem 2.4. Given a suit-
πœ‘(𝑑, π‘₯) on [0, 𝑇 ] × β„π‘‘ , we shall denote by βˆ‚π‘‘ πœ‘
2
its derivative with respect to 𝑑 and by π·πœ‘ and 𝐷 πœ‘ the Jacobian and the
Hessian matrix with respect to π‘₯, respectively.
ably dierentiable function
In the context of the setup introduced in the preceding Section 4.1, the
Hamilton-Jacobi-Bellman operator is given by
𝐻(π‘₯, 𝑝, 𝑄) :=
inf
(
(𝑒,π‘Ž)βˆˆπ‘ˆ ×ℝ𝑑
)
βˆ’ 𝐿𝑒,π‘Ž (π‘₯, 𝑝, 𝑄) ,
(π‘₯, 𝑝, 𝑄) ∈ ℝ𝑑 × β„π‘‘+1 × π•„π‘‘+1 ,
where
1
⊀
𝐿𝑒,π‘Ž (π‘₯, 𝑝, 𝑄) := πœ‡π‘‹,𝑀 (π‘₯, 𝑒)⊀ 𝑝+ Tr[πœŽπ‘‹,𝑀 πœŽπ‘‹,𝑀
(π‘₯, 𝑒, π‘Ž)𝑄],
2
(𝑒, π‘Ž) ∈ π‘ˆ ×ℝ𝑑
is the Dynkin operator with coecients
(
πœ‡π‘‹,𝑀 (π‘₯, 𝑒) :=
Since the set
π‘ˆ × β„π‘‘
πœ‡(π‘₯, 𝑒)
0
)
(
and
is unbounded,
𝐻
πœŽπ‘‹,𝑀 (π‘₯, 𝑒, π‘Ž) :=
𝜎(π‘₯, 𝑒)
π‘ŽβŠ€
)
.
may be discontinuous and viscos-
ity solution properties need to be stated in terms of the upper and lower
semicontinuous envelopes of
𝐻,
𝐻 βˆ— (π‘₯, 𝑝, 𝑄) :=
lim sup
𝐻(π‘₯β€² , 𝑝′ , 𝑄′ ),
(π‘₯β€² ,𝑝′ ,𝑄′ )β†’(π‘₯,𝑝,𝑄)
π»βˆ— (π‘₯, 𝑝, 𝑄) :=
lim inf
(π‘₯β€² ,𝑝′ ,𝑄′ )β†’(π‘₯,𝑝,𝑄)
19
𝐻(π‘₯β€² , 𝑝′ , 𝑄′ ).
The value function
𝑉
dened in (2.2) may also be discontinuous and so we
introduce
𝑉 βˆ— (𝑑, π‘₯, π‘š) :=
𝑉 (𝑑′ , π‘₯β€² , π‘šβ€² ),
lim sup
β€²
β€²
β€²
(𝑑 , π‘₯ , π‘š ) β†’ (𝑑, π‘₯, π‘š)
(𝑑′ , π‘₯β€² , π‘šβ€² ) ∈ int D
π‘‰βˆ— (𝑑, π‘₯, π‘š) :=
lim inf
(𝑑′ , π‘₯β€² , π‘šβ€² ) β†’ (𝑑, π‘₯, π‘š)
(𝑑′ , π‘₯β€² , π‘šβ€² ) ∈ int D
𝑉 (𝑑′ , π‘₯β€² , π‘šβ€² ).
Here int D denotes the parabolic interior; i.e, the interior of D βˆ– {𝑑 = 𝑇 } in
SΜ‚, where {𝑑 = 𝑇 } := {(𝑑, π‘₯, π‘š) ∈ SΜ‚ : 𝑑 = 𝑇 }. Moreover, we shall denote by
D the closure of D. The main result of this subsection is the following PDE.
Theorem 4.2.
Assume that 𝑉 is locally bounded on int D.
(i) The function 𝑉 βˆ— is a viscosity subsolution on D βˆ– {𝑑 = 𝑇 } of
βˆ’βˆ‚π‘‘ πœ‘ + π»βˆ— (β‹…, π·πœ‘, 𝐷2 πœ‘) ≀ 0.
(ii) The function π‘‰βˆ— is a viscosity supersolution on int D of
βˆ’βˆ‚π‘‘ πœ‘ + 𝐻 βˆ— (β‹…, π·πœ‘, 𝐷2 πœ‘) β‰₯ 0.
We refer to [5] for the various equivalent denitions of viscosity superand subsolutions. We merely mention that subsolution on
𝐴
means that
𝐴 which are local maxima of
𝑉 βˆ— βˆ’πœ‘ on 𝐴, where πœ‘ is a test function, and analogously for the supersolution.
the subsolution property is satised at points of
We shall not discuss in this generality the boundary condition and the
validity of a comparison principle. In the subsequent section, these will be
studied in some detail for the case of state constraints. We also refer to [1]
for the study of the boundary conditions in a similar framework.
Remark 4.3.
We observe that the domain of the PDE in Theorem 4.2 is
not given a priori; it is itself characterized by a control problem: if we dene
𝜈
𝑣(𝑑, π‘₯) := inf 𝐸[𝑔(𝑋𝑑,π‘₯
(𝑇 ))],
πœˆβˆˆπ’°π‘‘
(𝑑, π‘₯) ∈ S,
(4.4)
then
{
}
int D = (𝑑, π‘₯, π‘š) ∈ SΜ‚ : π‘š > 𝑣 βˆ— (𝑑, π‘₯), 𝑑 < 𝑇 ,
𝑣 βˆ— is the upper semicontinuous envelope of 𝑣 on [0, 𝑇 ) × β„π‘‘ . In particint D βˆ•= βˆ… since 𝑣 is locally bounded from above. In fact, in the present
where
ular,
setup, we also have
{
}
int D = (𝑑, π‘₯, π‘š) ∈ SΜ‚ : π‘š > 𝑣(𝑑, π‘₯), 𝑑 < 𝑇 .
(4.5)
Indeed, a well known randomization argument (e.g., [3, Remark 5.2]) yields
that
𝜈 (𝑇 ))]
𝑣(𝑑, π‘₯) = inf πœˆβˆˆπ’° 𝐸[𝑔(𝑋𝑑,π‘₯
for all
20
(𝑑, π‘₯) ∈ S;
i.e., the set
𝒰𝑑
in (4.4)
𝒰.
𝑣 inherits the upper semicontinuity of 𝐺
𝑣 = 𝑣 βˆ— . Using (4.5), we obtain that
{
} {
}
(𝑑, π‘₯, π‘š) ∈ SΜ‚ : π‘š β‰₯ 𝑣(𝑑, π‘₯) βŠ† (𝑑, π‘₯, π‘š) ∈ SΜ‚ : π‘š β‰₯ π‘£βˆ— (𝑑, π‘₯) = int D,
can be replaced by
Therefore,
(c.f. Lemma 4.1) and we have
where
π‘£βˆ—
is the lower semicontinuous envelope of
𝑣
on
[0, 𝑇 ] × β„π‘‘ ,
D βŠ† int D.
If furthermore
(𝑑, π‘₯),
𝑣
𝑣
𝑣,
(4.6)
is continuous and the inmum in (4.4) is attained for all
then the converse inclusion is also satised. In applications, it may be
desirable to have continuity of
by
and hence
𝑣
so that
int D
and
D
are described directly
(rather than a semicontinuous envelope). To analyze the continuity of
one can study the comparison principle for the Hamilton-Jacobi-Bellman
equation associated with the control problem (4.4).
tions 5 and 6] for the explicit computation of
𝑣
We refer to [8, Sec-
in an example from Mathe-
matical Finance.
The rest of this subsection is devoted to the proof of Theorem 4.2. We
rst state a version of Theorem 2.4 which is suitable to derive the PDE.
Lemma 4.4.
(i) Let 𝐡 be an open neighborhood of a point (𝑑, π‘₯, π‘š) ∈ D
such that 𝑉 (𝑑, π‘₯, π‘š) < ∞ and let πœ‘ : 𝐡 β†’ ℝ be a continuous function such
that 𝑉 ≀ πœ‘ on 𝐡 . For all πœ€ > 0 there exist (𝜈, 𝛼) ∈ 𝒰𝑑 × π’œπ‘‘ such that
[
]
𝜈
𝛼
𝛼
𝜈
𝑉 (𝑑, π‘₯, π‘š) ≀ 𝐸 πœ‘(𝜏, 𝑋𝑑,π‘₯
(𝜏 ), 𝑀𝑑,π‘š
(𝜏 )) + πœ€ and 𝑀𝑑,π‘š
(𝑇 ) β‰₯ 𝑔(𝑋𝑑,π‘₯
(𝑇 )),
𝜈 (𝑠), 𝑀 𝛼 (𝑠))
where 𝜏 is the rst exit time of (𝑠, 𝑋𝑑,π‘₯
𝑠β‰₯𝑑 from 𝐡 .
𝑑,π‘š
(ii) Let 𝐡 be an open neighborhood of a point (𝑑, π‘₯, π‘š) ∈ int D such that
𝐡 βŠ† D and let (𝜈, 𝛼) ∈ 𝒰𝑑 × π’œπ‘‘ . For any continuous function πœ‘ : 𝐡 β†’ ℝ
satisfying 𝑉 β‰₯ πœ‘ on 𝐡 and for any πœ€ > 0,
[
]
𝜈
𝛼
𝑉 (𝑑, π‘₯, π‘š + πœ€) β‰₯ 𝐸 πœ‘(𝜏, 𝑋𝑑,π‘₯
(𝜏 ), 𝑀𝑑,π‘š
(𝜏 )) ,
(4.7)
𝜈 (𝑠), 𝑀 𝛼 (𝑠))
where 𝜏 is the rst exit time of (𝑠, 𝑋𝑑,π‘₯
𝑠β‰₯𝑑 from 𝐡 .
𝑑,π‘š
Proof.
In view of Lemmata 2.2 and 4.1, part (i) is immediate from Theo-
rem 2.4(i).
For part (ii) we use the extension of Theorem 2.4(ii') as stated in Corollary 2.7 with
π’Ÿβˆ©D=π’Ÿ
π’Ÿ := 𝐡 .
Note that
is closed and hence
𝜈 (𝜏 ), 𝑀 𝛼 (𝜏 )) ∈ 𝐡 βŠ† D; in particular,
(𝜏, 𝑋𝑑,π‘₯
𝑑,π‘₯
𝜎 -compact.
We can now deduce the PDE for
𝑉
from the dynamic programming
principle in the form of Lemma 4.4. Although the arguments are the usual
ones, we shall indicate the proof, in particular to show that the relaxation
π‘š
+ πœ€
in (4.7) does not aect the PDE.
21
Proof of Theorem 4.2. (i) We rst prove the subsolution property. Let πœ‘
𝐢 1,2 -function and let (𝑑0 , π‘₯0 , π‘š0 ) ∈ D be such that 𝑑0 ∈ (0, 𝑇 ) and
(𝑑0 , π‘₯0 , π‘š0 ) is a maximum point of 𝑉 βˆ— βˆ’ πœ‘ satisfying
be a
(𝑉 βˆ— βˆ’ πœ‘)(𝑑0 , π‘₯0 , π‘š0 ) = 0.
(4.8)
Assume for contradiction that
(
)
βˆ’ βˆ‚π‘‘ πœ‘ + π»βˆ— (β‹…, π·πœ‘, 𝐷2 πœ‘) (𝑑0 , π‘₯0 , π‘š0 ) > 0.
D = int D by (4.6),
(𝑑0 , π‘₯0 , π‘š0 ) such that
Since
of
there exists a bounded open neighborhood
βˆ’βˆ‚π‘‘ πœ‘¯ βˆ’ 𝐿𝑒,π‘Ž (β‹…, π·πœ‘,
¯ 𝐷2 πœ‘)
¯ >0
on
𝐡 ∩ int D,
for all
𝐡 βŠ‚ SΜ‚
(𝑒, π‘Ž) ∈ π‘ˆ × β„π‘‘ ,
(4.9)
where
(
)
πœ‘(𝑠,
¯ 𝑦, 𝑛) := πœ‘(𝑑0 , π‘₯0 , π‘š0 ) + βˆ£π‘  βˆ’ 𝑑0 ∣2 + βˆ£π‘¦ βˆ’ π‘₯0 ∣4 + βˆ£π‘› βˆ’ π‘š0 ∣4 .
Moreover, we have
πœ‚ := min(πœ‘¯ βˆ’ πœ‘) > 0.
(4.10)
βˆ‚π΅
Given
πœ€ > 0,
let
(π‘‘πœ€ , π‘₯πœ€ , π‘šπœ€ ) ∈ 𝐡 ∩ int D
be such that
𝑉 (π‘‘πœ€ , π‘₯πœ€ , π‘šπœ€ ) β‰₯ 𝑉 βˆ— (𝑑0 , π‘₯0 , π‘š0 ) βˆ’ πœ€.
(4.11)
(𝜈, 𝛼) ∈ 𝒰𝑑 × π’œπ‘‘ such that π‘€π‘‘π›Όπœ€ ,π‘šπœ€ (𝑇 ) β‰₯ 𝑔(π‘‹π‘‘πœˆπœ€ ,π‘₯πœ€ (𝑇 )) and
𝜈
𝛼
let 𝜏 be the rst exit time of (𝑠, 𝑋𝑑 ,π‘₯ (𝑠), 𝑀𝑑 ,π‘š (𝑠))𝑠β‰₯𝑑 from 𝐡 . We recall
πœ€ πœ€
πœ€
πœ€
𝜈
𝛼
from Remark 2.3(ii) that (𝑠, 𝑋𝑑 ,π‘₯ (𝑠), 𝑀𝑑 ,π‘š (𝑠))𝑠β‰₯𝑑 remains in D on [𝑑, 𝑇 ],
πœ€ πœ€
πœ€
πœ€
and hence also in int D by (4.6). Now, it follows from Itô's formula and (4.9)
Consider arbitrary
that
[
]
πœ‘(𝑑
¯ πœ€ , π‘₯πœ€ , π‘šπœ€ ) β‰₯ 𝐸 πœ‘(𝜏,
¯ π‘‹π‘‘πœˆπœ€ ,π‘₯πœ€ (𝜏 ), π‘€π‘‘π›Όπœ€ ,π‘šπœ€ (𝜏 )) .
For
(π‘‘πœ€ , π‘₯πœ€ , π‘šπœ€ )
close enough to
(𝑑0 , π‘₯0 , π‘š0 ),
this implies that
]
πœ‘(𝑑
¯ 0 , π‘₯0 , π‘š0 ) β‰₯ 𝐸 πœ‘(𝜏,
¯ π‘‹π‘‘πœˆπœ€ ,π‘₯πœ€ (𝜏 ), π‘€π‘‘π›Όπœ€ ,π‘šπœ€ (𝜏 )) βˆ’ π‘œ(1)
[
as
πœ€ β†’ 0,
which, by (4.8), (4.10) and (4.11), leads to
[
]
𝑉 (π‘‘πœ€ , π‘₯πœ€ , π‘šπœ€ ) β‰₯ 𝐸 πœ‘(𝜏, π‘‹π‘‘πœˆπœ€ ,π‘₯πœ€ (𝜏 ), π‘€π‘‘π›Όπœ€ ,π‘šπœ€ (𝜏 )) + πœ‚ βˆ’ π‘œ(1).
This contradicts Lemma 4.4(i) for
πœ€>0
small enough.
(ii) We now prove the supersolution property. Let
and let
π‘‰βˆ— βˆ’ πœ‘
(𝑑0 , π‘₯0 , π‘š0 ) ∈ int D
be such that
(𝑑0 , π‘₯0 , π‘š0 )
πœ‘
be a
𝐢 1,2 -function
is a minimum point of
satisfying
(π‘‰βˆ— βˆ’ πœ‘)(𝑑0 , π‘₯0 , π‘š0 ) = 0.
Assume for contradiction that
(
)
βˆ’ βˆ‚π‘‘ πœ‘ + 𝐻 βˆ— (β‹…, π·πœ‘, 𝐷2 πœ‘) (𝑑0 , π‘₯0 , π‘š0 ) < 0.
22
(4.12)
Then there exist
(𝑑0 , π‘₯0 , π‘š0 )
such
(Λ†
𝑒, π‘Ž
Λ†) ∈ π‘ˆ × β„π‘‘ and
that 𝐡 βŠ† int D and
a bounded open neighborhood
βˆ’βˆ‚π‘‘ πœ‘¯ βˆ’ 𝐿𝑒ˆ,Λ†π‘Ž (β‹…, π·πœ‘,
¯ 𝐷2 πœ‘)
¯ <0
on
𝐡,
𝐡
of
(4.13)
where
(
)
πœ‘(𝑠,
¯ 𝑦, 𝑛) := πœ‘(𝑑0 , π‘₯0 , π‘š0 ) βˆ’ βˆ£π‘  βˆ’ 𝑑0 ∣2 + βˆ£π‘¦ βˆ’ π‘₯0 ∣4 + βˆ£π‘› βˆ’ π‘š0 ∣4 .
Note that
πœ‚ := min(πœ‘ βˆ’ πœ‘)
¯ > 0.
(4.14)
βˆ‚π΅
Given
πœ€ > 0,
let
(π‘‘πœ€ , π‘₯πœ€ , π‘šπœ€ ) ∈ 𝐡
be such that
𝑉 (π‘‘πœ€ , π‘₯πœ€ , π‘šπœ€ + πœ€) ≀ π‘‰βˆ— (𝑑0 , π‘₯0 , π‘š0 ) + πœ€.
Viewing
(Λ†
𝑒, π‘Ž
Λ†)
(4.15)
as a constant control, it follows from Itô's formula and (4.13)
that
[
]
πœ‘(𝑑
¯ πœ€ , π‘₯πœ€ , π‘šπœ€ ) ≀ 𝐸 πœ‘(𝜏,
¯ π‘‹π‘‘π‘’Λ†πœ€ ,π‘₯πœ€ (𝜏 ), π‘€π‘‘π‘ŽΛ†πœ€ ,π‘šπœ€ (𝜏 )) ,
𝜏 is the rst exit time of (𝑠, π‘‹π‘‘π‘’Λ†πœ€ ,π‘₯πœ€ (𝑠), π‘€π‘‘π‘ŽΛ†πœ€ ,π‘šπœ€ (𝑠))𝑠β‰₯𝑑 from 𝐡 .
(π‘‘πœ€ , π‘₯πœ€ , π‘šπœ€ ) close enough to (𝑑0 , π‘₯0 , π‘š0 ), this implies that
[
]
πœ‘(𝑑
¯ 0 , π‘₯0 , π‘š0 ) ≀ 𝐸 πœ‘(𝜏,
¯ π‘‹π‘‘π‘’Λ†πœ€ ,π‘₯πœ€ (𝜏 ), π‘€π‘‘π‘ŽΛ†πœ€ ,π‘šπœ€ (𝜏 )) + π‘œ(1) as πœ€ β†’ 0,
where
For
which, by (4.12), (4.14) and (4.15), leads to
[
]
𝑉 (π‘‘πœ€ , π‘₯πœ€ , π‘šπœ€ + πœ€) ≀ 𝐸 πœ‘(𝜏, π‘‹π‘‘π‘’Λ†πœ€ ,π‘₯πœ€ (𝜏 ), π‘€π‘‘π‘ŽΛ†πœ€ ,π‘šπœ€ (𝜏 )) βˆ’ πœ‚ + π‘œ(1).
πœ€>0
For
small enough, this yields a contradiction to Lemma 4.4(ii).
4.3 PDE for State Constraints
In this section, we discuss the Hamilton-Jacobi-Bellman PDE for the state
constraint problem (cf. Section 3) in the case where the state process is given
by a controlled SDE as introduced in Section 4.1 and required to stay in an
open set
π’ͺ βŠ† ℝ𝑑 .
Note that in this setup, the continuity conditions (3.2)
and (3.3) are satised.
We shall derive the PDE via Theorem 3.1.
The basic idea to guaran-
tee its condition (3.6) about right continuity in the constraint level runs as
follows.
Consider a control
small probability
𝛿 β‰₯ 0.
𝜈 ∈ 𝒰𝑑
such that
𝜈
𝑋𝑑,π‘₯
state constraint, by switching to some admissible
𝜈
𝑋𝑑,π‘₯
exits
π’ͺ.
π’ͺ with at most
𝜈 𝛿 , satisfying the
control 𝜈
Λ† shortly before
leaves
Then we shall construct a control
As a result,
πœˆπ›Ώ
coincides with
𝜈
on a set of large probability
and therefore the reward is similar. Along the lines of Lemma 2.8 we shall
then obtain the desired right continuity (cf. Lemma 4.7 below).
To make this work, we clearly need to have
[0, 𝑇 ] × π’ͺ,
¯ π‘₯) βˆ•= βˆ…
𝒰(𝑑,
for all
which is anyway necessary for the value function
23
𝑉¯
(𝑑, π‘₯)
in
from (3.1)
to be nite. However, we need a slightly stronger condition; namely, that
we can switch to an admissible control in a measurable way. A particularly
simple condition ensuring this, is the existence of an admissible feedback
control:
Assumption C.
such that, for all
Λ†
𝑋(𝑠)
= π‘₯+
∫
𝑠
𝑒
Λ†:π’ͺβ†’π‘ˆ
There exists a Lipschitz continuous mapping
(𝑑, π‘₯) ∈ [0, 𝑇 ] × π’ͺ,
the solution
(
)
Λ†
Λ†
πœ‡ 𝑋(π‘Ÿ),
𝑒
Λ†(𝑋(π‘Ÿ))
π‘‘π‘Ÿ +
𝑠
∫
Λ† 𝑑,π‘₯
𝑋
of
(
)
Λ†
Λ†
𝜎 𝑋(π‘Ÿ),
𝑒
Λ†(𝑋(π‘Ÿ))
π‘‘π‘Šπ‘Ÿ ,
𝑠 ∈ [𝑑, 𝑇 ]
𝑑
𝑑
(4.16)
satises
Λ† 𝑑,π‘₯ (𝑠) ∈ π’ͺ
𝑋
for all
𝑠 ∈ [𝑑, 𝑇 ], 𝑃 -a.s.
πœ‡(β‹…, 𝑒0 ) = 0 and 𝜎(β‹…, 𝑒0 ) = 0 for some 𝑒0 ∈ π‘ˆ , then Assumption C
is clearly satised for 𝑒
Λ† ≑ 𝑒0 . Or, under an additional smoothness condition,
Λ† 𝑑,π‘₯ will stay in π’ͺ if the Lipschitz function 𝑒
𝑋
Λ† satises
(
)
1
⊀
⊀
βˆ£π‘›πœŽβˆ£(β‹…, 𝑒
Λ†) = 0 and
𝑛 πœ‡ + Trace[𝐷𝑛 𝜎𝜎 ] (β‹…, 𝑒
Λ†) > 0 on βˆ‚π’ͺ,
2
If, e.g.,
where
𝑛
denotes the inner normal to
βˆ‚π’ͺ;
see also [11, Proposition 3.1] and
[7, Lemma III.4.3].
The following is a simple condition guaranteeing the uniform integrability
required in (2.31).
Assumption D.
Either
𝑓
πœ‡(π‘₯, 𝑒)
𝑒.
is bounded or the coecients
in the SDE (4.1) have linear growth in
π‘₯,
uniformly in
This assumption holds in particular if the control domain
Remark 4.5.
Assumption C implies that
and Assumption D implies that
𝑉¯
𝑉¯
π‘ˆ
and
𝜎(π‘₯, 𝑒)
is bounded.
is locally bounded from below
is locally bounded from above.
Next, we introduce the notation for the PDE related to the value function
𝑉¯
from (3.1). The associated Hamilton-Jacobi-Bellman operator is given by
¯
𝐻(π‘₯,
𝑝, 𝑄) := inf
π‘’βˆˆπ‘ˆ
(
)
¯ 𝑒 (π‘₯, 𝑝, 𝑄) ,
βˆ’πΏ
(π‘₯, 𝑝, 𝑄) ∈ ℝ𝑑 × β„π‘‘ × π•„π‘‘ ,
where the Dynkin operator is dened by
¯ 𝑒 (π‘₯, 𝑝, 𝑄) := πœ‡(π‘₯, 𝑒)⊀ 𝑝 + 1 Tr[𝜎𝜎 ⊀ (π‘₯, 𝑒)𝑄],
𝐿
2
𝑒 ∈ π‘ˆ.
Similarly as above, we introduce the semicontinuous envelopes
¯ βˆ— (π‘₯, 𝑝, 𝑄) :=
𝐻
lim sup
¯ β€² , 𝑝′ , 𝑄′ ),
𝐻(π‘₯
(π‘₯β€² , 𝑝′ , 𝑄′ ) β†’ (π‘₯, 𝑝, 𝑄)
(π‘₯β€² , 𝑝′ , 𝑄′ ) ∈ π’ͺ × β„π‘‘ × π•„π‘‘
¯ βˆ— (π‘₯, 𝑝, 𝑄) :=
𝐻
lim inf
(π‘₯β€² , 𝑝′ , 𝑄′ ) β†’ (π‘₯, 𝑝, 𝑄)
(π‘₯β€² , 𝑝′ , 𝑄′ ) ∈ π’ͺ × β„π‘‘ × π•„π‘‘
24
¯ β€² , 𝑝′ , 𝑄′ )
𝐻(π‘₯
(4.17)
and
𝑉¯ βˆ— (𝑑, π‘₯) :=
𝑉¯ (𝑑′ , π‘₯β€² ),
lim sup
(𝑑′ , π‘₯β€² ) β†’ (𝑑, π‘₯)
(𝑑′ , π‘₯β€² ) ∈ [0, 𝑇 ) × π’ͺ
𝑉¯βˆ— (𝑑, π‘₯) :=
lim inf
(𝑑′ , π‘₯β€² ) β†’ (𝑑, π‘₯)
(𝑑′ , π‘₯β€² ) ∈ [0, 𝑇 ) × π’ͺ
𝑉¯ (𝑑′ , π‘₯β€² ).
We can now state the Hamilton-Jacobi-Bellman PDE; the boundary condition is discussed in Proposition 4.11 below.
Theorem 4.6.
Assume that 𝑉¯ is locally bounded on [0, 𝑇 ) × π’ͺ.
(i) The function 𝑉¯ βˆ— is a viscosity subsolution on [0, 𝑇 ) × π’ͺ of
¯ βˆ— (β‹…, π·πœ‘, 𝐷2 πœ‘) ≀ 0.
βˆ’βˆ‚π‘‘ πœ‘ + 𝐻
(4.18)
(ii) Under Assumptions C and D, the function 𝑉¯βˆ— is a viscosity supersolution on [0, 𝑇 ) × π’ͺ of
¯ βˆ— (β‹…, π·πœ‘, 𝐷2 πœ‘) β‰₯ 0.
βˆ’βˆ‚π‘‘ πœ‘ + 𝐻
The proof is given below, after some auxiliary results.
the right-continuity condition (3.6) for the value function
We rst verify
𝑉 (𝑑, π‘₯, 𝑦, π‘š) intro-
duced in Section 3.
Lemma 4.7.
Let Assumptions C and D hold true. Then 𝑉 (𝑑, π‘₯, 1, 0+) =
𝑉 (𝑑, π‘₯, 1, 0) for all (𝑑, π‘₯) ∈ [0, 𝑇 ] × π’ͺ.
Proof. For 𝛿 > 0, let 𝜈 𝛿 ∈ 𝒰(𝑑, π‘₯, 𝛿) be as in Lemma 2.8. Then the process
𝜈 𝛿 dened in (3.5) satises π‘Œ 𝜈 𝛿 (𝑇 ) > 0 outside of a set of measure at
π‘Œπ‘‘,π‘₯,𝑦
𝑑,π‘₯,1
most 𝛿 . It follows that we can nd πœ€π›Ώ ∈ (0, 1 ∧ 𝑑(π‘₯)) such that the set
𝜈 (𝑇 ) ≀ πœ€ } satises 𝑃 [𝐴𝛿 ] ≀ 2𝛿 . Let 𝜏 𝛿 denote the rst time
𝐴𝛿 := {π‘Œπ‘‘,π‘₯,1
𝛿
πœˆπ›Ώ
when π‘Œπ‘‘,π‘₯,1 reaches the level πœ€π›Ώ and set
Λ† 𝛿 ),
Λ†(𝑋
πœˆΛœπ›Ώ := 𝜈 𝛿 1[𝑑,𝜏 𝛿 ] + 1(𝜏 𝛿 ,𝑇 ] 𝑒
Λ† 𝛿 is the solution of (4.16) on [𝜏 𝛿 , 𝑇 ] with initial condition given
𝑋
𝜈 𝛿 (𝜏 𝛿 ). Since the paths of π‘Œ 𝜈 𝛿 are nonincreasing, we have
Λ† 𝛿 (𝜏 𝛿 ) = 𝑋𝑑,π‘₯
𝑋
𝑑,π‘₯,1
where
𝛿
by
𝛿
𝜈˜
𝜈
lim 𝑃 [𝑋𝑑,π‘₯
(𝑇 ) βˆ•= 𝑋𝑑,π‘₯
(𝑇 )] ≀ lim 𝑃 [𝜏 𝛿 ≀ 𝑇 ] = lim 𝑃 [𝐴𝛿 ] = 0.
𝛿↓0
Next, we check that
trivial if
𝑓
𝛿↓0
𝜈 (𝑇 )), 𝜈 ∈ 𝒰}
{𝑓 (𝑋𝑑,π‘₯
𝛿↓0
is uniformly integrable.
This is
is bounded. Otherwise, Assumption D yields that the coecients
πœ‡(π‘₯, 𝑒) and 𝜎(π‘₯, 𝑒) have uniformly linear growth in π‘₯, and of course they are
𝜈
uniformly Lipschitz in π‘₯ as they are jointly Lipschitz. Thus {𝑋𝑑,π‘₯ (𝑇 ), 𝜈 ∈ 𝒰}
𝑝
is bounded in 𝐿 for any nite 𝑝 and the uniform integrability follows from the
quadratic growth assumption (4.2) on 𝑓 . It remains to apply Lemma 2.8.
25
Lemma 4.8.
Let Assumption C hold true and let 𝐡 be an open neighborhood
of (𝑑, π‘₯) ∈ [0, 𝑇 ] × π’ͺ. For all 𝜈 ∈ 𝒰𝑑 there exists 𝜈¯ ∈ 𝒰𝑑 such that
¯ π‘₯),
𝜈1[𝑑,𝜏 ] + 𝜈¯1(𝜏,𝑇 ] ∈ 𝒰(𝑑,
𝜈 (𝑠))
where 𝜏 is the rst exit time of (𝑠, 𝑋𝑑,π‘₯
𝑠β‰₯𝑑 from 𝐡 .
Proof.
Let
Λ† 𝜏,𝑋 𝜈 (𝜏 )
𝑋
𝑑,π‘₯
be the solution of (4.16) on
𝜈 (𝜏 ) at time 𝜏 .
𝑋𝑑,π‘₯
Λ† 𝜏,𝑋 𝜈 (𝜏 ) ).
𝜈¯ := 𝜈1[𝑑,𝜏 ] + 1(𝜏,𝑇 ] 𝑒
Λ†(𝑋
𝑑,π‘₯
integrable) initial condition
[𝜏, 𝑇 ]
with the (square
Then the claim holds true for
We have the following counterpart of Lemma 4.4.
Lemma 4.9.
(i) Let 𝐡 βŠ† [0, 𝑇 ] × β„π‘‘ be an open neighborhood of a point
(𝑑, π‘₯) ∈ [0, 𝑇 ]×π’ͺ such that 𝑉¯ (𝑑, π‘₯) is nite and let πœ‘ : 𝐡 β†’ ℝ be a continuous
¯ π‘₯) such
function such that 𝑉¯ ≀ πœ‘ on 𝐡 . For all πœ€ > 0 there exists 𝜈 ∈ 𝒰(𝑑,
that
[
]
𝜈
𝑉¯ (𝑑, π‘₯) ≀ 𝐸 πœ‘(𝜏, 𝑋𝑑,π‘₯
(𝜏 )) + πœ€,
𝜈 (𝑠))
where 𝜏 is the rst exit time of (𝑠, 𝑋𝑑,π‘₯
𝑠β‰₯𝑑 from 𝐡 .
(ii) Let Assumptions C and D hold true and let 𝐡 βŠ† [0, 𝑇 ] × π’ͺ be an
open neighborhood of (𝑑, π‘₯). For any 𝜈 ∈ 𝒰𝑑 and any continuous function
πœ‘ : 𝐡 β†’ ℝ satisfying 𝑉¯ β‰₯ πœ‘ on 𝐡 ,
[
]
𝜈
𝑉¯ (𝑑, π‘₯) β‰₯ 𝐸 πœ‘(𝜏, 𝑋𝑑,π‘₯
(𝜏 )) ,
𝜈 (𝑠))
where 𝜏 is the rst exit time of (𝑠, 𝑋𝑑,π‘₯
𝑠β‰₯𝑑 from 𝐡 .
Proof.
In view of Lemma 4.1, part (i) is immediate from Theorem 3.1(i).
Part (ii) follows from Theorem 3.1(ii) via Lemmata 4.7 and 4.8 .
Proof of Theorem 4.6.
The result follows from Lemma 4.9 by the arguments
in the proof of Theorem 4.2.
4.3.1 Boundary Condition and Uniqueness
In this section, we discuss the boundary condition and the uniqueness for
the PDE in Theorem 4.6. We shall work under a slightly stronger condition
on our setup.
Assumption D'.
linear growth in
The coecients
π‘₯,
uniformly in
πœ‡(π‘₯, 𝑒) and 𝜎(π‘₯, 𝑒) in the SDE (4.1) have
𝑒.
We also introduce the following regularity condition, which will be used
to prove the comparison theorem.
Denition 4.10.
Then
𝑀
is
Consider a set
of class β„›(π’ͺ)
π’ͺ βŠ† ℝ𝑑
if the following hold for any
26
𝑀 : [0, 𝑇 ] × π’ͺ β†’ ℝ.
(𝑑, π‘₯) ∈ [0, 𝑇 ) × βˆ‚π’ͺ:
and a function
(i) There exist
β„“ : ℝ+ β†’
π‘Ÿ > 0,
an open neighborhood
𝐡
of
π‘₯
in
ℝ𝑑
and a function
ℝ𝑑 such that
lim inf πœ€βˆ’1 βˆ£β„“(πœ€)∣ < ∞
πœ€β†’0
𝑦 + β„“(πœ€) + π‘œ(πœ€) ∈ π’ͺ
(ii) There exists a function
and
for all
(4.19)
𝑦 ∈π’ͺ∩𝐡
πœ† : ℝ+ β†’ ℝ+
and
πœ€ ∈ (0, π‘Ÿ).
(4.20)
such that
lim πœ†(πœ€) = 0 and
(
)
lim 𝑀 𝑑 + πœ†(πœ€), π‘₯ + β„“(πœ€) = 𝑀(𝑑, π‘₯).
(4.21)
πœ€β†’0
(4.22)
πœ€β†’0
πœ“ : ℝ+ β†’ ℝ𝑑 is any function of class π‘œ(πœ€),
then there exists π‘Ÿ > 0 such that 𝑦 + β„“(πœ€) + πœ“(πœ€) ∈ π’ͺ for all πœ€ ∈ (0, π‘Ÿ).
Note that (i) is a condition on the boundary of βˆ‚π’ͺ ; it can be seen as a
By (4.20) we mean that if
variant of the interior cone condition where the cone is replaced by a more
general shape. Condition (ii) essentially states that
𝑀
is continuous along
at least one curve approaching the boundary point through this shape. In
Proposition 4.12 below, we indicate a sucient condition for
β„›(π’ͺ),
𝑉¯βˆ—
to be of class
which is stated directly in terms of the given primitives. We shall see
in its proof that Denition 4.10 is well adapted to the problem at hand (see
also Remark A.4 below). Before that, let us state the uniqueness result.
Proposition 4.11. Let 𝑓 be continuous and let Assumptions C and D' hold
true. Then 𝑉¯ has quadratic growth and the boundary condition is attained
in the sense that
𝑉¯ βˆ— (𝑇, β‹…) ≀ 𝑓
and 𝑉¯βˆ— (𝑇, β‹…) β‰₯ 𝑓
on π’ͺ.
Assume in addition that 𝑉¯βˆ— is of class β„›(π’ͺ). Then
(i) 𝑉¯ is continuous on [0, 𝑇 ] × π’ͺ and admits a continuous extension to
[0, 𝑇 ] × π’ͺ,
(ii) 𝑉¯ is the unique (discontinuous) viscosity solution of the state constraint
problem
¯ π·πœ‘, 𝐷2 πœ‘) = 0, πœ‘(𝑇, β‹…) = 𝑓
βˆ’βˆ‚π‘‘ πœ‘ + 𝐻(β‹…,
in the class of functions having polynomial growth and having a lower
semicontinuous envelope of class β„›(π’ͺ).
Proof.
Recalling that
𝑓
has quadratic growth (4.2), it follows from stan-
dard estimates for the SDE (4.1) under Assumptions C and D' that
𝑉¯
has
quadratic growth and satises the boundary condition.
Assumption D' implies Assumption D and hence Theorem 4.6 yields
that
𝑉¯ βˆ—
and
𝑉¯βˆ—
are sub- and supersolutions, respectively.
sumption D' implies that
¯
𝐻
Moreover, As-
is continuous and satises Assumption E in
27
Appendix A; see Lemma A.3 below for the proof.
If
𝑉¯βˆ—
is of class
β„›(π’ͺ),
the comparison principle (Theorem A.1 and the subsequent remark, cf. Appendix A) yields that
𝑉¯ βˆ— = 𝑉¯βˆ—
[0, 𝑇 ]×π’ͺ; in particular, (i) holds.
on
Part (ii)
also follows from the comparison result.
We conclude this section with a sucient condition ensuring that
is of class
β„›(π’ͺ);
𝑉¯βˆ—
the idea is that the volatility should degenerate so that
the state process can be pushed away from the boundary. We remark that
conditions in a similar spirit exist in the previous literature (e.g., [11, 10]);
cf. Remark A.4 below.
Proposition 4.12.
Assume that 𝑉¯βˆ— is nite-valued on [0, 𝑇 ] × π’ͺ and that
π’ͺ, πœ‡ and 𝜎 satisfy the following conditions:
(i) There exists a 𝐢 1 -function 𝛿 , dened on a neighborhood of π’ͺ βŠ† ℝ𝑑 ,
such that 𝐷𝛿 is locally Lipschitz continuous and
𝛿 > 0 on π’ͺ,
𝛿 = 0 on βˆ‚π’ͺ,
𝛿 < 0 outside π’ͺ.
(ii) There exists a locally Lipschitz continuous mapping 𝑒
Λ‡ : ℝ𝑑 β†’ π‘ˆ such
that for all π‘₯ ∈ π’ͺ there exist an open neighborhood 𝐡 of π‘₯ and πœ„ > 0
satisfying
πœ‡(𝑧, 𝑒
Λ‡(𝑧))⊀ 𝐷𝛿(𝑦) β‰₯ πœ„ and 𝜎(𝑦, 𝑒
Λ‡(𝑦)) = 0 for all 𝑦 ∈ 𝐡∩π’ͺ and 𝑧 ∈ 𝐡.
(4.23)
Then 𝑉¯βˆ— is of class β„›(π’ͺ).
Proof.
Fix
(𝑑, π‘₯) ∈ [0, 𝑇 ) × βˆ‚π’ͺ and let 𝛿 , πœ„ and 𝐡
𝐡 is bounded. Consider 𝑦 ∈ π’ͺ ∩ 𝐡 .
assume that
and be as above; we may
Since
π‘₯β€² 7β†’ πœ‡(π‘₯β€² , 𝑒
Λ‡(π‘₯β€² ))
is
locally Lipschitz, the ordinary dierential equation
∫
π‘₯
Λ‡(𝑠) = 𝑦 +
𝑠
(
)
πœ‡ π‘₯
Λ‡(π‘Ÿ), 𝑒
Λ‡(Λ‡
π‘₯(π‘Ÿ)) π‘‘π‘Ÿ
0
has a unique solution
π‘₯
ˇ𝑦 (𝑠) ∈ π’ͺ
For
for
πœ€ ∈ [0, 𝑇π‘₯ ),
π‘₯
ˇ𝑦
on some interval
𝑠 ∈ (0, πœ€],
for
πœ€>0
[0, 𝑇𝑦 ).
Then (4.23) ensures that
small enough, for all
𝑦 ∈ π’ͺ ∩ 𝐡.
(4.24)
we set
β„“(πœ€) := π‘₯
Λ‡π‘₯ (πœ€) βˆ’ π‘₯,
πœ†(πœ€) := πœ€.
Then (4.21) is clearly satised. Using the continuity of
πœ‡
and
𝑒
Λ‡,
we see that
πœ€βˆ’1 βˆ£β„“(πœ€)∣ β†’ πœ‡(π‘₯, 𝑒
Λ‡(π‘₯)),
which implies (4.19). Moreover, using that
πœ‡
and
𝑒
Λ‡
are locally Lipschitz, we
nd that
π‘₯
Λ‡π‘₯ (π‘Ÿ) βˆ’ π‘₯
ˇ𝑦 (π‘Ÿ) = π‘₯ βˆ’ 𝑦 + 𝑂(π‘Ÿ).
28
Together with (4.23), (4.24) and the local Lipschitz continuity of πœ‡, 𝑒
Λ‡ and
𝐷𝛿 , this implies that for 𝑦 ∈ π’ͺ suciently close to π‘₯ and πœ€ > 0 small enough,
(
)
𝛿 𝑦 + β„“(πœ€) + π‘œ(πœ€)
(
)
=𝛿 π‘¦βˆ’π‘₯+π‘₯
Λ‡π‘₯ (πœ€) + π‘œ(πœ€)
∫ πœ€
(
)⊀ (
)
= 𝛿(𝑦) +
πœ‡ π‘₯
Λ‡π‘₯ (π‘Ÿ), 𝑒
Λ‡(Λ‡
π‘₯π‘₯ (π‘Ÿ)) 𝐷𝛿 𝑦 βˆ’ π‘₯ + π‘₯
Λ‡π‘₯ (π‘Ÿ) π‘‘π‘Ÿ + π‘œ(πœ€)
∫ 0πœ€
(
)⊀ (
)
ˇ𝑦 (π‘Ÿ), 𝑒
Λ‡(π‘₯ βˆ’ 𝑦 + π‘₯
ˇ𝑦 (π‘Ÿ)) 𝐷𝛿 π‘₯
ˇ𝑦 (π‘Ÿ) π‘‘π‘Ÿ + 𝑂(πœ€2 ) + π‘œ(πœ€)
= 𝛿(𝑦) + πœ‡ π‘₯ βˆ’ 𝑦 + π‘₯
0
β‰₯ πœ€πœ„ + π‘œ(πœ€),
πœ€ > 0 small enough. This implies (4.20).
(𝑠, 𝑦) ∈ [0, 𝑇 ) × π’ͺ close to (𝑑, π‘₯). For πœ€ > 0 small enough,
¯ + πœ†(πœ€), π‘₯
𝜈 πœ€ ∈ 𝒰(𝑠
ˇ𝑦 (πœ€)) such that
[ ( πœˆπœ€
)]
(
)
𝐸 𝑓 𝑋𝑠+πœ†(πœ€),Λ‡π‘₯𝑦 (πœ€) (𝑇 ) β‰₯ 𝑉¯ 𝑠 + πœ†(πœ€), π‘₯
ˇ𝑦 (πœ€) βˆ’ πœ€.
which is strictly positive for
Consider
can nd
we
Recall the degeneracy condition in (4.23). Setting
𝜈¯πœ€ := 𝑒
Λ‡(Λ‡
π‘₯𝑦 )1[𝑠,𝑠+πœ†(πœ€)] + 1(𝑠+πœ†(πœ€),𝑇 ] 𝜈 πœ€ ,
we obtain that
[ ( 𝜈¯πœ€
)]
𝑉¯ (𝑠, 𝑦) β‰₯ 𝐸 𝑓 𝑋𝑠,𝑦
(𝑇 )
[ ( πœˆπœ€
)]
= 𝐸 𝑓 𝑋𝑠+πœ†(πœ€),Λ‡
π‘₯𝑦 (πœ€) (𝑇 )
(
)
β‰₯ 𝑉¯ 𝑠 + πœ†(πœ€), π‘₯
ˇ𝑦 (πœ€) βˆ’ πœ€.
π‘₯
ˇ𝑦 (πœ€) β†’ π‘₯
Λ‡π‘₯ (πœ€) as 𝑦 β†’ π‘₯, this leads to
(
)
(
)
𝑉¯βˆ— (𝑑, π‘₯) β‰₯ 𝑉¯βˆ— 𝑑 + πœ†(πœ€), π‘₯
Λ‡π‘₯ (πœ€) βˆ’ πœ€ = 𝑉¯βˆ— 𝑑 + πœ†(πœ€), π‘₯ + β„“(πœ€) βˆ’ πœ€
Recalling that
for
πœ€>0
small enough, which implies
(
)
𝑉¯βˆ— (𝑑, π‘₯) β‰₯ lim sup 𝑉¯βˆ— 𝑑 + πœ†(πœ€), π‘₯ + β„“(πœ€)
πœ€β†’0
(
)
β‰₯ lim inf 𝑉¯βˆ— 𝑑 + πœ†(πœ€), π‘₯ + β„“(πœ€)
πœ€β†’0
Hence
β‰₯ 𝑉¯βˆ— (𝑑, π‘₯).
(
)
limπœ€β†’0 𝑉¯βˆ— 𝑑 + πœ†(πœ€), π‘₯ + β„“(πœ€) = 𝑉¯βˆ— (𝑑, π‘₯);
i.e., (4.22) holds for
𝑉¯βˆ— .
4.3.2 On Closed State Constraints
Recall that the value function
𝑉¯
considered above corresponds to the con-
straint that the state processes remains in the open set
π’ͺ.
We can similarly
consider the closed constraint; i.e.,
{
𝜈
𝜈
(𝑇 ))] : 𝜈 ∈ 𝒰𝑑 , 𝑋𝑑,π‘₯
(𝑠) ∈ π’ͺ
𝑉 (𝑑, π‘₯) := sup 𝐸[𝑓 (𝑋𝑑,π‘₯
29
for all
𝑠 ∈ [𝑑, 𝑇 ], 𝑃 -a.s.
}
set, the constraint function
𝑉¯
do not apply to 𝑉 . Indeed, for the closed
𝐺 in Section 3 would not be u.s.c. and hence the
The arguments used above for
derivation of Theorem 3.1 fails; note that the upper semicontinuity is essential for the covering argument in the proof of Theorem 2.4(ii),(ii'). Moreover,
the switching argument in the proof of Lemma 4.8 cannot be imitated since,
given that the state process
𝜈
𝑋𝑑,π‘₯
βˆ‚π’ͺ, it
exit π’ͺ .
hits the boundary
know which trajectories of the state will actually
is not possible to
However, we shall see that, if a comparison principle holds, then the
π’ͺ
dynamic programming principle for the open constraint
characterize the value function
apply the PDE for
𝑉¯
𝑉
associated to
π’ͺ.
is enough to fully
More precisely, we shall
and its comparison principle to deduce that
𝑉 = 𝑉¯
under certain conditions. Of course, the basic observation that this equality
holds under suitable conditions is not new; see, e.g., [10]. We shall use the
following assumption.
Assumption C'. Assumption C holds with 𝑒ˆ dened on π’ͺ.
Corollary 4.13. Let 𝑓 be continuous, let Assumptions C' and D' hold true
and assume that 𝑉¯βˆ— is of class β„›(π’ͺ). Then 𝑉 = 𝑉¯ on [0, 𝑇 ] × π’ͺ.
We recall that
𝑉¯
admits a (unique) continuous extension to
[0, 𝑇 ] × π’ͺ
under the stated conditions, so the assertion makes sense.
Proof.
The easier direction of the dynamic programming principle for
𝑉
can
be obtained as above, so the result of Lemma 4.9(i) still holds. It then follows
by the same arguments as in the proof of Theorem 4.6 and Proposition 4.11
that the upper semicontinuous envelope
satisfying
βˆ—
𝑉 (𝑇, β‹…) ≀ 𝑓
on
π’ͺ.
𝑉
βˆ—
is a viscosity subsolution of (4.18)
As in the proof of Proposition 4.11, we can
βˆ—
𝑉 ≀ 𝑉¯βˆ— on
[0, 𝑇 ] × π’ͺ by the
apply the comparison principle of Theorem A.1 to deduce that
[0, 𝑇 ] × π’ͺ.
On the other hand, we clearly have
𝑉¯ ≀ 𝑉
on
denitions of these value functions. Therefore, we have
βˆ—
𝑉¯βˆ— ≀ 𝑉 βˆ— ≀ 𝑉 ≀ 𝑉¯βˆ— = 𝑉¯ βˆ— ,
where
π‘‰βˆ—
denotes the lower semicontinuous envelope of
𝑉
and the last equal-
ity is due to Proposition 4.11. It follows that all these functions coincide.
A Comparison for State Constraint Problems
In this appendix we provide, by adapting the usual techniques, a fairly general comparison theorem for state constraint problems which is suitable for
the applications in Proposition 4.11 and Corollary 4.13.
In the following,
β„‹
denotes a continuous mapping from
to ℝ which is nonincreasing in its third variable, π’ͺ
ℝ𝑑 , and 𝜌 > 0 is a xed constant. We consider the
is a given open subset of
equation
πœŒπœ‘ βˆ’ βˆ‚π‘‘ πœ‘ + β„‹(β‹…, π·πœ‘, 𝐷2 πœ‘) = 0
30
ℝ𝑑 × β„π‘‘ × π•„π‘‘
(A.1)
and the following condition on
Assumption E.
β„‹.
𝛼 > 0 such that
)
lim inf β„‹(𝑦, π‘ž, π‘Œ πœ‚ ) βˆ’ β„‹(π‘₯, 𝑝, 𝑋 πœ‚ )
πœ‚β†“0
(
)
(
)
≀ 𝛼 ∣π‘₯ βˆ’ π‘¦βˆ£ 1 + βˆ£π‘žβˆ£ + 𝑛2 ∣π‘₯ βˆ’ π‘¦βˆ£ + (1 + ∣π‘₯∣)βˆ£π‘ βˆ’ π‘žβˆ£ + (1 + ∣π‘₯∣2 )βˆ£π‘„βˆ£
There exists
(
for all
(π‘₯, 𝑦) ∈ π’ͺ with ∣π‘₯ βˆ’ π‘¦βˆ£ ≀ 1 and for all (𝑝, π‘ž, 𝑄) ∈ ℝ𝑑 × β„π‘‘ × π•„2𝑑 ,
βŠ‚ 𝕄𝑑 × π•„π‘‘ and 𝑛 β‰₯ 1 such that
( πœ‚
)
𝑋
0
≀ 𝐴𝑛 + πœ‚π΄2𝑛 for all πœ‚ > 0,
0 βˆ’π‘Œ πœ‚
(𝑋 πœ‚ , π‘Œ πœ‚ )πœ‚>0
where
𝐴𝑛 := 𝑛
2
(
𝐼𝑑 βˆ’πΌπ‘‘
βˆ’πΌπ‘‘ 𝐼𝑑
)
+ 𝑄.
Theorem A.1.
Let Assumption E hold true. Let 𝑀1 be an u.s.c. viscosity
subsolution on π’ͺ and let 𝑀2 be an l.s.c. viscosity supersolution on π’ͺ of (A.1).
If 𝑀1 and 𝑀2 have polynomial growth on π’ͺ and if 𝑀2 is of class β„›(π’ͺ), then
𝑀2 β‰₯ 𝑀1 on {𝑇 } × π’ͺ
Remark A.2.
implies
𝑀2 β‰₯ 𝑀1 on [0, 𝑇 ] × π’ͺ.
Our result also applies to the equation
βˆ’βˆ‚π‘‘ πœ‘ + β„‹(β‹…, π·πœ‘, 𝐷2 πœ‘) = 0,
provided that
(A.2)
β„‹ is homogeneous of degree one with respect to its second and
third argument, as it is the case for the Hamilton-Jacobi-Bellman operators
in the body of this paper. Indeed,
only if
(𝑑, π‘₯) 7β†’
π‘’πœŒπ‘‘ 𝑀
𝑀1
is then a subsolution of (A.2) if and
1 (𝑑, π‘₯) is a subsolution of (A.1), and similarly for the
supersolution. Further extensions could also be considered but are beyond
the scope of this paper.
Proof of Theorem A.1. Assume that 𝑀2 β‰₯ 𝑀1 on {𝑇 } × π’ͺ. Let 𝑝 β‰₯ 1 and
𝐢 > 0 be such that 𝑀1 (𝑑, π‘₯) βˆ’ 𝑀2 (𝑑, π‘₯) ≀ 𝐢(1 + ∣π‘₯βˆ£π‘ ) for all (𝑑, π‘₯) ∈ [0, 𝑇 ] × π’ͺ.
Assume for contradiction that sup(𝑀1 βˆ’ 𝑀2 ) > 0, then we can nd πœ„ > 0 and
(𝑑0 , π‘₯0 ) ∈ [0, 𝑇 ] × π’ͺ such that
𝜁 := (𝑀1 βˆ’ 2πœ™ βˆ’ 𝑀2 )(𝑑0 , π‘₯0 ) = max (𝑀1 βˆ’ 2πœ™ βˆ’ 𝑀2 ) > 0,
(A.3)
[0,𝑇 ]×π’ͺ
where
πœ™(𝑑, π‘₯) := πœ„π‘’βˆ’πœ…π‘‘ (1 + ∣π‘₯∣2𝑝 ).
Here
πœ…>0
is a xed constant which is large enough to ensure that
π‘š(𝑑, π‘₯) :=
(A.4)
2
(
2
)
βˆ’ πœŒπœ™(𝑑, π‘₯) + βˆ‚π‘‘ πœ™(𝑑, π‘₯) + 𝛼 (1 + ∣π‘₯∣)βˆ£π·πœ™(𝑑, π‘₯)∣ + (1 + ∣π‘₯∣ )∣𝐷 πœ™(𝑑, π‘₯)∣
31
[0, 𝑇 ] × β„π‘‘ ,
is nonpositive on
where
𝛼
Note that by the assumption that
(𝑑0 , π‘₯0 ) ∈ [0, 𝑇 ) × π’ͺ.
Case 1: π‘₯0 ∈ βˆ‚π’ͺ.
For all
is the constant from Assumption E.
𝑀2 β‰₯ 𝑀1
on
{𝑇 } × π’ͺ,
we must have
𝑛 β‰₯ 1, there exist (𝑑𝑛 , π‘₯𝑛 , 𝑠𝑛 , 𝑦 𝑛 ) ∈ ([0, 𝑇 ]×π’ͺ)2
satisfying
Φ𝑛 (𝑑𝑛 , π‘₯𝑛 , 𝑠𝑛 , 𝑦 𝑛 ) =
Φ𝑛 ,
max
(A.5)
([0,𝑇 ]×π’ͺ)2
where
Φ𝑛 (𝑑, π‘₯, 𝑠, 𝑦) := 𝑀1 (𝑑, π‘₯) βˆ’ 𝑀2 (𝑠, 𝑦) βˆ’ Ξ˜π‘› (𝑑, π‘₯, 𝑠, 𝑦)
and
)
1 (
Ξ˜π‘› (𝑑, π‘₯, 𝑠, 𝑦) := 𝑛2 βˆ£π‘‘ + πœ†(π‘›βˆ’1 ) βˆ’ π‘ βˆ£2 + πœ€βˆ£π‘₯ + β„“(π‘›βˆ’1 ) βˆ’ π‘¦βˆ£2
2
+ βˆ£π‘‘ βˆ’ 𝑑0 ∣2 + ∣π‘₯ βˆ’ π‘₯0 ∣4 + πœ™(𝑑, π‘₯) + πœ™(𝑠, 𝑦),
β„“
πœ€ > 0.
with
πœ†
and
given for
π‘₯0
as in the statement of the Denition 4.10, and
Note that (A.5), the assumption that
𝑀2
satises (4.22), and (A.3)
imply that
(
)
Φ𝑛 (𝑑𝑛 , π‘₯𝑛 , 𝑠𝑛 , 𝑦 𝑛 ) β‰₯ Φ𝑛 𝑑0 , π‘₯0 , 𝑑0 + πœ†(π‘›βˆ’1 ), π‘₯0 + β„“(π‘›βˆ’1 )
= (𝑀1 βˆ’ 2πœ™ βˆ’ 𝑀2 )(𝑑0 , π‘₯0 ) + π‘œ(1)
= 𝜁 + π‘œ(1)
as
𝑛 β†’ ∞.
(A.6)
𝑀1 , 𝑀2 and the denition of πœ™, it fol𝑛 𝑛 𝑛 𝑛
lows that, after passing to a subsequence, (𝑑 , π‘₯ , 𝑠 , 𝑦 ) converges to some
∞
∞
∞
∞
2
(𝑑 , π‘₯ , 𝑑 , π‘₯ ) ∈ ([0, 𝑇 ] × π’ͺ) . We then have
Recalling the growth condition on
𝜁 = (𝑀1 βˆ’ 2πœ™ βˆ’ 𝑀2 )(𝑑0 , π‘₯0 )
= max (𝑀1 βˆ’ 2πœ™ βˆ’ 𝑀2 )
[0,𝑇 ]×π’ͺ
β‰₯ (𝑀1 βˆ’ 2πœ™ βˆ’ 𝑀2 )(π‘‘βˆž , π‘₯∞ ) βˆ’ βˆ£π‘‘βˆž βˆ’ 𝑑0 ∣2 βˆ’ ∣π‘₯∞ βˆ’ π‘₯0 ∣4
)
1 (
βˆ’ lim sup 𝑛2 βˆ£π‘‘π‘› + πœ†(π‘›βˆ’1 ) βˆ’ 𝑠𝑛 ∣2 + πœ€βˆ£π‘₯𝑛 + β„“(π‘›βˆ’1 ) βˆ’ 𝑦 𝑛 ∣2
π‘›β†’βˆž 2
β‰₯ lim inf Φ𝑛 (𝑑𝑛 , π‘₯𝑛 , 𝑠𝑛 , 𝑦 𝑛 )
π‘›β†’βˆž
β‰₯ 𝜁,
where (A.6) was used in the last step. After passing to a subsequence, we
deduce that
(𝑑𝑛 , π‘₯𝑛 , 𝑠𝑛 , 𝑦 𝑛 ) β†’ (𝑑0 , π‘₯0 , 𝑑0 , π‘₯0 ),
(A.7)
𝑀1 (𝑑𝑛 , π‘₯𝑛 ) βˆ’ 𝑀2 (𝑠𝑛 , 𝑦 𝑛 ) β†’ (𝑀1 βˆ’ 𝑀2 )(𝑑0 , π‘₯0 ),
𝑠𝑛
Since
=
𝑑𝑛
+
πœ†(π‘›βˆ’1 )
+
π‘œ(π‘›βˆ’1 ),
𝑦𝑛
=
π‘₯𝑛
(𝑑0 , π‘₯0 ) ∈ [0, 𝑇 ) × βˆ‚π’ͺ, it follows from
∈ [0, 𝑇 ) × π’ͺ for 𝑛 large enough.
(𝑠𝑛 , 𝑦 𝑛 )
32
+
β„“(π‘›βˆ’1 )
+
π‘œ(π‘›βˆ’1 ).
(A.8)
(A.9)
(4.20), (4.21) and (A.9) that
Let
2,+
𝒫 π’ͺ 𝑀1
2,βˆ’
𝒫 π’ͺ 𝑀2
and
be the closed parabolic super- and subjets as
dened in [5, Section 8]. From the Crandall-Ishii lemma [5, Theorem 8.3] we
obtain, for each
πœ‚ > 0,
elements
2,+
(π‘Žπ‘› , 𝑝𝑛 , π‘‹πœ‚π‘› ) ∈ 𝒫 π’ͺ 𝑀1 (𝑑𝑛 , π‘₯𝑛 )
and
2,βˆ’
(𝑏𝑛 , π‘ž 𝑛 , π‘Œπœ‚π‘› ) ∈ 𝒫 π’ͺ 𝑀2 (𝑠𝑛 , 𝑦 𝑛 )
such that
π‘Žπ‘› = βˆ‚π‘‘ Ξ˜π‘› (𝑑𝑛 , π‘₯𝑛 , 𝑠𝑛 , 𝑦 𝑛 ),
𝑏𝑛 = βˆ’βˆ‚π‘  Ξ˜π‘› (𝑑𝑛 , π‘₯𝑛 , 𝑠𝑛 , 𝑦 𝑛 ),
𝑝𝑛 = 𝐷π‘₯ Ξ˜π‘› (𝑑𝑛 , π‘₯𝑛 , 𝑠𝑛 , 𝑦 𝑛 ),
π‘ž 𝑛 = βˆ’π·π‘¦ Ξ˜π‘› (𝑑𝑛 , π‘₯𝑛 , 𝑠𝑛 , 𝑦 𝑛 ),
(
π‘‹πœ‚π‘›
0
0 βˆ’π‘Œπœ‚π‘›
)
≀ 𝐴𝑛 + πœ‚π΄2𝑛 ,
𝐴𝑛 = 𝐷2 Ξ˜π‘› (𝑑𝑛 , π‘₯𝑛 , 𝑠𝑛 , 𝑦 𝑛 ); i.e.,
)
(
) ( 2 𝑛 𝑛
𝐼𝑑 βˆ’πΌπ‘‘
𝐷 πœ™(𝑑 , π‘₯ ) + 𝑂(∣π‘₯𝑛 βˆ’ π‘₯0 ∣2 )
0
.
𝐴𝑛 = πœ€π‘›2
+
βˆ’πΌπ‘‘ 𝐼𝑑
0
𝐷2 πœ™(𝑠𝑛 , 𝑦 𝑛 )
where
In view of the sub- and supersolution properties of
(𝑠𝑛 , 𝑦 𝑛 ) ∈ [0, 𝑇 ) × π’ͺ
for
𝑛
𝑀1
and
𝑀2 ,
the fact that
large, and Assumption E, we deduce that
Δ𝑛 := 𝜌(𝑀1 (𝑑𝑛 , π‘₯𝑛 ) βˆ’ 𝑀2 (𝑠𝑛 , 𝑦 𝑛 ))
≀ 2(𝑑𝑛 βˆ’ 𝑑0 ) + βˆ‚π‘‘ πœ™(𝑑𝑛 , π‘₯𝑛 ) + βˆ‚π‘  πœ™(𝑠𝑛 , 𝑦 𝑛 )
(
)
+ π›Όβˆ£π‘₯𝑛 βˆ’ 𝑦 𝑛 ∣ 1 + βˆ£π‘ž 𝑛 ∣ + πœ€π‘›2 ∣π‘₯𝑛 βˆ’ 𝑦 𝑛 ∣
(
)
+ 𝛼 (1 + ∣π‘₯𝑛 ∣)βˆ£π‘ž 𝑛 βˆ’ 𝑝𝑛 ∣ + (1 + ∣π‘₯𝑛 ∣2 )βˆ£π‘„π‘› ∣ ,
where
𝑛
𝑄 :=
(
𝐷2 πœ™(𝑑𝑛 , π‘₯𝑛 ) + 𝑂(∣π‘₯𝑛 βˆ’ π‘₯0 ∣2 )
0
0
𝐷2 πœ™(𝑠𝑛 , 𝑦 𝑛 )
By the denitions of
𝑝𝑛
and
π‘žπ‘›,
)
.
it follows that
Δ𝑛 ≀ 2(𝑑𝑛 βˆ’ 𝑑0 ) + βˆ‚π‘‘ πœ™(𝑑𝑛 , π‘₯𝑛 ) + βˆ‚π‘  πœ™(𝑠𝑛 , 𝑦 𝑛 )
(
)
+ π›Όβˆ£π‘₯𝑛 βˆ’ 𝑦 𝑛 ∣ 1 + βˆ£π·πœ™(𝑠𝑛 , 𝑦 𝑛 )∣ + πœ€π‘›2 ∣π‘₯𝑛 + β„“(π‘›βˆ’1 ) βˆ’ 𝑦 𝑛 ∣ + πœ€π‘›2 ∣π‘₯𝑛 βˆ’ 𝑦 𝑛 ∣
(
)
+ 𝛼(1 + ∣π‘₯𝑛 ∣) 4∣π‘₯𝑛 βˆ’ π‘₯0 ∣3 + βˆ£π·πœ™(𝑑𝑛 , π‘₯𝑛 )∣ + βˆ£π·πœ™(𝑠𝑛 , 𝑦 𝑛 )∣
(
)
+ 𝛼(1 + ∣π‘₯𝑛 ∣2 ) ∣𝐷2 πœ™(𝑑𝑛 , π‘₯𝑛 )∣ + ∣𝐷2 πœ™(𝑠𝑛 , 𝑦 𝑛 )∣ + 𝑂(∣π‘₯𝑛 βˆ’ π‘₯0 ∣2 ) .
Recalling (A.7)(A.9), letting
π‘›β†’βˆž
leads to
(
)2
𝜌(𝑀1 βˆ’ 𝑀2 )(𝑑0 , π‘₯0 ) ≀ 2βˆ‚π‘‘ πœ™(𝑑0 , π‘₯0 ) + π›Όπœ€ lim inf 𝑛ℓ(π‘›βˆ’1 )
π‘›β†’βˆž
(
)
+ 2𝛼 (1 + ∣π‘₯0 ∣)βˆ£π·πœ™(𝑑0 , π‘₯0 )∣ + (1 + ∣π‘₯0 ∣2 )∣𝐷2 πœ™(𝑑0 , π‘₯0 )∣ ,
33
which, by (4.19) and the denition of
π‘š
in (A.4), implies
𝜌(𝑀1 βˆ’ 2πœ™ βˆ’ 𝑀2 )(𝑑0 , π‘₯0 ) ≀ 2π‘š(𝑑0 , π‘₯0 )
after letting
πœ€ β†’ 0.
Since
πœ… > 0 has been chosen so that π‘š ≀ 0 on [0, 𝑇 ]×ℝ𝑑 ,
this contradicts (A.3).
Case 2: π‘₯0 ∈ π’ͺ.
This case is handled similarly by using
)
1 (
Ξ˜π‘› (𝑑, π‘₯, 𝑠, 𝑦) := 𝑛2 βˆ£π‘‘ βˆ’ π‘ βˆ£2 + ∣π‘₯ βˆ’ π‘¦βˆ£2 + βˆ£π‘‘ βˆ’ 𝑑0 ∣2 + ∣π‘₯ βˆ’ π‘₯0 ∣4 + πœ™(𝑑, π‘₯) + πœ™(𝑠, 𝑦).
2
After taking a subsequence, the corresponding sequence of maximum points
(𝑑𝑛 , π‘₯𝑛 , 𝑠𝑛 , 𝑦 𝑛 )𝑛β‰₯1 again converges
𝑛 large enough. The rest of the
to
(𝑑0 , π‘₯0 , 𝑑0 , π‘₯0 ),
so that
π‘₯𝑛 , 𝑦 𝑛 ∈ π’ͺ
for
proof follows the same arguments as in
Case 1.
Lemma A.3.
¯ dened in
Under Assumption D', the operator 𝐻
ises Assumption E.
Proof.
(4.17)
sat-
The argument is standard. We rst observe that
(
)
¯ 𝑒 (𝑦, π‘ž, π‘Œ πœ‚ ) + 𝐿
¯ 𝑒 (π‘₯, 𝑝, 𝑋 πœ‚ ) = πœ‡(π‘₯, 𝑒) βˆ’ πœ‡(𝑦, 𝑒) ⊀ π‘ž + πœ‡(π‘₯, 𝑒)⊀ (𝑝 βˆ’ π‘ž)
βˆ’πΏ
1 βˆ‘ 𝑖
+
Ξ£ (π‘₯, 𝑦, 𝑒)⊀ Ξžπœ‚ Σ𝑖 (π‘₯, 𝑦, 𝑒),
2
1≀𝑖≀𝑑
where
(
Ξ£(π‘₯, 𝑦, 𝑒) :=
and
Σ𝑖
𝜎(π‘₯, 𝑒)
𝜎(𝑦, 𝑒)
denotes the 𝑖th column of
)
and
Ξ£.
Since
πœ‚
(
Ξ :=
π‘‹πœ‚
0
0 βˆ’π‘Œ πœ‚
)
πœ‡ is Lipschitz continuous and has
uniformly linear growth by Assumption D', we have
(
)⊀
(
)
πœ‡(π‘₯, 𝑒) βˆ’ πœ‡(𝑦, 𝑒) π‘ž + πœ‡(π‘₯, 𝑒)⊀ (𝑝 βˆ’ π‘ž) ≀ 𝛼 ∣π‘₯ βˆ’ π‘¦βˆ£βˆ£π‘žβˆ£ + (1 + ∣π‘₯∣)βˆ£π‘ βˆ’ π‘žβˆ£
for some constant
𝛼 > 0.
Recall from the statement of Assumption E the
condition that
πœ‚
Ξ ≀ 𝐴𝑛 +
Since
𝜎
πœ‚π΄2𝑛 ,
where
𝐴𝑛 := 𝑛
2
(
𝐼𝑑 βˆ’πΌπ‘‘
βˆ’πΌπ‘‘ 𝐼𝑑
)
+ 𝑄.
is Lipschitz continuous and satises Assumption D', it follows that
Σ𝑖 (π‘₯, 𝑦, 𝑒)⊀ Ξžπœ‚ Σ𝑖 (π‘₯, 𝑦, 𝑒) ≀ 𝑛2 ∣𝜎 𝑖 (π‘₯, 𝑒) βˆ’ 𝜎 𝑖 (𝑦, 𝑒)∣2
(
)
+ 𝛼 1 + ∣π‘₯∣2 + βˆ£π‘¦βˆ£2 βˆ£π‘„βˆ£ + 𝑂(πœ‚),
𝛼, where 𝜎 𝑖 denotes the 𝑖th column of 𝜎 . We conclude
by using the Lipschitz continuity of 𝜎 and the condition that ∣π‘₯ βˆ’ π‘¦βˆ£ ≀ 1.
possibly after enlarging
34
Remark A.4.
We conclude with some comments on Theorem A.1 and
Proposition 4.12, and related results in the literature.
The main issue in proving comparison with state constraints is to avoid
boundary points; i.e., that
𝑦𝑛
(see the proof of Theorem A.1) ends up on
the boundary. One classical way to ensure this, is to use a perturbation of
∣π‘₯ βˆ’ π‘¦βˆ£2
in a suitable inward direction, like the function
β„“
above. Moreover,
this requires the supersolution to be continuous at the boundary points,
along the direction of perturbation; cf. (4.22).
In [11], the inner normal
𝑛(π‘₯)
π‘₯ ∈ βˆ‚π’ͺ is used as
π‘₯ (cf. (A.9)); therefore,
at the boundary point
an inward direction. In the proof, one is only close to
the comparison result [11, Theorem 2.2] requires the existence of a truncated
cone around
𝑛(π‘₯)
which stays inside the domain, and the continuity of the
subsolution along the directions that it generates. Our condition (4.20) is
less restrictive than the corresponding requirement in [11]:
we only need
πœ€ β†’
7
β„“(πœ€), cf. (4.22), rather than all lines
function πœ† appears because we consider parabolic
the continuity along the curve
in a neighborhood. The
equations, whereas [11] focuses on the elliptic case.
In Proposition 4.12 we give conditions (certainly not the most general
possible) ensuring that the value function is of class
β„›(π’ͺ).
They should
be compared to [11, Condition (A3)], which is used to verify the continuity
assumption of [11, Theorem 2.2]. Our conditions are stronger in the sense
that they are imposed around the boundary and not only at the boundary;
on the other hand, we require
𝐢 1 -regularity
of the boundary whereas [11]
3
requires 𝐢 .
In [10], a slightly dierent technique is used, based on ideas from [9].
First, it is assumed that at each boundary point
π‘₯, there exists a xed control
which kills the volatility at the neighboring boundary points and keeps a
truncated cone around the drift in the domain. This is similar to our (4.23),
except that our control is not xed; on the other hand, we assume that it
π‘₯. Thus, the conditions in [10] are
2
not directly comparable to ours; e.g., if π’ͺ is the unit disk in ℝ , π‘ˆ = [βˆ’1, 1],
2
πœ‡(π‘₯, 𝑒) = βˆ’π‘₯ and 𝜎(π‘₯, 𝑒) = ∣π‘₯ βˆ’ π‘’βˆ£πΌ2 , then (4.23) is satised (with 𝛿 given
by the Euclidean distance near the boundary and 𝑒
Λ‡(π‘₯) = π‘₯2 ), while [10,
kills the volatility in a neighborhood of
Condition (2.1)] is not.
Second, in [10], the state constraint problem is
transformed so as to introduce a Neumann-type boundary condition and
construct a suitable test function which, as a function of its rst component,
turns out to be a uniformly strict supersolution of the Neumann boundary
condition. The construction in [10] heavily relies on the assumption that the
coecients are bounded; cf. the beginning of [10, Section 3] and the proof
of [10, Theorem 3.1].
35
References
[1] B. Bouchard, R. Elie, and C. Imbert. Optimal control under stochastic target
constraints. SIAM J. Control Optim., 48(5):35013531, 2010.
[2] B. Bouchard, R. Elie, and N. Touzi. Stochastic target problems with controlled
loss. SIAM J. Control Optim., 48(5):31233150, 2009.
[3] B. Bouchard and N. Touzi. Weak dynamic programming principle for viscosity
solutions. SIAM J. Control Optim., 49(3):948962, 2011.
[4] B. Bouchard and T.N. Vu. A stochastic target approach for P&L matching
problems. Mathematics of Operation Research, 37(3):526558, 2012.
[5] M. Crandall, H. Ishii, and P.-L. Lions. User's guide to viscosity solutions of
second order partial dierential equations. Bull. Amer. Math. Soc., 27(1):167,
1992.
[6] W. H. Fleming and H. M. Soner. Controlled Markov Processes and Viscosity
Solutions. Springer, New York, 2nd edition, 2006.
[7] M. I. Freidlin. Functional Integration and Partial Dierential Equations.
Princeton University Press, 1985.
[8] H. Föllmer and P. Leukert. Ecient hedging: cost versus shortfall risk. Finance and Stochastics, 4(2):117146, 2000.
[9] H. Ishii and S. Koike. A new formulation of state constraint problems for
rst-order PDEs. SIAM J. Control Optim., 34(2):554571, 1996.
[10] H. Ishii and P. Loreti. A class of stochastic optimal control problems with
state constraint. Indiana Univ. Math. J., 51(5):11671196, 2002.
[11] M. A. Katsoulakis. Viscosity solutions of second order fully nonlinear elliptic
equations with state constraints. Indiana Univ. Math. J., 43(2):493519, 1994.
[12] J.-M. Lasry and P.-L. Lions. Nonlinear elliptic equations with singular boundary conditions and stochastic control with state constraints. I. The model
problem. Math. Ann., 283(4):583630, 1989.
[13] H. M. Soner. Optimal control with state-space constraint. I. SIAM J. Control
Optim., 24(3):552561, 1986.
[14] H. M. Soner. Optimal control with state-space constraint. II. SIAM J. Control
Optim., 24(6):11101122, 1986.
[15] H. M. Soner and N. Touzi. Dynamic programming for stochastic target problems and geometric ows. J. Eur. Math. Soc. (JEMS), 4(3):201236, 2002.
[16] P. Soravia. Optimality principles and representation formulas for viscosity
solutions of Hamilton-Jacobi equations. I. Equations of unbounded and degenerate control problems without uniqueness. Adv. Dierential Equations,
4(2):275296, 1999.
[17] P. Soravia. Optimality principles and representation formulas for viscosity
solutions of Hamilton-Jacobi equations. II. Equations of control problems with
state constraints. Dierential Integral Equations, 12(2):275293, 1999.
[18] J. Yong and X. Y. Zhou. Stochastic Controls. Hamiltonian Systems and HJB
Equations. Springer, New York, 1999.
36