PDF

A Quasi-Sure Approach to the Control of
Non-Markovian Stochastic Dierential Equations
Marcel Nutz
βˆ—
First version: June 16, 2011. This version: March 18, 2012.
Abstract
We study stochastic dierential equations (SDEs) whose drift and
diusion coecients are path-dependent and controlled. We construct
a value process on the canonical path space, considered simultaneously
under a family of singular measures, rather than the usual family of
processes indexed by the controls. This value process is characterized by a second order backward SDE, which can be seen as a nonMarkovian analogue of the Hamilton-Jacobi-Bellman partial dierential equation. Moreover, our value process yields a generalization of
the 𝐺-expectation to the context of SDEs.
Keywords Stochastic optimal control, non-Markovian SDE, second order BSDE,
𝐺-expectation, random 𝐺-expectation, volatility uncertainty, risk measure
AMS 2000 Subject Classications 93E20, 49L20, 60H10, 60G44, 91B30
Acknowledgements
Financial support by Swiss National Science Foundation
Grant PDFM2-120424/1 is gratefully acknowledged.
The author thanks
Shige Peng, Mete Soner, Nizar Touzi and Jianfeng Zhang for stimulating
discussions and two anonymous referees for helpful comments.
1
Introduction
We consider a controlled stochastic dierential equation (SDE) of the form
∫
𝑋𝑑 = π‘₯ +
𝑑
∫
πœ‡(π‘Ÿ, 𝑋, πœˆπ‘Ÿ ) π‘‘π‘Ÿ +
0
where
𝜈
𝜎(π‘Ÿ, 𝑋, πœˆπ‘Ÿ ) π‘‘π‘Šπ‘Ÿ ,
0 ≀ 𝑑 ≀ 𝑇,
(1.1)
0
π‘Š
πœ‡(π‘Ÿ, 𝑋, πœˆπ‘Ÿ )
is an adapted control process,
Lipschitz-continuous coecients
βˆ—
𝑑
is a Brownian motion and the
and
𝜎(π‘Ÿ, 𝑋, πœˆπ‘Ÿ )
Dept. of Mathematics, Columbia University, New York,
1
may depend on
[email protected]
{𝑋𝑠 , 0 ≀ 𝑠 ≀ π‘Ÿ} of the solution. Denoting by 𝑋 𝜈 the solution corresponding to 𝜈 , we are interested in the stochastic optimal control
the past trajectory
problem
𝑉0 := sup 𝐸[πœ‰(𝑋 𝜈 )],
(1.2)
𝜈
where
πœ‰ is a given functional.
The standard approach to such a non-Markovian
control problem (cf. [9, 10]) is to consider for each control
𝜈
the associated
value process
π½π‘‘πœˆ =
where
of a
ess sup
𝜈˜: 𝜈˜=𝜈 on [0,𝑑]
(ℱ𝑑 ) is the given ltration.
forward
𝐸[πœ‰(𝑋 𝜈˜ )βˆ£β„±π‘‘ ],
The dependence on
(1.3)
𝜈
reects the presence
component in the optimization problem.
The situation is quite dierent in Markovian optimal control (cf. [12]),
where one uses a single value function which depends on certain state variables but not on a control. This is essential to describe the value function by
a dierential equation, such as the Hamilton-Jacobi-Bellman PDE, which is
the main merit of the dynamic programming approach. It is worth noting
that this equation is always
backward
in time.
An analogous description
for (1.3) via backward SDEs (BSDEs, cf. [20]) is available for certain popular problems such as utility maximization with power or exponential utility
functions (e.g., [13, 18]) or drift control (e.g., [10]). However, this relies on a
𝐽𝜈
on 𝜈 .
very particular algebraic structure which allows for a separation of
a backward part independent of
𝜈
and a forward part depending
into
In this paper, we consider the problem (1.2) on the canonical space by
recasting it as
𝜈
𝑉0 = sup 𝐸 𝑃 [πœ‰(𝐡)],
(1.4)
𝜈
where
π‘ƒπœˆ
is the distribution of
π‘‹πœˆ
and
𝐡
is the canonical process, and we
describe its dynamic value by a single value process
𝑉
𝑉 = {𝑉𝑑 (πœ”)}.
Formally,
corresponds to a value function in the Markovian sense if we see the whole
trajectory of the controlled system as a state variable.
Even though (1.2)
has features of coupled forward-backward type, the value process is dened
in a purely backward manner: one may say that by constructing
𝑉
on the
whole canonical space, we essentially calculate the value for all possible outcomes of the forward part. An important ingredient in the same vein is that
𝑉 is dened quasi-surely under the family of mutually singular measures
{𝑃 𝜈 }. Rather than forming a family of processes as in (1.3), the necessary
information is stored in a single process which is dened on a large part
of the probability space; indeed, the process
𝜈
thought of as an analogue of 𝐽 .
𝑉
seen under
𝑃 𝜈
should be
Clearly, this is a necessary step to ob-
tain a (second order) backward SDE. We remark that [22] considered the
same control problem (1.2) and also made a connection to nonlinear expectations. However, in [22], the value process was considered only under the
given probability measure.
2
We rst consider a fairly regular functional
πœ‰
and dene
𝑉𝑑 (πœ”)
as a con-
ditional version of (1.4). Applying and advancing ideas from [28] and [17],
regular conditional probability distributions are used to dene
ery πœ”
𝑉𝑑 (πœ”)
for
ev-
and prove a pathwise dynamic programming principle (Theorem 3.2).
In a second step, we enlarge the class of functionals
πœ‰
to an
𝐿1 -type
space
and prove that the value process admits a (quasi-sure) càdlàg modication
(Theorem 5.1).
We also show that the value process falls into the class of sublinear ex-
πœ‰ is considered as a random variable on
πœ‰ 7β†’ 𝑉𝑑 can be seen as a generalization of
which, by [7], corresponds to the case πœ‡ ≑ 0 and
pectations studied in [19]. Indeed, if
the canonical space, the mapping
𝐺-expectation [23, 24],
𝜎(π‘Ÿ, 𝑋, πœˆπ‘Ÿ ) = πœˆπ‘Ÿ , where the SDE (1.1) degenerates to a stochastic integral.
Moreover, 𝑉𝑑 can be seen as a variant of the random 𝐺-expectation [17]; cf.
the
Remark 6.5.
Finally, we characterize
𝑉
by a second order backward SDE (2BSDE) in
the spirit of [19]; cf. Theorem 6.4. The second order is clearly necessary since
the Hamilton-Jacobi-Bellman PDE for the Markovian case is fully nonlinear,
while ordinary BSDEs correspond to semilinear equations.
2BSDEs were
introduced in [5], and in [27] for the non-Markovian case. We refer to [27]
for the precise relation between 2BSDEs in the quasi-sure formulation and
fully nonlinear parabolic PDEs.
We remark that our approach is quite dierent from the (backward)
stochastic
partial
dierential equations studied in [21] for a similar con-
trol problem (mainly for uncontrolled volatility) and in [14, 15, 2, 3, 4] for
so-called pathwise stochastic control problems.
The relation to the path-
dependent PDEs, introduced very recently in [25], is yet to be explored.
The remainder of the paper is organized as follows. In Section 2 we detail
the controlled SDE and its conditional versions, dene the value process for
the case when
πœ‰
is uniformly continuous and establish its regularity.
The
pathwise dynamic programming principle is proved in Section 3. In Section 4
we extend the value process to a more general class of functionals
πœ‰
and
state its quasi-sure representation. The càdlàg modication is constructed
in Section 5. In the concluding Section 6, we provide the Hamilton-JacobiBellman 2BSDE and interpret the value process as a variant of the random
𝐺-expectation.
2
Construction of the Value Function
In this section, we rst introduce the setting and notation. Then, we dene
the value function
𝑉𝑑 (πœ”)
for a uniformly continuous reward functional
examine the regularity of
𝑉𝑑 .
3
πœ‰
and
2.1
Notation
We x a constant
𝑇 > 0 and let Ξ© := 𝐢([0, 𝑇 ];
R𝑑 )
be the canonical space of
βˆ₯πœ”βˆ₯𝑇 := sup0≀𝑠≀𝑇 βˆ£πœ”π‘  ∣,
where ∣ β‹… ∣ is the Euclidean norm. We denote by 𝐡 the canonical process
𝐡𝑑 (πœ”) = πœ”π‘‘ , by 𝑃0 the Wiener measure, and by 𝔽 = {ℱ𝑑 }0≀𝑑≀𝑇 the (raw)
ltration generated by 𝐡 . Unless otherwise stated, probabilistic notions
requiring a ltration (such as adaptedness) refer to 𝔽.
For any probability measure 𝑃 on Ξ© and any (𝑑, πœ”) ∈ [0, 𝑇 ] × Ξ©, we can
πœ”
construct the corresponding regular conditional probability distribution 𝑃𝑑 ;
πœ”
cf. [29, Theorem 1.3.4]. We recall that 𝑃𝑑 is a probability kernel on ℱ𝑑 × β„±π‘‡ ;
πœ”
πœ”
i.e., 𝑃𝑑 is a probability measure on (Ξ©, ℱ𝑇 ) for xed πœ” and πœ” 7β†’ 𝑃𝑑 (𝐴) is
πœ”
ℱ𝑑 -measurable for each 𝐴 ∈ ℱ𝑇 . Moreover, the expectation under 𝑃𝑑 is the
conditional expectation under 𝑃 :
continuous paths equipped with the uniform norm
πœ”
𝐸 𝑃𝑑 [πœ‰] = 𝐸 𝑃 [πœ‰βˆ£β„±π‘‘ ](πœ”) 𝑃 -a.s.
πœ‰
whenever
is
ℱ𝑇 -measurable
and bounded. Finally,
the set of paths that coincide with
πœ”
up to
𝑃𝑑 πœ” β€² ∈ Ξ© : πœ” β€² = πœ”
{
πœ”
While
π‘ƒπ‘‘πœ”
π‘ƒπ‘‘πœ”
is concentrated on
𝑑,
on
}
[0, 𝑑] = 1.
(2.1)
is not dened uniquely by these properties, we choose and x one
version for each triplet
(𝑑, πœ”, 𝑃 ).
Ω𝑑 := {πœ” ∈ 𝐢([𝑑, 𝑇 ]; ℝ𝑑 ) : πœ”π‘‘ = 0} the
shifted canonical space of paths starting at the origin. For πœ” ∈ Ξ©, the
𝑑
𝑑
𝑑
shifted path πœ” ∈ Ξ© is dened by πœ”π‘Ÿ := πœ”π‘Ÿ βˆ’ πœ”π‘‘ for 𝑑 ≀ π‘Ÿ ≀ 𝑇 , so that
𝑑
𝑑
Ξ© = {πœ” : πœ” ∈ Ξ©}. Moreover, we denote by 𝑃0𝑑 the Wiener measure on Ω𝑑
𝑑
𝑑
𝑑
and by 𝔽 = {β„±π‘Ÿ }π‘‘β‰€π‘Ÿβ‰€π‘‡ the (raw) ltration generated by 𝐡 , which can be
𝑑
identied with the canonical process on Ξ© .
Given two paths πœ” and πœ”
˜ , their concatenation at 𝑑 is the (continuous)
Let
𝑑 ∈ [0, 𝑇 ].
We denote by
path dened by
(πœ” βŠ—π‘‘ πœ”
˜ )π‘Ÿ := πœ”π‘Ÿ 1[0,𝑑) (π‘Ÿ) + (πœ”π‘‘ + πœ”
˜ π‘Ÿπ‘‘ )1[𝑑,𝑇 ] (π‘Ÿ),
Given an
ℱ𝑇 -measurable
random variable
conditioned random variable
πœ‰ 𝑑,πœ”
on
Ξ©
πœ‰
on
Ξ©
and
0 ≀ π‘Ÿ ≀ 𝑇.
πœ” ∈ Ξ©,
we dene the
by
πœ‰ 𝑑,πœ” (˜
πœ” ) := πœ‰(πœ” βŠ—π‘‘ πœ”
˜ ),
πœ”
˜ ∈ Ω.
πœ‰ 𝑑,πœ” (˜
πœ” ) = πœ‰ 𝑑,πœ” (˜
πœ” 𝑑 ); in particular, πœ‰ 𝑑,πœ” can also be seen as a ran𝑑
dom variable on Ξ© . Then πœ”
˜ 7β†’ πœ‰ 𝑑,πœ” (˜
πœ” ) is ℱ𝑇𝑑 -measurable and moreover,
𝑑,πœ”
πœ‰
depends only on the restriction of πœ” to [0, 𝑑]. We note that for an 𝔽progressively measurable process {π‘‹π‘Ÿ , π‘Ÿ ∈ [𝑠, 𝑇 ]}, the conditioned process
{π‘‹π‘Ÿπ‘‘,πœ” , π‘Ÿ ∈ [𝑑, 𝑇 ]} is 𝔽𝑑 -progressively measurable. If 𝑃 is a probability on Ξ©,
𝑑,πœ” on β„± 𝑑 dened by
the measure 𝑃
𝑇
Note that
𝑃 𝑑,πœ” (𝐴) := π‘ƒπ‘‘πœ” (πœ” βŠ—π‘‘ 𝐴),
𝐴 ∈ ℱ𝑇𝑑 ,
where
4
πœ” βŠ—π‘‘ 𝐴 := {πœ” βŠ—π‘‘ πœ”
˜: πœ”
˜ ∈ 𝐴},
is again a probability by (2.1). We then have
𝐸𝑃
𝑑,πœ”
πœ”
[πœ‰ 𝑑,πœ” ] = 𝐸 𝑃𝑑 [πœ‰] = 𝐸 𝑃 [πœ‰βˆ£β„±π‘‘ ](πœ”) 𝑃 -a.s.
Analogous notation will be used when
πœ”βˆˆ
Ω𝑠 , where
πœ‰
𝑇 . We denote by Ω𝑠𝑑
0≀𝑠≀𝑑≀
Ω𝑠 to [𝑠, 𝑑], equipped with βˆ₯πœ”βˆ₯[𝑠,𝑑]
𝑠
identied with {πœ” ∈ Ξ© : πœ”π‘Ÿ = πœ”π‘‘ for
restriction of
Ω𝑠𝑑 can be
Ω𝑑 is analogous.
2.2
Let
Ω𝑠 and
:= {πœ”βˆ£[𝑠,𝑑] : πœ” ∈ Ω𝑠 } the
:= supπ‘Ÿβˆˆ[𝑠,𝑑] βˆ£πœ”π‘Ÿ ∣. Note that
π‘Ÿ ∈ [𝑑, 𝑇 ]}. The meaning of
is a random variable on
The Controlled SDE
π‘ˆ
be a nonempty Borel subset of
β„π‘š
for some
π‘š ∈ β„•.
We consider two
given functions
πœ‡ : [0, 𝑇 ] × Ξ© × π‘ˆ β†’ ℝ𝑑
and
𝜎 : [0, 𝑇 ] × Ξ© × π‘ˆ β†’ ℝ𝑑×𝑑 ,
the drift and diusion coecients, such that
(𝑑, πœ”) 7β†’ πœ‡(𝑑, 𝑋(πœ”), πœˆπ‘‘ (πœ”))
and
(𝑑, πœ”) 7β†’ 𝜎(𝑑, 𝑋(πœ”), πœˆπ‘‘ (πœ”)) are progressively measurable for any continuous
𝑋 and any π‘ˆ -valued progressively measurable process 𝜈 .
In particular, πœ‡(𝑑, πœ”, 𝑒) and 𝜎(𝑑, πœ”, 𝑒) depend only on the past trajectory
{πœ”π‘Ÿ , π‘Ÿ ∈ [0, 𝑑]}, for any 𝑒 ∈ π‘ˆ . Moreover, we assume that there exists a
constant 𝐾 > 0 such that
adapted process
βˆ£πœ‡(𝑑, πœ”, 𝑒) βˆ’ πœ‡(𝑑, πœ” β€² , 𝑒)∣ + ∣𝜎(𝑑, πœ”, 𝑒) βˆ’ 𝜎(𝑑, πœ” β€² , 𝑒)∣ ≀ 𝐾βˆ₯πœ” βˆ’ πœ” β€² βˆ₯𝑑
(2.2)
(𝑑, πœ”, πœ” β€² , 𝑒) ∈ [0, 𝑇 ]×Ξ©×Ξ©×π‘ˆ . We denote by 𝒰 the set of all π‘ˆ -valued
progressively measurable processes 𝜈 such that
∫ 𝑇
∫ 𝑇
βˆ£πœ‡(π‘Ÿ, 𝑋, πœˆπ‘Ÿ )∣ π‘‘π‘Ÿ < ∞ and
∣𝜎(π‘Ÿ, 𝑋, πœˆπ‘Ÿ )∣2 π‘‘π‘Ÿ < ∞
(2.3)
for all
0
0
hold path-by-path for any continuous adapted process
𝑋.
Given
𝜈 ∈ 𝒰,
the
stochastic dierential equation
𝑑
∫
𝑋𝑑 = π‘₯ +
πœ‡(π‘Ÿ, 𝑋, πœˆπ‘Ÿ ) π‘‘π‘Ÿ +
0
has a
𝑃0 -a.s.
we denote by
𝑑
∫
𝜎(π‘Ÿ, 𝑋, πœˆπ‘Ÿ ) π‘‘π΅π‘Ÿ ,
0≀𝑑≀𝑇
unique strong solution for any initial condition
𝑋(0, π‘₯, 𝜈).
π‘₯ ∈ ℝ𝑑 ,
𝑃0
which
We shall denote by
𝑃¯ (0, π‘₯, 𝜈) := 𝑃0 ∘ 𝑋(0, π‘₯, 𝜈)βˆ’1
the distribution of
under
0
𝑋(0, π‘₯, 𝜈)
on
Ξ©
(2.4)
and by
(
)βˆ’1
𝑃 (0, π‘₯, 𝜈) := 𝑃0 ∘ 𝑋(0, π‘₯, 𝜈)0
𝑋(0, π‘₯, 𝜈)0 ≑ 𝑋(0, π‘₯, 𝜈) βˆ’ π‘₯; i.e., the solution which is
0
translated to start at the origin. Note that 𝑃 (0, π‘₯, 𝜈) is concentrated on Ξ©
0
and can therefore be seen as a probability measure on Ξ© .
the distribution of
We shall work under the following nondegeneracy condition.
5
Assumption 2.1.
Throughout this paper, we assume that
𝔽𝑋
where
(π‘₯, 𝜈)
𝔽𝑋
𝑃0
𝑃0
βŠ‡π”½
𝑃0 -augmentation
ℝ𝑑 × π’° .
is the
varies over
𝑋 = 𝑋(0, π‘₯, 𝜈),
for all
(2.5)
of the ltration generated by
𝑋
and
One can construct situations where Assumption 2.1 fails. For example,
π‘₯ = 0, 𝜎 ≑ 1 and πœ‡(π‘Ÿ, 𝑋, πœˆπ‘Ÿ ) = πœˆπ‘Ÿ , then (2.5) fails for a suitable choice
of 𝜈 ; see, e.g., [11]. The following is a positive result which covers many
if
applications.
Remark 2.2.
Let
𝜎
be strictly positive denite and assume that
is a progressively measurable functional of
𝑋
𝜎(π‘Ÿ, 𝑋, πœˆπ‘Ÿ ).
and
πœ‡(π‘Ÿ, 𝑋, πœˆπ‘Ÿ )
Then (2.5)
holds true.
Note that the latter assumption is satised in particular when
controlled; i.e.,
πœ‡
is un-
πœ‡(π‘Ÿ, πœ”, 𝑒) = πœ‡(π‘Ÿ, πœ”).
Proof.
∫
Let 𝑋 = 𝑋(0, π‘₯, 𝜈). As
𝜎𝜎 ⊀ (π‘Ÿ, 𝑋, πœˆπ‘Ÿ ) π‘‘π‘Ÿ is adapted to
the quadratic variation of
the ltration generated by
𝑋 , the
𝑋 . In
process
view of
our assumptions, it follows that
∫
𝑀 :=
∫
𝜎(π‘Ÿ, 𝑋, πœˆπ‘Ÿ ) π‘‘π΅π‘Ÿ = 𝑋 βˆ’ π‘₯ βˆ’
has the same property. Hence
the ltration generated by
Remark 2.3.
𝐡=
∫
πœ‡(π‘Ÿ, 𝑋, πœˆπ‘‘ ) π‘‘π‘Ÿ
𝜎(π‘Ÿ, 𝑋, πœˆπ‘Ÿ )βˆ’1 π‘‘π‘€π‘Ÿ
is again adapted to
𝑋.
For some applications, in particular when the SDE is of ge-
ometric form, requiring (2.5) to hold for all
instead x the initial condition
require (2.5) only for that
π‘₯
π‘₯ ∈ ℝ𝑑
is too strong. One can
throughout the paper, then it suces to
π‘₯.
𝑑 ∈ [0, 𝑇 ] an SDE on [𝑑, 𝑇 ] × Ξ©π‘‘ induced by πœ‡
and 𝜎 . Of course, the second argument of πœ‡ and 𝜎 requires a path on [0, 𝑇 ],
so that it is necessary to specify a history for the SDE on [0, 𝑑]. This role
is played by an arbitrary path πœ‚ ∈ Ξ©. Given πœ‚ , we dene the conditioned
Next, we introduce for xed
coecients
πœ‡π‘‘,πœ‚ : [0, 𝑇 ] × Ξ©π‘‘ × π‘ˆ β†’ ℝ𝑑 ,
πœ‡π‘‘,πœ‚ (π‘Ÿ, πœ”, 𝑒) := πœ‡(π‘Ÿ, πœ‚ βŠ—π‘‘ πœ”, 𝑒),
𝜎 𝑑,πœ‚ : [0, 𝑇 ] × Ξ©π‘‘ × π‘ˆ β†’ ℝ𝑑×𝑑 ,
𝜎 𝑑,πœ‚ (π‘Ÿ, πœ”, 𝑒) := 𝜎(π‘Ÿ, πœ‚ βŠ—π‘‘ πœ”, 𝑒).
πœ” is a path not necesat (π‘Ÿ, πœ”, 𝑒) depends only
(More precisely, these functions are dened also when
sarily starting at the origin, but clearly their value
6
on
πœ” 𝑑 .)
We observe that the Lipschitz condition (2.2) is inherited; indeed,
βˆ£πœ‡π‘‘,πœ‚ (π‘Ÿ, πœ”, 𝑒) βˆ’ πœ‡π‘‘,πœ‚ (π‘Ÿ, πœ” β€² , 𝑒)∣+∣𝜎 𝑑,πœ‚ (π‘Ÿ, πœ”, 𝑒) βˆ’ 𝜎 𝑑,πœ‚ (π‘Ÿ, πœ” β€² , 𝑒)∣
≀ 𝐾βˆ₯πœ‚ βŠ—π‘‘ πœ” βˆ’ πœ‚ βŠ—π‘‘ πœ” β€² βˆ₯[𝑑,π‘Ÿ]
= 𝐾βˆ₯πœ” 𝑑 βˆ’ πœ” ′𝑑 βˆ₯[𝑑,π‘Ÿ]
≀ 2𝐾βˆ₯πœ” βˆ’ πœ” β€² βˆ₯[𝑑,π‘Ÿ] .
𝔽𝑑 -progressively∫measurable, π‘ˆ -valued pro𝑇
2
cesses 𝜈 such that
𝑑 βˆ£πœ‡(π‘Ÿ, 𝑋, πœˆπ‘Ÿ )∣ π‘‘π‘Ÿ < ∞ and 𝑑 ∣𝜎(π‘Ÿ, 𝑋, πœˆπ‘Ÿ )∣ π‘‘π‘Ÿ < ∞ for
𝑑
𝑑
any continuous 𝔽 -adapted process 𝑋 = {π‘‹π‘Ÿ , π‘Ÿ ∈ [0, 𝑇 ]}. For 𝜈 ∈ 𝒰 , the
We denote by
𝒰 𝑑 the
βˆ«π‘‡
set of all
SDE
∫
𝑠
∫
𝑑,πœ‚
𝑠
πœ‡ (π‘Ÿ, 𝑋, πœˆπ‘Ÿ ) π‘‘π‘Ÿ +
𝑋𝑠 = πœ‚π‘‘ +
𝜎 𝑑,πœ‚ (π‘Ÿ, 𝑋, πœˆπ‘Ÿ ) π‘‘π΅π‘Ÿπ‘‘ ,
𝑑≀𝑠≀𝑇
under
𝑃0𝑑
𝑑
𝑑
(2.6)
has a unique solution
𝑋(𝑑, πœ‚, 𝜈)
on
[𝑑, 𝑇 ].
Similarly as above, we dene
(
)βˆ’1
𝑃 (𝑑, πœ‚, 𝜈) := 𝑃0𝑑 ∘ 𝑋(𝑑, πœ‚, 𝜈)𝑑
𝑋(𝑑, πœ‚, 𝜈)𝑑 ≑ 𝑋(𝑑, πœ‚, 𝜈) βˆ’ πœ‚π‘‘ on Ω𝑑 . Note that this
notation 𝑃 (0, π‘₯, 𝜈) if π‘₯ is seen as a constant path.
to be the distribution of
consistent with the
2.3
is
The Value Function
We can now dene the value function for the case when the reward functional
πœ‰
is an element of
functions on
UC𝑏 (Ξ©), the space of bounded uniformly continuous
Ξ©.
Denition 2.4.
Given
𝑑 ∈ [0, 𝑇 ]
and
πœ‰ ∈ UC𝑏 (Ξ©),
we dene the value
function
𝑉𝑑 (πœ”) = 𝑉𝑑 (πœ‰; πœ”) = sup 𝐸 𝑃 (𝑑,πœ”,𝜈) [πœ‰ 𝑑,πœ” ],
(𝑑, πœ”) ∈ [0, 𝑇 ] × Ξ©.
(2.7)
πœˆβˆˆπ’° 𝑑
The function
πœ‰
is xed throughout Sections 2 and 3 and hence often
suppressed in the notation. In view of the double dependence on
the measurability of
𝑉𝑑
is not obvious.
πœ”
in (2.7),
We have the following regularity
result.
Proposition 2.5. Let 𝑑 ∈ [0, 𝑇 ] and πœ‰ ∈ UC𝑏 (Ξ©). Then 𝑉𝑑 ∈ UC𝑏 (Ω𝑑 ) and
in particular 𝑉𝑑 is ℱ𝑑 -measurable. More precisely,
βˆ£π‘‰π‘‘ (πœ”) βˆ’ 𝑉𝑑 (πœ” β€² )∣ ≀ 𝜌(βˆ₯πœ” βˆ’ πœ” β€² βˆ₯𝑑 )
for all πœ”, πœ” β€² ∈ Ξ©
with a modulus of continuity 𝜌 depending only on πœ‰ , the Lipschitz constant
𝐾 and the time horizon 𝑇 .
7
The rst source of regularity for
𝑉𝑑
is our assumption that
πœ‰
is uniformly
continuous; the second one is the Lipschitz property of the SDE. Before
stating the proof of the proposition, we examine the latter aspect in detail.
Lemma 2.6. Let πœ“ ∈ UC𝑏 (Ω𝑑 ). There exists a modulus of continuity 𝜌𝐾,𝑇,πœ“ ,
depending only on 𝐾, 𝑇 and the minimal modulus of continuity of πœ“ , such
that
(
)
𝐸 𝑃 (𝑑,πœ”,𝜈) [πœ“] βˆ’ 𝐸 𝑃 (𝑑,πœ”β€² ,𝜈) [πœ“] ≀ 𝜌𝐾,𝑇,πœ“ βˆ₯πœ” βˆ’ πœ” β€² βˆ₯𝑑
for all 𝑑 ∈ [0, 𝑇 ], 𝜈 ∈ 𝒰 𝑑 and πœ”, πœ” β€² ∈ Ξ©.
Proof.
𝑑
𝐸[ β‹… ] := 𝐸 𝑃0 [ β‹… ] to alleviate the notation. Let πœ”, πœ”
¯ ∈ Ξ©, set
𝑑
¯
𝑋 := 𝑋(𝑑, πœ”, 𝜈) and 𝑋 := 𝑋(𝑑, πœ”
¯ , 𝜈), and recall that 𝑋 = 𝑋 βˆ’ 𝑋𝑑 = 𝑋 βˆ’ πœ”π‘‘
¯π‘‘ = 𝑋
¯ βˆ’πœ”
and similarly 𝑋
¯π‘‘.
𝑑
𝑑
𝑑
(i) We begin with a standard SDE estimate. Let 𝑋 = 𝑀 + 𝐴 and
¯π‘‘ = 𝑀
¯ 𝑑 + 𝐴¯π‘‘ be the semimartingale decompositions and 𝑑 ≀ 𝜏 ≀ 𝑇 be
𝑋
𝑑 ¯ 𝑑 , 𝐴𝑑 , 𝐴
¯π‘‘ are bounded on [𝑑, 𝜏 ]. Then Itô's
a stopping time such that 𝑀 , 𝑀
formula and the Lipschitz property (2.2) of 𝜎 yield that
∫ 𝜏
𝑑
𝑑 2
¯
¯ πœˆπ‘Ÿ )∣2 π‘‘π‘Ÿ
𝐸[βˆ£π‘€πœ βˆ’ π‘€πœ ∣ ] ≀ 𝐸
∣𝜎(π‘Ÿ, πœ” βŠ—π‘‘ 𝑋, πœˆπ‘Ÿ ) βˆ’ 𝜎(π‘Ÿ, πœ”
¯ βŠ—π‘‘ 𝑋,
𝑑
∫ 𝜏
2
¯ 2 π‘‘π‘Ÿ
≀𝐾 𝐸
βˆ₯πœ” βŠ—π‘‘ 𝑋 βˆ’ πœ”
¯ βŠ—π‘‘ 𝑋βˆ₯
π‘Ÿ
βˆ«π‘‘ 𝜏
(
)
¯ 𝑑 βˆ₯[𝑑,π‘Ÿ] 2 π‘‘π‘Ÿ
≀ 𝐾 2𝐸
βˆ₯πœ” βˆ’ πœ”
¯ βˆ₯𝑑 + βˆ₯𝑋 𝑑 βˆ’ 𝑋
𝑑
∫ 𝑇
[
]
2
2
2
¯ 𝑑 βˆ₯2
𝐸 βˆ₯𝑋 𝑑 βˆ’ 𝑋
≀ 2𝐾 𝑇 βˆ₯πœ” βˆ’ πœ”
¯ βˆ₯𝑑 + 2𝐾
[𝑑,π‘Ÿβˆ§πœ ] π‘‘π‘Ÿ.
We set
𝑑
Hence, Doob's maximal inequality implies that
[
]
𝑑
¯ 𝑑 βˆ₯2
¯π‘‘ 2
𝐸 βˆ₯𝑀 𝑑 βˆ’ 𝑀
[𝑑,𝜏 ] ≀ 4𝐸[βˆ£π‘€πœ βˆ’ π‘€πœ ∣ ]
≀ 8𝐾 2 𝑇 βˆ₯πœ” βˆ’ πœ”
¯ βˆ₯2𝑑 + 8𝐾 2
𝑇
∫
𝑑
Moreover, using the Lipschitz property (2.2) of
βˆ£π΄π‘‘π‘  βˆ’ 𝐴¯π‘‘𝑠 ∣ ≀
∫
πœ‡,
[
]
¯ 𝑑 βˆ₯2
𝐸 βˆ₯𝑋 𝑑 βˆ’ 𝑋
[𝑑,π‘Ÿβˆ§πœ ] π‘‘π‘Ÿ.
we also have that
𝑠
¯ πœˆπ‘Ÿ )∣ π‘‘π‘Ÿ
βˆ£πœ‡(π‘Ÿ, πœ” βŠ—π‘‘ 𝑋, πœˆπ‘Ÿ ) βˆ’ πœ‡(π‘Ÿ, πœ”
¯ βŠ—π‘‘ 𝑋,
𝑑
∫ 𝑠
(
)
¯ 𝑑 βˆ₯[𝑑,π‘Ÿ] π‘‘π‘Ÿ
≀𝐾
βˆ₯πœ” βˆ’ πœ”
¯ βˆ₯𝑑 + βˆ₯𝑋 𝑑 βˆ’ 𝑋
𝑑
for all
𝑑≀𝑠≀𝑇
and then Jensen's inequality yields that
]
𝐸 βˆ₯𝐴 βˆ’ 𝐴¯π‘‘ βˆ₯2[𝑑,𝜏 ] ≀ 2𝐾 2 𝑇 2 βˆ₯πœ” βˆ’ πœ”
¯ βˆ₯2𝑑 + 2𝐾 2 𝑇
[
𝑑
8
∫
𝑑
𝑇
[
]
¯ 𝑑 βˆ₯2
𝐸 βˆ₯𝑋 𝑑 βˆ’ 𝑋
[𝑑,π‘Ÿβˆ§πœ ] π‘‘π‘Ÿ.
Hence, we have shown that
]
¯ 𝑑 βˆ₯2
𝐸 βˆ₯𝑋 βˆ’ 𝑋
¯ βˆ₯2𝑑 + 𝐢0
[𝑑,𝜏 ] ≀ 𝐢0 βˆ₯πœ” βˆ’ πœ”
𝑑
[
where
𝐢0
depends only on
𝐾
and
𝑇,
∫
𝑇
𝑑
[
]
¯ 𝑑 βˆ₯2
𝐸 βˆ₯𝑋 𝑑 βˆ’ 𝑋
[𝑑,π‘Ÿβˆ§πœ ] π‘‘π‘Ÿ,
and we conclude by Gronwall's lemma
that
[
]
¯ 𝑑 βˆ₯2
𝐸 βˆ₯𝑋 𝑑 βˆ’ 𝑋
¯ βˆ₯2𝑑 ,
[𝑑,𝜏 ] ≀ 𝐢βˆ₯πœ” βˆ’ πœ”
𝐢 := 𝐢0 𝑒𝐢0 𝑇 .
By the continuity of their sample paths, there exists a localizing sequence
(πœπ‘› )𝑛β‰₯1
each 𝑛.
of stopping times such that
¯ 𝑑 , 𝐴𝑑 , 𝐴¯π‘‘
𝑀 𝑑, 𝑀
are bounded on
[𝑑, πœπ‘› ] for
Therefore, monotone convergence and the previous inequality yield
that
[
]
¯ 𝑑 βˆ₯2
𝐸 βˆ₯𝑋 𝑑 βˆ’ 𝑋
¯ βˆ₯2𝑑 .
[𝑑,𝑇 ] ≀ 𝐢βˆ₯πœ” βˆ’ πœ”
(ii) Let
𝜌˜ be
(2.8)
the minimal (nondecreasing) modulus of continuity for
πœ“,
{
}
𝜌˜(𝑧) := sup βˆ£πœ“(˜
πœ” ) βˆ’ πœ“(˜
πœ” β€² )∣ : πœ”
˜, πœ”
˜ β€² ∈ Ω𝑑 , βˆ₯˜
πœ”βˆ’πœ”
˜ β€² βˆ₯[𝑑,𝑇 ] ≀ 𝑧 ,
𝜌 is a bounded continuous function
satisfying 𝜌(0) = 0 and 𝜌 β‰₯ 𝜌
˜. Let 𝑃 := 𝑃 (𝑑, πœ”, 𝜈) and 𝑃¯ := 𝑃 (𝑑, πœ”
¯ , 𝜈), then
¯
¯ 𝑑 , respectively; therefore,
𝑃 and 𝑃 are the distributions of 𝑋 𝑑 and 𝑋
𝑃
[ (
)]
¯ 𝑑 βˆ₯[𝑑,𝑇 ] . (2.9)
¯ 𝑑 )] ≀ 𝐸 𝜌 βˆ₯𝑋 𝑑 βˆ’ 𝑋
𝐸 [πœ“] βˆ’ 𝐸 𝑃¯ [πœ“] = 𝐸[πœ“(𝑋 𝑑 ) βˆ’ πœ“(𝑋
and let
𝜌
be the concave hull of
𝜌˜.
Then
Moreover, Jensen's inequality and (2.8) yield that
[ (
)]
( [
])
¯ 𝑑 βˆ₯[𝑑,𝑇 ] ≀ 𝜌 𝐸 βˆ₯𝑋 𝑑 βˆ’ 𝑋
¯ 𝑑 βˆ₯[𝑑,𝑇 ]
𝐸 𝜌 βˆ₯𝑋 𝑑 βˆ’ 𝑋
( [
]1/2 )
¯ 𝑑 βˆ₯2
≀ 𝜌 𝐸 βˆ₯𝑋 𝑑 βˆ’ 𝑋
[𝑑,𝑇 ]
(√
)
≀ 𝜌 𝐢βˆ₯πœ” βˆ’ πœ”
¯ βˆ₯𝑑
√
𝑃
𝑃¯
for every 𝑛. In view of (2.9), we have ∣𝐸 [πœ“] βˆ’ 𝐸 [πœ“]∣ ≀ 𝜌( 𝐢βˆ₯πœ” βˆ’ πœ”
¯ βˆ₯𝑑 );
√
i.e., the result holds for 𝜌𝐾,𝑇,πœ“ (𝑧) := 𝜌( 𝐢𝑧).
After these preparations, we can prove the continuity of
Proof of Proposition 2.5.
𝑉𝑑 .
To disentangle the double dependence on
πœ” in (2.7),
we rst consider the function
(πœ‚, πœ”) 7β†’ 𝐸 𝑃 (𝑑,πœ‚,𝜈) [πœ‰ 𝑑,πœ” ],
Since
πœ‰ ∈ UC𝑏 (Ξ©),
(πœ‚, πœ”) ∈ Ξ© × Ξ©.
there exists a modulus of continuity
βˆ£πœ‰(πœ”) βˆ’ πœ‰(πœ” β€² )∣ ≀ 𝜌(πœ‰) (βˆ₯πœ” βˆ’ πœ” β€² βˆ₯𝑇 ),
Therefore, we have for all
βˆ£πœ‰ 𝑑,πœ” (˜
πœ”) βˆ’ πœ‰
πœ”
˜ ∈ Ω𝑑
𝑑,πœ” β€²
𝜌(πœ‰)
for
πœ‰;
i.e.,
πœ”, πœ” β€² ∈ Ξ©.
that
(˜
πœ” )∣ = βˆ£πœ‰(πœ” βŠ—π‘‘ πœ”
˜ ) βˆ’ πœ‰(πœ” β€² βŠ—π‘‘ πœ”
˜ )∣
≀ 𝜌(πœ‰) (βˆ₯πœ” βŠ—π‘‘ πœ”
˜ βˆ’ πœ” β€² βŠ—π‘‘ πœ”
˜ βˆ₯𝑇 )
= 𝜌(πœ‰) (βˆ₯πœ” βˆ’ πœ” β€² βˆ₯𝑑 ).
9
(2.10)
For xed
Fix
for
πœ“πœ” ;
πœ‚ ∈ Ξ©, it follows that
𝑃 (𝑑,πœ‚,𝜈) 𝑑,πœ”
β€² 𝐸
[πœ‰ ] βˆ’ 𝐸 𝑃 (𝑑,πœ‚,𝜈) [πœ‰ 𝑑,πœ” ] ≀ 𝜌(πœ‰) (βˆ₯πœ” βˆ’ πœ” β€² βˆ₯𝑑 ).
πœ”βˆˆΞ©
and let
πœ“πœ” := πœ‰ 𝑑,πœ” .
Then
𝜌(πœ‰)
(2.11)
yields a modulus of continuity
πœ” . Thus
πœ‚ 7β†’ 𝐸 𝑃 (𝑑,πœ‚,𝜈) [πœ“πœ” ] admits a modulus of
𝑇, 𝐾, πœ‰ . In view of (2.11), we conclude
in particular, this modulus of continuity is uniform in
Lemma 2.6 implies that the mapping
continuity
πœŒπ‘‡,𝐾,πœ‰
depending only on
that
𝑃 (𝑑,πœ‚,𝜈) 𝑑,πœ”
β€²
β€² 𝐸
[πœ‰ ] βˆ’ 𝐸 𝑃 (𝑑,πœ‚ ,𝜈) [πœ‰ 𝑑,πœ” ] ≀ 𝜌(πœ‰) (βˆ₯πœ” βˆ’ πœ” β€² βˆ₯𝑑 ) + πœŒπ‘‡,𝐾,πœ‰ (βˆ₯πœ‚ βˆ’ πœ‚ β€² βˆ₯𝑑 )
for all
πœ‚, πœ‚ β€² , πœ”, πœ” β€² ∈ Ξ©
and in particular that
𝑃 (𝑑,πœ”,𝜈) 𝑑,πœ”
β€²
β€² 𝐸
[πœ‰ ] βˆ’ 𝐸 𝑃 (𝑑,πœ” ,𝜈) [πœ‰ 𝑑,πœ” ] ≀ 𝜌(βˆ₯πœ” βˆ’ πœ” β€² βˆ₯𝑑 ),
𝜌 := 𝜌(πœ‰) + πœŒπ‘‡,𝐾,πœ‰
πœ”, πœ” β€² ∈ Ξ©. Passing to the supremum over 𝜈 ∈ 𝒰 𝑑 , we
βˆ£π‘‰π‘‘ (πœ”) βˆ’ 𝑉𝑑 (πœ” β€² )∣ ≀ 𝜌(βˆ₯πœ” βˆ’ πœ” β€² βˆ₯𝑑 ) for all πœ”, πœ” β€² ∈ Ξ©, which was the
for all
3
(2.12)
obtain that
claim.
Pathwise Dynamic Programming
In this section, we provide a pathwise dynamic programming principle which
is fundamental for the subsequent sections. As we are working in the weak
formulation (1.4), the arguments used here are similar to, e.g., [26], while
[22] gives a related construction in the strong formulation (i.e., working only
under
𝑃0 ).
We assume in this section that the following conditional version of Assumption 2.1 holds true; however, we shall see later (Lemma 4.4) that this
extended assumption holds automatically outside certain nullsets.
Assumption 3.1.
Throughout Section 3, we assume that
𝔽𝑋
for all
𝑃0𝑑
βŠ‡ 𝔽𝑑
for
𝑋 := 𝑋(𝑑, πœ‚, 𝜈),
(3.1)
(𝑑, πœ‚, 𝜈) ∈ [0, 𝑇 ] × Ξ© × π’° 𝑑 .
The main result of this section is the following dynamic programming
principle.
We shall also provide more general, quasi-sure versions of this
result later (the nal form being Theorem 5.2).
Theorem 3.2. Let
Then
0 ≀ 𝑠 ≀ 𝑑 ≀ 𝑇 , πœ‰ ∈ UC𝑏 (Ξ©) and set π‘‰π‘Ÿ (β‹…) = π‘‰π‘Ÿ (πœ‰; β‹…).
[
]
𝑉𝑠 (πœ”) = sup 𝐸 𝑃 (𝑠,πœ”,𝜈) (𝑉𝑑 )𝑠,πœ”
πœˆβˆˆπ’° 𝑠
10
for all πœ” ∈ Ξ©.
(3.2)
We remark that in view of Proposition 2.5, we may see
mapping
UC𝑏 (Ξ©) β†’ UC𝑏 (Ξ©)
𝑉𝑠 = 𝑉𝑠 ∘ 𝑉𝑑
πœ‰ 7β†’ π‘‰π‘Ÿ (πœ‰; β‹…)
as a
and recast (3.2) as the semigroup property
on
UC𝑏 (Ξ©)
0 ≀ 𝑠 ≀ 𝑑 ≀ 𝑇.
for all
(3.3)
Some auxiliary results are needed for the proof of Theorem 3.2, which is
stated at the end of this section. We start with the (well known) observation
that conditioning the solution of an SDE yields the solution of a suitably
conditioned SDE.
Lemma 3.3. Let
then
¯ := 𝑋(𝑠, πœ”
0 ≀ 𝑠 ≀ 𝑑 ≀ 𝑇 , 𝜈 ∈ 𝒰 𝑠 and πœ”
¯ ∈ Ξ©. If 𝑋
¯ , 𝜈),
𝑃0𝑑 -a.s.
(
)
¯ 𝑑,πœ” = 𝑋 𝑑, πœ”
¯
𝑋
¯ βŠ—π‘  𝑋(πœ”),
𝜈 𝑑,πœ”
for all πœ” ∈ Ω𝑠 .
Proof.
¯ , we have
πœ” ∈ Ω𝑠 . Using the denition and the ow property of 𝑋
∫ π‘Ÿ
∫ π‘Ÿ
𝑠,¯
πœ”
¯
¯
¯ πœˆπ‘’ ) 𝑑𝐡𝑒𝑠
π‘‹π‘Ÿ = πœ”
¯π‘  +
πœ‡ (𝑒, 𝑋, πœˆπ‘’ ) 𝑑𝑒 +
𝜎 𝑠,¯πœ” (𝑒, 𝑋,
𝑠
βˆ«π‘  π‘Ÿ
∫ π‘Ÿ
¯π‘‘ +
¯ πœˆπ‘’ ) 𝑑𝑒 +
¯ πœˆπ‘’ ) 𝑑𝐡 𝑠 𝑃 𝑠 -a.s.
=𝑋
πœ‡(𝑒, πœ”
¯ βŠ—π‘  𝑋,
𝜎(𝑒, πœ”
¯ βŠ—π‘  𝑋,
𝑒
0
Let
𝑑
for all
π‘Ÿ ∈ [𝑑, 𝑇 ].
𝑑
Hence, using that
𝑠
the increments of 𝐡 ,
(𝑃0𝑠 )𝑑,πœ” = 𝑃0𝑑
by the
𝑃0𝑠 -independence
of
∫ π‘Ÿ
∫ π‘Ÿ
𝑑,πœ”
𝑑,πœ” 𝑑,πœ”
𝑑,πœ”
¯
¯ 𝑑,πœ” , πœˆπ‘’π‘‘,πœ” ) 𝑑𝐡𝑒𝑑 𝑃0𝑑 -a.s.
¯
¯
¯ βŠ—π‘  𝑋 , πœˆπ‘’ ) 𝑑𝑒+ 𝜎(𝑒, πœ”
¯ βŠ—π‘  𝑋
π‘‹π‘Ÿ = 𝑋𝑑 + πœ‡(𝑒, πœ”
𝑑
𝑑
(3.4)
Since
¯
𝑋
is adapted, we have
¯ 𝑑,πœ” (β‹…) = 𝑋(πœ”
¯ βŠ—π‘‘ β‹…) = 𝑋(πœ”)
¯
𝑋
on
[𝑠, 𝑑]
and in
particular
¯ 𝑑,πœ” = πœ”
¯
¯ 𝑑,πœ” = πœ‚ βŠ—π‘‘ 𝑋
¯ 𝑑,πœ” ,
πœ”
¯ βŠ—π‘  𝑋
¯ βŠ—π‘  𝑋(πœ”)
βŠ—π‘‘ 𝑋
Therefore, recalling that
¯ π‘Ÿπ‘‘,πœ” = πœ‚π‘‘ +
𝑋
∫
π‘Ÿ
¯π‘  = πœ”
𝑋
¯π‘ ,
¯ 𝑑,πœ”
𝑋
¯
πœ‚ := πœ”
¯ βŠ—π‘  𝑋(πœ”).
(3.4) can be stated as
¯ 𝑑,πœ” , 𝜈 𝑑,πœ” ) 𝑑𝑒 +
πœ‡ (𝑒, 𝑋
𝑑,πœ‚
∫
𝑑
i.e.,
for
π‘Ÿ
¯ 𝑑,πœ” , 𝜈 𝑑,πœ” ) 𝑑𝐡 𝑑
𝜎 𝑑,πœ‚ (𝑒, 𝑋
𝑒
𝑃0𝑑 -a.s.;
𝑑
solves the SDE (2.6) for the parameters
(𝑑, πœ‚, 𝜈 𝑑,πœ” ).
Now the result
follows by the uniqueness of the solution to this SDE.
Given
𝑑 ∈ [0, 𝑇 ]
and
πœ” ∈ Ξ©,
we dene
{
}
𝒫(𝑑, πœ”) = 𝑃 (𝑑, πœ”, 𝜈) : 𝜈 ∈ 𝒰 𝑑 .
These sets have the following invariance property.
11
(3.5)
Lemma 3.4. Let 0 ≀ 𝑠 ≀ 𝑑 ≀ 𝑇 and πœ”¯ ∈ Ξ©. If 𝑃
for 𝑃 -a.e. πœ” ∈ Ω𝑠 .
𝑃 𝑑,πœ” ∈ 𝒫(𝑑, πœ”
¯ βŠ—π‘  πœ”)
Proof.
∈ 𝒫(𝑠, πœ”
¯ ), then
𝑃 ∈ 𝒫(𝑠, πœ”
¯ ), we have 𝑃 = 𝑃 (𝑠, πœ”
¯ , 𝜈) for some 𝜈 ∈ 𝒰 𝑠 ; i.e.,
setting 𝑋 := 𝑋(𝑠, πœ”
¯ , 𝜈), 𝑃 is the distribution of
∫ β‹…
∫ β‹…
𝑋𝑠 =
πœ‡(π‘Ÿ, πœ”
¯ βŠ—π‘  𝑋, πœˆπ‘Ÿ ) π‘‘π‘Ÿ +
𝜎(π‘Ÿ, πœ”
¯ βŠ—π‘  𝑋, πœˆπ‘Ÿ ) π‘‘π΅π‘Ÿπ‘  under 𝑃0𝑠 .
Since
𝑠
𝑠
πœ‡
Λ†π‘Ÿ :=∫ πœ‡(π‘Ÿ, πœ”
¯ βŠ—π‘  βˆ«π‘‹, πœˆπ‘Ÿ ) and 𝜎
Λ†π‘Ÿ := 𝜎(π‘Ÿ, πœ”
¯ βŠ—π‘  𝑋, πœˆπ‘Ÿ ) and see the above as
β‹…
β‹…
𝑠
integral
Λ†π‘Ÿ π‘‘π‘Ÿ + 𝑠 𝜎
Λ†π‘Ÿ π‘‘π΅π‘Ÿ rather than an SDE. As in [26, Lemma 2.2],
π‘ πœ‡
We set
the
the nondegeneracy assumption (3.1) implies the existence of a progressively
measurable transformation
π›½πœˆ : Ω𝑠 β†’ Ω𝑠
(depending on
π›½πœˆ (𝑋 𝑠 ) = 𝐡 𝑠
𝑠, πœ”
¯, 𝜈)
such that
𝑃0𝑠 -a.s.
(3.6)
Furthermore, a rather tedious calculation as in the proof of [26, Lemma 4.1]
shows that
𝑃
𝑑,πœ”
=
𝑃0𝑑
(∫
∘
β‹…
πœ‡
Λ†π‘Ÿπ‘‘,π›½πœˆ (πœ”) π‘‘π‘Ÿ
∫
+
𝑠
β‹…
𝜎
Λ†π‘Ÿπ‘‘,π›½πœˆ (πœ”) π‘‘π΅π‘Ÿπ‘‘
)βˆ’1
for
𝑃 -a.e. πœ” ∈ Ω𝑠 .
𝑠
πœ”
Λ‡ := π›½πœˆ (πœ”), we have
(
)
(
)
= πœ‡ π‘Ÿ, πœ”
¯ βŠ—π‘  𝑋 𝑑,Λ‡πœ” , πœˆπ‘Ÿπ‘‘,Λ‡πœ” = πœ‡ π‘Ÿ, πœ”
¯ βŠ—π‘  𝑋(Λ‡
πœ” ) βŠ—π‘‘ 𝑋 𝑑,Λ‡πœ” , πœˆπ‘Ÿπ‘‘,Λ‡πœ”
Note that, abbreviating
𝜈 (πœ”)
πœ‡
ˆ𝑑,𝛽
π‘Ÿ
and similarly for
𝜎
Λ† 𝑑,π›½πœˆ (πœ”) .
Hence, we deduce by Lemma 3.3 that
𝑃 𝑑,πœ” = 𝑃 (𝑑, πœ”
¯ βŠ—π‘  𝑋(Λ‡
πœ” ), 𝜈 𝑑,Λ‡πœ” )
𝑃 -a.e. πœ” ∈ Ω𝑠 .
for
In view of (3.6), we have
𝑋 𝑠 (π›½πœˆ (𝐡 𝑠 )) = 𝐡 𝑠
i.e.,
𝑋(Λ‡
πœ” )𝑠 = 𝑋 𝑠 (Λ‡
πœ”) = πœ”
for
𝑃 -a.e. πœ” ∈ Ω𝑠 ,
𝑃 𝑑,πœ” = 𝑃 (𝑑, πœ”
¯ βŠ—π‘  πœ”, 𝜈 𝑑,Λ‡πœ” )
In particular,
𝑃 -a.s.;
(3.7)
and we conclude that
for
𝑃 -a.e. πœ” ∈ Ω𝑠 .
(3.8)
𝑃 𝑑,πœ” ∈ 𝒫(𝑑, πœ”
¯ βŠ—π‘  πœ”).
Lemma 3.5 (Pasting). Let
0 ≀ 𝑠 ≀ 𝑑 ≀ 𝑇, πœ”
¯ ∈ Ξ©, 𝜈 ∈ 𝒰 𝑠 and set
𝑖
𝑃 := 𝑃 (𝑠, πœ”
¯ , 𝜈), 𝑋 := 𝑋(𝑠, πœ”
¯ , 𝜈). Let (𝐸 )0≀𝑖≀𝑁 be a nite ℱ𝑑𝑠 -measurable
𝑠
𝑖
𝑑
partition of Ξ© , 𝜈 ∈ 𝒰 for 1 ≀ 𝑖 ≀ 𝑁 and dene 𝜈¯ ∈ 𝒰 𝑠 by
[
]
𝑁
βˆ‘
𝑠
𝑖 𝑑
𝑠
𝜈¯(πœ”) := 1[𝑠,𝑑) 𝜈(πœ”) + 1[𝑑,𝑇 ] 𝜈(πœ”)1𝐸 0 (𝑋(πœ”) ) +
𝜈 (πœ” )1𝐸 𝑖 (𝑋(πœ”) ) .
𝑖=1
Then 𝑃¯ := 𝑃 (𝑠, πœ”¯ , 𝜈¯) satises 𝑃¯ = 𝑃 on
𝑃¯ 𝑑,πœ” = 𝑃 (𝑑, πœ”
¯ βŠ—π‘  πœ”, 𝜈 𝑖 )
ℱ𝑑𝑠
and
for 𝑃 -a.e. πœ” ∈ 𝐸 𝑖 ,
12
1 ≀ 𝑖 ≀ 𝑁.
Proof.
As 𝜈
¯ = 𝜈 on [𝑠, 𝑑), we have 𝑋(𝑠, πœ”
¯ , 𝜈¯) = 𝑋 on [𝑠, 𝑑] and in particular
𝑃¯ = 𝑃 on ℱ𝑑𝑠 . Let 1 ≀ 𝑖 ≀ 𝑁 , we show that 𝑃¯ 𝑑,πœ” = 𝑃 (𝑑, πœ”
¯ βŠ—π‘  πœ”, 𝜈 𝑖 ) for
𝑃 -a.e. πœ” ∈ 𝐸 𝑖 . Recall from (3.8) that
𝑃¯ 𝑑,πœ” = 𝑃 (𝑑, πœ”
¯ βŠ—π‘  πœ”, 𝜈¯π‘‘,π›½πœˆ¯ (πœ”) )
where
π›½πœˆ¯
for
𝑃¯ -a.e. πœ” ∈ Ω𝑠 ,
is dened as in (3.6). Since both sides of this equality depend only
on the restriction of
πœ”
to
[𝑠, 𝑑] and 𝑋(𝑠, πœ”
¯ , 𝜈¯) = 𝑋
𝑃¯ 𝑑,πœ” = 𝑃 (𝑑, πœ”
¯ βŠ—π‘  πœ”, 𝜈¯π‘‘,Λ‡πœ” )
on
[𝑠, 𝑑], we also have that
𝑃 -a.e. πœ” ∈ Ω𝑠 ,
for
(3.9)
πœ”
Λ‡ = π›½πœˆ (πœ”) is dened as below (3.6). Note that by (3.7), πœ” ∈ 𝐸 𝑖
implies 𝑋(Λ‡
πœ” )𝑠 ∈ 𝐸 𝑖 under 𝑃 . (More precisely, if 𝐴 βŠ† Ω𝑠 is a set such that
𝐴 βŠ† 𝐸 𝑖 𝑃 -a.s., then {𝑋(Λ‡
πœ” )𝑠 : πœ” ∈ 𝐴} βŠ† 𝐸 𝑖 𝑃 -a.s.) In fact, since 𝑋 is
𝑖
𝑠
adapted and 𝐸 ∈ ℱ𝑑 , we even have that
where
𝑋(Λ‡
πœ” βŠ—π‘‘ πœ”
˜ )𝑠 ∈ 𝐸 𝑖
By the denition of
𝜈¯,
πœ”
˜ ∈ Ω𝑑 ,
for all
for
𝑃 -a.e. πœ” ∈ 𝐸 𝑖 .
we conclude that
𝜈¯π‘‘,Λ‡πœ” (˜
πœ” ) = 𝜈¯(Λ‡
πœ” βŠ—π‘‘ πœ”
˜ ) = 𝜈 𝑖 ((Λ‡
πœ” βŠ—π‘‘ πœ”
˜ )𝑑 ) = 𝜈 𝑖 (˜
πœ” ),
πœ”
˜ ∈ Ω𝑑 ,
for
𝑃 -a.e. πœ” ∈ 𝐸 𝑖 .
In view of (3.9), this yields the claim.
Remark 3.6.
In [26] and [17], it was possible to use a pasting of measures
as follows: in the notation of Lemma 3.5, it was possible to specify measures
𝑃 𝑖 on Ω𝑑 , corresponding
to certain admissible controls, and use the pasting
[
]
βˆ‘π‘
0
Λ†
𝑃 (𝐴) := 𝑃 (𝐴 ∩ 𝐸 ) + 𝑖=1 𝐸 𝑃 𝑃 𝑖 (𝐴𝑑,πœ” )1𝐸 𝑖 (πœ”) to obtain a measure in Ω𝑠
which again corresponded to some admissible control and satised
𝑃ˆ 𝑑,πœ” = 𝑃 𝑖
𝑃ˆ -a.e. πœ” ∈ 𝐸 𝑖 ,
for
(3.10)
which was then used in the proof of the dynamic programming principle.
This is not possible in our SDE-driven setting.
𝑃ˆ
Indeed, suppose that
𝑃 (𝑠, πœ”
¯ , 𝜈) for some πœ”
¯ and 𝜈 , then we see from (3.8) that,
𝑑,πœ”
Λ†
when 𝜎 is general, 𝑃
will depend explicitly on πœ” , which contradicts (3.10).
is of the form
Therefore, the subsequent proof uses an argument where (3.10) holds only
at one specic
πœ” ∈ 𝐸𝑖;
on the rest of
𝐸𝑖,
we conne ourselves to controlling
the error.
We can now show the dynamic programming principle. Apart from the
dierence remarked above, the basic pattern of the proof is the same as in
[26, Proposition 4.7].
Proof of Theorem 3.2.
Using the notation (3.5), our claim (3.2) can be stated
as
sup
𝑃 βˆˆπ’«(𝑠,¯
πœ”)
[
]
𝐸 𝑃 πœ‰ 𝑠,¯πœ” =
sup
[
]
𝐸 𝑃 (𝑉𝑑 )𝑠,¯πœ”
𝑃 βˆˆπ’«(𝑠,¯
πœ”)
13
for all
πœ”
¯ ∈ Ξ©.
(3.11)
(i) We rst show the inequality ≀ in (3.11). Fix
πœ”
¯ ∈ Ξ© and 𝑃 ∈ 𝒫(𝑠, πœ”
¯ ).
𝑃 𝑑,πœ” ∈ 𝒫(𝑑, πœ”
¯ βŠ—π‘  πœ”) for 𝑃 -a.e. πœ” ∈ Ω𝑠 and hence that
]
]
𝑑,πœ” [
𝑑,πœ” [ 𝑑,¯
𝐸𝑃
(πœ‰ 𝑠,¯πœ” )𝑑,πœ” = 𝐸 𝑃
πœ‰ πœ” βŠ—π‘  πœ”
]
β€²[
≀
sup
𝐸 𝑃 πœ‰ 𝑑,¯πœ”βŠ—π‘  πœ”
Lemma 3.4 shows that
𝑃 β€² βˆˆπ’«(𝑑,¯
πœ” βŠ—π‘  πœ”)
= 𝑉𝑑 (¯
πœ” βŠ—π‘  πœ”)
= 𝑉𝑑𝑠,¯πœ” (πœ”)
Since
𝑉𝑑
for
𝑃 -a.e. πœ” ∈ Ω𝑠 .
is measurable by Proposition 2.5, we can take
𝑃 (π‘‘πœ”)-expectations
on both sides to obtain that
[ 𝑑,πœ” [
[
]
]]
[
]
𝐸 𝑃 πœ‰ 𝑠,¯πœ” = 𝐸 𝑃 𝐸 𝑃
(πœ‰ 𝑠,¯πœ” )𝑑,πœ” ≀ 𝐸 𝑃 𝑉𝑑𝑠,¯πœ” .
𝑃 ∈ 𝒫(𝑠, πœ”
¯ ) on both sides and obtain the claim.
(ii) We now show the inequality β‰₯ in (3.11). Fix πœ”
¯ ∈ Ξ©, 𝜈 ∈ 𝒰 𝑠 and
let 𝑃 = 𝑃 (𝑠, πœ”
¯ , 𝜈). We x πœ€ > 0 and construct a countable cover of the state
We take the supremum over
space as follows.
Let
πœ”
Λ† ∈ Ω𝑠 . By the denition of 𝑉𝑑 (¯
πœ” βŠ—π‘  πœ”
Λ† ),
:= 𝑃 (𝑑, πœ”
¯ βŠ—π‘  πœ”
Λ† , 𝜈 (Λ†πœ”) ) satises
there exists
𝜈 (Λ†πœ”) ∈ 𝒰 𝑑
such
(Λ†
πœ”)
that 𝑃
𝑉𝑑 (¯
πœ” βŠ—π‘  πœ”
Λ† ) ≀ 𝐸𝑃
(πœ”)
Λ†
[πœ‰ 𝑑,¯πœ”βŠ—π‘  πœ”Λ† ] + πœ€.
(3.12)
𝐡(πœ€, πœ”
Λ† ) βŠ† Ω𝑠 denote the open βˆ₯ β‹… βˆ₯[𝑠,𝑑] -ball of radius πœ€ around πœ”
Λ† . Since
)
is
a
separable
(quasi-)metric
space
and
therefore
Lindelöf,
there
[𝑠,𝑑]
𝑖
𝑠
𝑖
𝑖
exists a sequence (Λ†
πœ” )𝑖β‰₯1 in Ξ© such that the balls 𝐡 := 𝐡(πœ€, πœ”
Λ† ) form a
𝑠
𝑖
𝑠
cover of Ξ© . As an βˆ₯ β‹… βˆ₯[𝑠,𝑑] -open set, each 𝐡 is ℱ𝑑 -measurable and hence
Let
(Ω𝑠 , βˆ₯ β‹… βˆ₯
𝐸 1 := 𝐡 1 ,
denes a partition
𝐸 𝑖+1 := 𝐡 𝑖+1 βˆ– (𝐸 1 βˆͺ β‹… β‹… β‹… βˆͺ 𝐸 𝑖 ),
𝑖β‰₯1
(𝐸 𝑖 )𝑖β‰₯1 of Ω𝑠 . Replacing 𝐸 𝑖 by
( 𝑖
)
𝐸 βˆͺ {Λ†
πœ” 𝑖 } βˆ– {Λ†
πœ” 𝑗 : 𝑗 β‰₯ 1, 𝑗 βˆ•= 𝑖}
if necessary, we may assume that
πœ”
Λ† 𝑖 ∈ 𝐸𝑖
for
𝑖 β‰₯ 1.
We set
𝜈 𝑖 := 𝜈 (Λ†πœ”
𝑖)
and
𝑃 𝑖 := 𝑃 (𝑑, πœ”
¯ βŠ—π‘  πœ”
Λ† 𝑖 , 𝜈 𝑖 ).
Next, we paste the controls
𝜈 𝑖.
Fix
𝑁 ∈ β„• and let 𝐴𝑁 := 𝐸 1 βˆͺ β‹… β‹… β‹… βˆͺ 𝐸 𝑁 ,
Ω𝑠 . Let 𝑋 := 𝑋(𝑠, πœ”
¯ , 𝜈), dene
𝑐
1
𝑁
then {𝐴𝑁 , 𝐸 , . . . , 𝐸 } is a nite partition of
[
𝑠
𝜈¯(πœ”) := 1[𝑠,𝑑) 𝜈(πœ”) + 1[𝑑,𝑇 ] 𝜈(πœ”)1𝐴𝑐𝑁 (𝑋(πœ”) ) +
𝑁
βˆ‘
𝑖
𝑑
𝑠
]
𝜈 (πœ” )1𝐸 𝑖 (𝑋(πœ”) )
𝑖=1
𝑃¯ := 𝑃 (𝑠, πœ”
¯ , 𝜈¯). Then, by Lemma 3.5, we have 𝑃¯ = 𝑃 on ℱ𝑑𝑠 and
˜ 𝑖 , for some subset 𝐸
˜ 𝑖 βŠ† 𝐸 𝑖 of full measure
= 𝑃 (𝑑, πœ”
¯ βŠ—π‘  πœ”, 𝜈 𝑖 ) for all πœ” ∈ 𝐸
and let
𝑃¯ 𝑑,πœ”
𝑃 . Let
us assume for the moment that
Λœπ‘–
πœ”
ˆ𝑖 ∈ 𝐸
for
14
1 ≀ 𝑖 ≀ 𝑁,
(3.13)
then we may conclude that
𝑖
𝑃¯ 𝑑,Λ†πœ” = 𝑃 𝑖
1 ≀ 𝑖 ≀ 𝑁.
for
Recall from Proposition 2.5 that
𝑉𝑑
(3.14)
admits a modulus of continuity
𝜌(𝑉𝑑 ) .
Moreover, we obtain similarly as in (2.10) that there exists a modulus of
continuity
𝜌(πœ‰)
such that
β€²
βˆ£πœ‰ 𝑑,¯πœ”βŠ—π‘  πœ” βˆ’ πœ‰ 𝑑,¯πœ”βŠ—π‘  πœ” ∣ ≀ 𝜌(πœ‰) (βˆ₯πœ” βˆ’ πœ” β€² βˆ₯[𝑠,𝑑] ).
Let
πœ” ∈ 𝐸 𝑖 βŠ† Ω𝑠
for some
1 ≀ 𝑖 ≀ 𝑁,
βˆ₯πœ” βˆ’ πœ”
Λ† 𝑖 βˆ₯[𝑠,𝑑] < πœ€.
then
Together
with (3.12) and (3.14), we obtain that
𝑉𝑑𝑠,¯πœ” (πœ”) ≀ 𝑉𝑑𝑠,¯πœ” (Λ†
πœ” 𝑖 ) + 𝜌(𝑉𝑑 ) (πœ€)
𝑖
𝑖
≀ 𝐸 𝑃 [πœ‰ 𝑑,¯πœ”βŠ—π‘  πœ”Λ† ] + πœ€ + 𝜌(𝑉𝑑 ) (πœ€)
ˆ𝑖
¯ 𝑑,πœ”
= 𝐸𝑃
𝑖
[πœ‰ 𝑑,¯πœ”βŠ—π‘  πœ”Λ† ] + πœ€ + 𝜌(𝑉𝑑 ) (πœ€).
(3.15)
Recall from (2.12) that the mapping
β€²
β€²
𝑖
πœ” β€² 7β†’ 𝐸 𝑃 (𝑑,πœ” ,𝜈 ) [πœ‰ 𝑑,πœ” ]
is uniformly continuous with a modulus
πœ”βˆˆ
𝜌˜
independent of
𝑖
and
𝑁.
Since
𝐸 𝑖 , it follows that
ˆ𝑖
¯ 𝑑,πœ”
𝐸𝑃
¯ 𝑑,πœ”
𝑖
[πœ‰ 𝑑,¯πœ”βŠ—π‘  πœ”Λ† ] βˆ’ 𝐸 𝑃
= 𝐸 𝑃 (𝑑,¯πœ”βŠ—π‘  πœ”Λ†
≀ 𝜌˜(πœ€)
Setting
for
𝑖 ,𝜈 𝑖 )
[πœ‰ 𝑑,¯πœ”βŠ—π‘  πœ” ]
𝑖
𝑃¯ -a.e. πœ” ∈ 𝐸 𝑖 .
𝜌(πœ€) := 𝜌˜(πœ€) + πœ€ + 𝜌(𝑉𝑑 ) (πœ€)
¯ 𝑑,πœ”
𝐸𝑃
𝑖
[πœ‰ 𝑑,¯πœ”βŠ—π‘  πœ”Λ† ] βˆ’ 𝐸 𝑃 (𝑑,¯πœ”βŠ—π‘  πœ”,𝜈 ) [πœ‰ 𝑑,¯πœ”βŠ—π‘  πœ” ]
and noting that
¯ 𝑑,πœ”
[πœ‰ 𝑑,¯πœ”βŠ—π‘  πœ” ] = 𝐸 𝑃
(3.16)
¯
[(πœ‰ 𝑠,¯πœ” )𝑑,πœ” ] = 𝐸 𝑃 [πœ‰ 𝑠,¯πœ” βˆ£β„±π‘‘π‘  ](πœ”),
the inequalities (3.15) and (3.16) imply that
]
¯[
𝑉𝑑𝑠,¯πœ” (πœ”) ≀ 𝐸 𝑃 πœ‰ 𝑠,¯πœ” ℱ𝑑𝑠 (πœ”) + 𝜌(πœ€)
for
on
(3.17)
𝑃¯ -a.e. (and thus 𝑃 -a.e.) πœ” ∈ 𝐸 𝑖 . This holds for all 1 ≀ 𝑖 ≀ 𝑁 .
ℱ𝑑𝑠 , taking 𝑃 -expectations yields
¯
𝐸 𝑃 [𝑉𝑑𝑠,¯πœ” 1𝐴𝑁 ] ≀ 𝐸 𝑃𝑁 [πœ‰ 𝑠,¯πœ” 1𝐴𝑁 ] + 𝜌(πœ€),
𝑃¯π‘ = 𝑃¯ to recall the dependence on 𝑁 .
𝑐
¯
𝑃𝑁 (𝐴𝑁 ) = 𝑃 (𝐴𝑐𝑁 ) β†’ 0 as 𝑁 β†’ ∞. In view of
where we write
have
¯
¯
¯
¯
As
𝑃 = 𝑃¯
(3.18)
Since
𝐴𝑁 ↑ Ξ© 𝑠 ,
we
𝐸 𝑃𝑁 [πœ‰ 𝑠,¯πœ” 1𝐴𝑁 ] = 𝐸 𝑃𝑁 [πœ‰ 𝑠,¯πœ” ] βˆ’ 𝐸 𝑃𝑁 [πœ‰ 𝑠,¯πœ” 1𝐴𝑐𝑁 ] ≀ 𝐸 𝑃𝑁 [πœ‰ 𝑠,¯πœ” ] + βˆ₯πœ‰βˆ₯∞ 𝑃𝑁 (𝐴𝑐𝑁 ),
15
we conclude from (3.18) that
¯
𝐸 𝑃 [𝑉𝑑𝑠,¯πœ” ] ≀ lim sup 𝐸 𝑃𝑁 [πœ‰ 𝑠,¯πœ” ] + 𝜌(πœ€) ≀
𝑁 β†’βˆž
Since
β€²
𝐸 𝑃 [πœ‰ 𝑠,¯πœ” ] + 𝜌(πœ€).
sup
𝑃 β€² βˆˆπ’«(𝑠,¯
πœ”)
𝑃 ∈ 𝒫(𝑠, πœ”
¯ ) was arbitrary, letting πœ€ β†’ 0 completes the proof of
(3.11).
It remains to argue that our assumption (3.13) does not entail a loss of
˜ 𝑖 for some 𝑖. Then there are two
πœ”
ˆ𝑖 ∈
/ 𝐸
𝑖
possible cases. The case 𝑃 (𝐸 ) = 0 is easily seen to be harmless; recall
𝑖
that the measure 𝑃 was xed throughout the proof. In the case 𝑃 (𝐸 ) > 0,
𝑖
𝑖
˜ ) > 0 and in particular 𝐸
˜ βˆ•= βˆ…. Thus we can replace
we also have 𝑃 (𝐸
𝑖
𝑖
˜
πœ”
Λ† by an arbitrary element of 𝐸 (which can be chosen independently of
𝑁 ). Using the continuity of the value function (Proposition 2.5) and of the
generality.
Indeed, assume that
reward function (2.12), we see that the above arguments still apply if we add
an additional modulus of continuity in (3.15).
4
Extension of the Value Function
In this section, we extend the value function
of random variables
𝑉𝑑
πœ‰,
πœ‰ 7β†’ 𝑉𝑑 (πœ‰; β‹…) to an 𝐿1 -type space
in the spirit of, e.g., [8]. While the construction of
in the previous section required a precise analysis πœ” by
πœ” ,
we can now
move towards a more probabilistic presentation. In particular, we shall often
write
𝑉𝑑 (πœ‰)
for the random variable
πœ” 7β†’ 𝑉𝑑 (πœ‰; πœ”).
For reasons explained in Remark 4.2 below, we x from now on an initial
condition
π‘₯ ∈ ℝ𝑑
and let
𝒫π‘₯ := {𝑃 (0, π‘₯, 𝜈) : 𝜈 ∈ 𝒰}
𝑠 = 0. Given a random variable
πœ“ on Ξ©, we write πœ“ π‘₯ as a shorthand for πœ“ 0,π‘₯ ≑ πœ“(π‘₯βŠ—0 β‹…). We also write 𝑉𝑑π‘₯ (πœ‰)
π‘₯
for (𝑉𝑑 (πœ‰)) .
𝑝
Given 𝑝 ∈ [1, ∞), we dene 𝐿𝒫 to be the space of ℱ𝑇 -measurable random
π‘₯
variables 𝑋 satisfying
be the corresponding set of measures at time
βˆ₯𝑋βˆ₯𝐿𝑝 := sup βˆ₯𝑋βˆ₯𝐿𝑝 (𝑃 ) < ∞,
𝒫π‘₯
𝑃 βˆˆπ’«π‘₯
βˆ₯𝑋βˆ₯𝑝𝐿𝑝 (𝑃 ) := 𝐸 𝑃 [βˆ£π‘‹βˆ£π‘ ]. More precisely, we identify functions which are
𝑝
equal 𝒫π‘₯ -quasi-surely, so that 𝐿𝒫 becomes a Banach space. (Two functions
π‘₯
are equal 𝒫π‘₯ -quasi-surely, 𝒫π‘₯ -q.s. for short, if they are equal 𝑃 -a.s. for all
𝑃 ∈ 𝒫π‘₯ .) Furthermore, given 𝑑 ∈ [0, 𝑇 ],
where
𝕃𝑝𝒫π‘₯ (ℱ𝑑 )
is dened as the
βˆ₯ β‹… βˆ₯𝐿𝑝
𝒫π‘₯
-closure of
UC𝑏 (Ω𝑑 ) βŠ† 𝐿𝑝𝒫π‘₯ .
𝐿𝑝𝒫π‘₯ -convergent sequence has a 𝒫π‘₯ -q.s. convergent subsequence,
𝑝
any element of 𝕃𝒫 (ℱ𝑑 ) has an ℱ𝑑 -measurable representative. For brevity,
π‘₯
𝑝
𝑝
we shall often write 𝕃𝒫 for 𝕃𝒫 (ℱ𝑇 ).
π‘₯
π‘₯
Since any
16
Remark 4.1.
𝐿𝑝𝒫π‘₯ is
πœ‰ ∈
The space
𝒫π‘₯ -quasi
𝕃𝑝𝒫π‘₯
can be described as follows.
πœ‰
uniformly continuous if
We say that
πœ‰β€²
has a representative
with
πœ€ > 0 there exists an open set 𝐺 βŠ† Ξ© such that
𝑃 (𝐺) < πœ€ for all 𝑃 ∈ 𝒫 and such that the restriction πœ‰ β€² βˆ£Ξ©βˆ–πΊ is uniformly con𝑝
𝑝
tinuous. Then 𝕃𝒫 consists of all πœ‰ ∈ 𝐿𝒫 such that πœ‰ is 𝒫π‘₯ -quasi uniformly
π‘₯
π‘₯
continuous and limπ‘›β†’βˆž βˆ₯πœ‰1{βˆ£πœ‰βˆ£β‰₯𝑛} βˆ₯𝐿𝑝 = 0. Moreover, If 𝒫π‘₯ is weakly rela𝒫π‘₯
𝑝
tively compact, then 𝕃𝒫 contains all bounded continuous functions on Ξ©.
π‘₯
the property that for all
The proof is the same as in [17, Proposition 5.2], which, in turn, followed
an argument of [7].
𝕃1𝒫π‘₯ ,
π‘₯ ∈ ℝ𝑑 .
Before extending the value function to
working under a xed initial condition
Remark 4.2.
let us explain why we are
There is no fundamental obstruction to writing the theory
without xing the initial condition
be more elegant if stated using
𝒫¯
π‘₯;
in fact, most of the results would
instead of
𝒫π‘₯ ,
where
𝒫¯
is the set of all
distributions of the form (2.4), with arbitrary initial condition. However, the
set
𝒫¯
is very large and therefore the corresponding space
𝕃1𝒫¯
is very small,
which is undesirable for the domain of our extended value function. As an
illustration, consider a random variable of the form
where
𝑓 : ℝ𝑑 β†’ ℝ
πœ‰(πœ”) := 𝑓 (πœ”0 )
on
Ξ©,
is a measurable function. Then
βˆ₯πœ‰βˆ₯𝐿1¯ = sup βˆ£π‘“ (π‘₯)∣;
𝒫
i.e.,
πœ‰
is in
𝐿1𝒫¯
only when
𝑓
π‘₯βˆˆβ„π‘‘
is uniformly bounded. As a second issue in the
same vein, it follows from the Arzelà-Ascoli theorem that the set
𝒫¯
is never
weakly relatively compact. The latter property, which is satised by
example when
πœ‡
and
𝜎
𝒫π‘₯
for
are bounded, is sometimes useful in the context of
quasi-sure analysis.
Lemma 4.3. Let 𝑝 ∈ [1, ∞). The mapping 𝑉𝑑π‘₯ on UC𝑏 (Ξ©) is 1-Lipschitz,
βˆ₯𝑉𝑑π‘₯ (πœ‰) βˆ’ 𝑉𝑑π‘₯ (πœ“)βˆ₯𝐿𝑝 ≀ βˆ₯πœ‰ βˆ’ πœ“βˆ₯𝐿𝑝
𝒫π‘₯
𝒫π‘₯
for all πœ‰, πœ“ ∈ UC𝑏 (Ξ©).
As a consequence, 𝑉𝑑π‘₯ uniquely extends to a Lipschitz-continuous mapping
𝑉𝑑π‘₯ : 𝕃𝑝𝒫π‘₯ (ℱ𝑇 ) β†’ 𝕃𝑝𝒫π‘₯ (ℱ𝑑 ).
Proof.
The argument is standard and included only for completeness. Note
βˆ£πœ‰ βˆ’πœ“βˆ£π‘ is again in UC𝑏 (Ξ©). The denition of 𝑉𝑑π‘₯ and Jensen's inequality
π‘₯
π‘₯
𝑝
π‘₯
𝑝
π‘₯
𝑝
imply that βˆ£π‘‰π‘‘ (πœ‰) βˆ’ 𝑉𝑑 (πœ“)∣ ≀ 𝑉𝑑 (βˆ£πœ‰ βˆ’ πœ“βˆ£) ≀ 𝑉𝑑 (βˆ£πœ‰ βˆ’ πœ“βˆ£ ). Therefore,
[
]1/𝑝
βˆ₯𝑉𝑑π‘₯ (πœ‰) βˆ’ 𝑉𝑑π‘₯ (πœ“)βˆ₯𝐿𝑝 ≀ sup 𝐸 𝑃 𝑉𝑑π‘₯ (βˆ£πœ‰ βˆ’ πœ“βˆ£π‘ )
= sup 𝐸 𝑃 [βˆ£πœ‰ βˆ’ πœ“βˆ£π‘ ]1/𝑝 ,
that
𝒫π‘₯
𝑃 βˆˆπ’«π‘₯
𝑃 βˆˆπ’«π‘₯
where the equality is due to (3.2) applied with
𝑠 = 0.
(For the case
𝑠 = 0, the
additional Assumption 3.1 was not used in the previous section.) Recalling
𝑉𝑑π‘₯ maps
𝑝
𝕃𝒫π‘₯ (ℱ𝑑 ).
from Proposition 2.5 that
𝑝
extension maps 𝕃𝒫 to
π‘₯
UC𝑏 (Ξ©)
17
to
UC𝑏 (Ω𝑑 ),
it follows that the
4.1
Quasi-Sure Properties of the Extension
In this section, we provide some auxiliary results of technical nature. The
rst one will (quasi-surely) allow us to appeal to the results in the previous
section without imposing Assumption 3.1. This is desirable since we would
like to end up with quasi-sure theorems whose statements do not involve
regular conditional probability distributions.
Lemma 4.4. Assumption 2.1 implies that Assumption 3.1 holds for
quasi-every πœ‚ ∈ Ξ© satisfying πœ‚0 = π‘₯.
𝒫π‘₯ -
For the proof of this lemma, we shall use the following result.
Lemma 4.5. Let
π‘Œ and 𝑍 be continuous adapted processes, 𝑑 ∈ [0, 𝑇 ] and
let 𝑃 be a probability measure on Ξ©. Then π”½π‘Œ
π”½π‘Œ 𝑑,πœ”
Proof.
𝑃 𝑑,πœ”
βŠ‡ 𝔽𝑍
𝑑,πœ”
𝑃
βŠ‡ 𝔽𝑍 implies that
for 𝑃 -a.e. πœ” ∈ Ξ©.
The assumption implies that there exists a progressively measurable
transformation
it follows that
𝛽 : Ξ© β†’ Ξ© such that 𝑍 = 𝛽(π‘Œ ) 𝑃 -a.s. For 𝑃 -a.e. πœ” ∈ Ξ©,
𝑍(πœ” βŠ—π‘‘ β‹…) = 𝛽(π‘Œ (πœ” βŠ—π‘‘ β‹…)) 𝑃 𝑑,πœ” -a.s., which, in turn, yields the
result.
Proof of Lemma 4.4.
Let
𝑋 := 𝑋(𝑑, πœ‚, 𝜈)
with
πœ‚ 0 = π‘₯,
we have to show that
𝑃0𝑑
βŠ‡ 𝔽𝑑 whenever πœ‚ 0 is outside some 𝒫π‘₯ -polar set. Hence we shall x an
Λ† ∈ 𝒫π‘₯ and show that the result holds on a set of full measure 𝑃ˆ .
arbitrary 𝑃
Λ† ∈ 𝒫π‘₯ , then 𝑃ˆ = 𝑃 (0, π‘₯, πœˆΛ†) for some πœˆΛ† ∈ 𝒰 and 𝑃ˆ is concentrated on
Let 𝑃
Λ† 0 , where 𝑋
Λ† := 𝑋(0, π‘₯, πœˆΛ†). That is, recalling that πœ‚0 = 𝑋
Λ† 0 = π‘₯,
the image of 𝑋
Λ†
we may assume that πœ‚ = 𝑋(πœ”) for some πœ” ∈ Ξ©. Let
𝔽𝑋
𝜈¯(˜
πœ” ) := 1[0,𝑑) πœˆΛ†(˜
πœ” ) + 1[𝑑,𝑇 ] 𝜈(˜
πœ” 𝑑 ).
(4.1)
¯ := 𝑋(0, π‘₯, 𝜈¯) satises 𝑋
¯ =𝑋
Λ† on [0, 𝑑] and hence we may assume
𝑋
¯
πœ‚ = 𝑋(πœ”)
on [0, 𝑑]. Using Lemma 3.3 and (4.1), we deduce that
(
)
¯ 𝑑,πœ” = 𝑋 𝑑, π‘₯ βŠ— 𝑋(πœ”),
¯
𝑋
𝜈¯π‘‘,πœ” = 𝑋(𝑑, πœ‚, 𝜈) = 𝑋.
Then
that
Since
𝔽 βŠ† 𝔽𝑋¯
𝑃0
by Assumption 2.1, we conclude that
𝔽𝑑
by using Lemma 4.5 with
𝑍
𝑃0𝑑
βŠ† 𝔽𝑋¯ 𝑑,πœ”
𝑃0𝑑
πœ‰ 7β†’ 𝑉𝑑π‘₯ (πœ‰)
on
𝕃1𝒫π‘₯
𝑃0𝑑
being the canonical process.
The next two results show that (for
mapping
= 𝔽𝑋
πœ‡β‰‘0
and
𝜎
positive denite) the
falls into the general class of sublinear expecta-
tions considered in [19], whose techniques we shall apply in the subsequent
18
section. More precisely, the two lemmas below yield the validity of its main
condition [19, Assumption 4.1].
The following property is known as
stability under pasting and well known
to be important in non-Markovian control. It should not be confused with the
pasting discussed in Remark 3.6, where the considered measures correspond
to dierent points in time.
Lemma 4.6. Let 𝜏 be an 𝔽-stopping time and let Ξ› ∈ β„±πœ . Let 𝑃, 𝑃 1 , 𝑃 2 ∈ 𝒫π‘₯
satisfy 𝑃 1 = 𝑃 2 = 𝑃 on β„±πœ . Then
[
]
𝑃¯ (𝐴) := 𝐸 𝑃 𝑃 1 (π΄βˆ£β„±πœ )1Ξ› + 𝑃 2 (π΄βˆ£β„±πœ )1Λ𝑐 ,
𝐴 ∈ ℱ𝑇
denes an element of 𝒫π‘₯ .
Proof.
𝑃¯
It follows from the denition of the conditional expectation that
is
a probability measure which is characterized by the properties
𝑃¯ = 𝑃
on
β„±πœ
and
¯ 𝜏 (πœ”),πœ”
𝑃
{
(𝑃 1 )𝜏 (πœ”),πœ”
=
(𝑃 2 )𝜏 (πœ”),πœ”
for
for
𝑃 -a.e. πœ” ∈ Ξ›,
𝑃 -a.e. πœ” ∈ Λ𝑐 .
(4.2)
𝜈, 𝜈 1 , 𝜈 2 ∈ 𝒰 be such that 𝑃 = 𝑃 (0, π‘₯, 𝜈) and 𝑃 𝑖 = 𝑃 (0, π‘₯, 𝜈 𝑖 ) for 𝑖 = 1, 2.
Moreover, let 𝑋 := 𝑋(0, π‘₯, 𝜈), dene 𝜈
¯ ∈ 𝒰 by
Let
𝜈¯π‘Ÿ (πœ”) := 1[0,𝜏 (𝑋(πœ”)0 )) (π‘Ÿ)πœˆπ‘Ÿ (πœ”)
[
]
+ 1[𝜏 (𝑋(πœ”)0 ),𝑇 ] (π‘Ÿ) πœˆπ‘Ÿ1 (πœ”)1Ξ› (𝑋(πœ”)0 ) + πœˆπ‘Ÿ2 (πœ”)1Λ𝑐 (𝑋(πœ”)0 )
and let
π‘ƒβˆ— := 𝑃 (0, π‘₯, 𝜈¯) ∈ 𝒫π‘₯ . We show that π‘ƒβˆ— satises the three prop𝜈 = 𝜈¯ on [0, 𝜏 (𝑋 0 )) implies that π‘ƒβˆ— = 𝑃 on β„±πœ .
erties from (4.2). Indeed,
Moreover, as in (3.8),
𝜏 (πœ”),πœ”
π‘ƒβˆ—
)
(
0
= 𝑃 𝜏 (πœ”), π‘₯ βŠ—0 πœ”, 𝜈¯πœ (𝑋(πœ”) ),Λ‡πœ”
for
𝑃 -a.e. πœ” ∈ Ξ©.
Similarly as below (3.9), we also have that
0 ),Λ‡
πœ”
𝜈¯πœ (𝑋(πœ”)
(
)
0
= 𝜈1 πœ”
Λ‡ βŠ—πœ (𝑋(πœ”)0 ) β‹… = (𝜈 1 )𝜏 (𝑋(πœ”) ),Λ‡πœ”
for
𝑃 -a.e. πœ” ∈ Ξ›.
Therefore,
𝜏 (πœ”),πœ”
π‘ƒβˆ—
(
)
0
= 𝑃 𝜏 (πœ”), π‘₯ βŠ—0 πœ”, (𝜈 1 )𝜏 (𝑋(πœ”) ),Λ‡πœ” = (𝑃 1 )𝜏 (πœ”),πœ”
for
𝑃 -a.e. πœ” ∈ Ξ›.
An analogous argument establishes the third property from (4.2) and we
conclude that
𝑃¯ = π‘ƒβˆ— ∈ 𝒫π‘₯ .
The second property is the quasi-sure representation of
result which will be generalized in Theorem 5.2 below.
19
𝑉𝑑π‘₯ (πœ‰)
on
𝕃1𝒫π‘₯ ,
a
Lemma 4.7. Let 𝑑 ∈ [0, 𝑇 ] and πœ‰ ∈ 𝕃1𝒫π‘₯ . Then
β€²
𝑃 -a.s.
𝑉𝑑π‘₯ (πœ‰) = ess sup𝑃 𝐸 𝑃 [πœ‰ π‘₯ βˆ£β„±π‘‘ ]
𝑃 β€² βˆˆπ’«
for all 𝑃 ∈ 𝒫π‘₯ ,
(4.3)
π‘₯ (ℱ𝑑 ,𝑃 )
where 𝒫π‘₯ (ℱ𝑑 , 𝑃 ) := {𝑃 β€² ∈ 𝒫π‘₯ : 𝑃 β€² = 𝑃 on ℱ𝑑 }.
Proof.
Recall that Lemma 4.4 allows us to appeal to the results of Section 3.
(i) We rst prove the inequality ≀ for
πœ‰ ∈ UC𝑏 (Ξ©).
Fix
𝑃 ∈ 𝒫π‘₯ .
We
use Step (ii) of the proof of Theorem 3.2, in particular (3.17), for the special
case 𝑠 = 0 and obtain that for given πœ€ > 0 and 𝑁 β‰₯ 1 there exists a measure
𝑃¯π‘ ∈ 𝒫π‘₯ (ℱ𝑑 , 𝑃 ) such that
¯
𝑉𝑑π‘₯ (πœ”) ≀ 𝐸 𝑃𝑁 [πœ‰ π‘₯ βˆ£β„±π‘‘ ](πœ”) + 𝜌(πœ€) for 𝑃 -a.s. πœ” ∈ 𝐸 1 βˆͺ β‹… β‹… β‹… βˆͺ 𝐸 𝑁 ,
βˆͺ
π‘₯
π‘₯
𝑖
0
where 𝑉𝑑 (πœ”) := 𝑉𝑑 (πœ‰; πœ”). Since
𝑖β‰₯1 𝐸 = Ξ© = Ξ© 𝑃 -a.s., we deduce that
¯
𝑉𝑑π‘₯ (πœ”) ≀ sup 𝐸 𝑃𝑁 [πœ‰ π‘₯ βˆ£β„±π‘‘ ](πœ”) + 𝜌(πœ€)
for
𝑃 -a.s. πœ” ∈ Ξ©.
𝑁 β‰₯1
The claim follows by letting
πœ€ β†’ 0.
(ii) Next, we show the inequality β‰₯ in (4.3) for
𝑃, 𝑃 β€²
∈
πœ‰ ∈ UC𝑏 (Ξ©). Fix
∈ 𝒫(𝑑, π‘₯ βŠ—0 πœ”) for 𝑃 β€² -a.s. πœ” ∈ Ξ© by
applied with 𝑠 := 𝑑 and 𝑑 := 𝑇 yields that
𝒫π‘₯ and recall that (𝑃 β€² )𝑑,πœ”
Lemma 3.4. Therefore, (3.2)
𝑉𝑑π‘₯ (πœ”) = 𝑉𝑑 (π‘₯ βŠ—0 πœ”) β‰₯ 𝐸 (𝑃
β€² )𝑑,πœ”
[πœ‰ 𝑑,π‘₯βŠ—0 πœ” ] = 𝐸 (𝑃
β€² )𝑑,πœ”
β€²
[(πœ‰ π‘₯ )𝑑,πœ” ] = 𝐸 𝑃 [πœ‰ π‘₯ βˆ£β„±π‘  ](πœ”)
𝑃 β€² -a.s. on ℱ𝑑 . If 𝑃 β€² ∈ 𝒫π‘₯ (ℱ𝑑 , 𝑃 ), then 𝑃 β€² = 𝑃 on ℱ𝑑 and the inequality holds
β€²
also 𝑃 -a.s. The claim follows as 𝑃 ∈ 𝒫π‘₯ (ℱ𝑑 , 𝑃 ) was arbitrary.
(iii) So far, we have proved the result for πœ‰ ∈ UC𝑏 (Ξ©). The general
1
case πœ‰ ∈ 𝕃𝒫 can be derived by an approximation argument exploiting the
π‘₯
stability under pasting (Lemma 4.6). We omit the details since the proof is
exactly the same as in [17, Theorem 5.4].
5
Path Regularity for the Value Process
𝒫π‘₯ -modication for 𝑉 π‘₯ (πœ‰); that is, a
π‘₯
π‘₯
π‘Œπ‘‘ = 𝑉𝑑 (πœ‰) 𝒫π‘₯ -q.s. for all 𝑑 ∈ [0, 𝑇 ]. (Recall
ℝ𝑑 has been xed.) To this end, we extend the
In this section, we construct a càdlàg
càdlàg process
π‘Œ
π‘₯ such that
that the initial condition
π‘₯∈
𝔽+ = {ℱ𝑑+ }0≀𝑑≀𝑇 be the minimal right+
𝒫
continuous ltration containing 𝔽 and we augment 𝔽 by the collection 𝒩 π‘₯
of (𝒫π‘₯ , ℱ𝑇 )-polar sets to obtain the ltration
raw ltration
𝔽
as in [19]: we let
𝔾 = {𝒒𝑑 }0≀𝑑≀𝑇 ,
We note that
𝔾 depends on π‘₯ ∈ ℝ𝑑
𝒒𝑑 := ℱ𝑑+ ∨ 𝒩 𝒫π‘₯ .
since
𝒩 𝒫π‘₯
does, but for brevity, we shall
not indicate this in the notation. In fact, the dependence on
20
π‘₯ is not crucial:
we could also work with
is
𝒫π‘₯ -q.s.
𝔽+ , at the expense of obtaining a modication which
equal to a càdlàg process rather than being càdlàg itself.
We recall that in the quasi-sure setting, value processes similar to the one
under consideration do not admit càdlàg modications in general; indeed,
while the right limit exists quasi-surely, it need not be a modication (cf.
[19]). Both the regularity of
πœ‰ ∈ 𝕃1𝒫π‘₯
and the regularity induced by the SDE
are crucial for the following result.
Theorem 5.1. Let
πœ‰ ∈ 𝕃1𝒫π‘₯ . There exists a (𝒫π‘₯ -q.s. unique) 𝔾-adapted
càdlàg 𝒫π‘₯ -modication β„° π‘₯ (πœ‰) = {ℰ𝑑π‘₯ (πœ‰)}π‘‘βˆˆ[0,𝑇 ] of {𝑉𝑑π‘₯ (πœ‰)}π‘‘βˆˆ[0,𝑇 ] . Moreover,
β€²
𝑃 -a.s.
ℰ𝑑π‘₯ (πœ‰) = ess sup𝑃 𝐸 𝑃 [πœ‰ π‘₯ βˆ£π’’π‘‘ ]
𝑃 β€² βˆˆπ’«π‘₯ (𝒒𝑑 ,𝑃 )
for all 𝑃 ∈ 𝒫π‘₯ ,
(5.1)
for all 𝑑 ∈ [0, 𝑇 ].
Proof.
In view of Lemmata 4.6 and 4.7, we obtain exactly as in [19, Propo-
sition 4.5] that there exists a
𝒫π‘₯ -q.s. unique 𝔾-adapted càdlàg process β„° π‘₯ (πœ‰)
satisfying (5.1) and
π‘₯
ℰ𝑑π‘₯ (πœ‰) = 𝑉𝑑+
(πœ‰) := lim π‘‰π‘Ÿπ‘₯ (πœ‰) 𝒫π‘₯ -q.s.
π‘Ÿβ†“π‘‘
for all
0 ≀ 𝑑 < 𝑇.
The observation made there is that (4.3) implies that
supermartingale for all
𝑃 ∈ 𝒫π‘₯ ,
𝑉 π‘₯ (πœ‰)
(5.2)
is a
(𝑃, 𝔽)-
so that one can use the standard modi-
cation argument for supermartingales under each
𝑃 π‘₯
[6, Theorem VI.2], also yields that 𝐸 [ℰ𝑑 (πœ‰)βˆ£β„±π‘‘+ ]
𝑃 . This argument, cf.
≀ 𝑉𝑑π‘₯ (πœ‰) 𝑃 -a.s. and in
particular
𝐸 𝑃 [ℰ𝑑π‘₯ (πœ‰)] ≀ 𝐸 𝑃 [𝑉𝑑π‘₯ (πœ‰)]
for all
𝑃 ∈ 𝒫π‘₯ .
Hence, it remains to show that
ℰ𝑑π‘₯ (πœ‰) β‰₯ 𝑉𝑑π‘₯ (πœ‰) 𝒫π‘₯ -q.s.
for
(5.3)
𝑑 ∈ [0, 𝑇 ), which is the part that is known to fail in a more general setting.
We give the proof in several steps.
𝕃1𝒫π‘₯ (ℱ𝑑 ), and in fact even to
UC𝑏 (Ω𝑑 ) if a suitable representative is chosen. Let πœ‰ ∈ UC𝑏 (Ξ©), π‘Ÿ ∈ (𝑑, 𝑇 ] and
π‘₯
π‘₯
set π‘‰π‘Ÿ := π‘‰π‘Ÿ (πœ‰). By Proposition 2.5, there exists a modulus of continuity 𝜌
independent of π‘Ÿ such that
(i) We rst show that
ℰ𝑑π‘₯
maps
UC𝑏 (Ξ©)
to
βˆ£π‘‰π‘Ÿπ‘₯ (πœ”) βˆ’ π‘‰π‘Ÿπ‘₯ (πœ” β€² )∣ ≀ 𝜌(βˆ₯πœ” βˆ’ πœ” β€² βˆ₯π‘Ÿ ).
Hence the
𝒫π‘₯ -q.s.
limit from (5.2) satises
βˆ£π‘‰π‘‘+ (πœ”) βˆ’ 𝑉𝑑+ (πœ” β€² )∣ ≀ 𝜌(βˆ₯πœ” βˆ’ πœ” β€² βˆ₯π‘Ÿ )
Since
ℰ𝑑π‘₯ (πœ‰) = 𝑉𝑑+ (πœ‰),
taking the limit
for all
π‘Ÿβ†“π‘‘
π‘Ÿ ∈ (𝑑, 𝑇 ] ∩ β„š,
yields that
βˆ£β„°π‘‘π‘₯ (πœ‰)(πœ”) βˆ’ ℰ𝑑π‘₯ (πœ‰)(πœ” β€² )∣ ≀ 𝜌(βˆ₯πœ” βˆ’ πœ” β€² βˆ₯𝑑 ) 𝒫π‘₯ -q.s.
21
𝒫π‘₯ -q.s.
ℰ𝑑π‘₯ (πœ‰) co∈ 𝕃1𝒫π‘₯ (ℱ𝑑 ).
∈ 𝕃1𝒫π‘₯ and
By a variant of Tietze's extension theorem, cf. [16], this implies that
π‘₯
incides 𝒫π‘₯ -q.s. with an element of UC𝑏 (Ω𝑑 ). In particular, ℰ𝑑 (πœ‰)
π‘₯
(ii) Next, we show that ℰ𝑑 is Lipschitz-continuous. Let πœ‰, πœ“
𝑑𝑛 ↓ 𝑑.
Using (5.2), Fatou's lemma and Lemma 4.3, we obtain that
βˆ₯ℰ𝑑π‘₯ (πœ‰) βˆ’ ℰ𝑑π‘₯ (πœ“)βˆ₯𝐿1 = lim𝑛 βˆ£π‘‰π‘‘π‘₯𝑛 (πœ‰) βˆ’ 𝑉𝑑π‘₯𝑛 (πœ“)∣𝐿1
𝒫π‘₯
𝒫π‘₯
[
]
𝑃
π‘₯
= sup 𝐸 lim𝑛 βˆ£π‘‰π‘‘π‘› (πœ‰) βˆ’ 𝑉𝑑π‘₯𝑛 (πœ“)∣
𝑃 βˆˆπ’«π‘₯
[
]
≀ sup lim inf 𝐸 𝑃 βˆ£π‘‰π‘‘π‘₯𝑛 (πœ‰) βˆ’ 𝑉𝑑π‘₯𝑛 (πœ“)∣
𝑛
𝑃 βˆˆπ’«π‘₯
≀ βˆ₯πœ‰ βˆ’ πœ“βˆ₯𝐿1 .
𝒫π‘₯
1
𝑛
𝑛
(iii) Let πœ‰ ∈ 𝕃𝒫 . Then there exist πœ‰ ∈ UC𝑏 (Ξ©) such that πœ‰
π‘₯
π‘₯
𝑛
π‘₯
1
π‘₯
𝑛
and in thus ℰ𝑑 (πœ‰ ) β†’ ℰ𝑑 (πœ‰) in 𝐿𝒫 by Step (ii). Since ℰ𝑑 (πœ‰ ) ∈
π‘₯
1
1
Step (i) and since 𝕃𝒫 (ℱ𝑑 ) is closed in 𝐿𝒫 , we conclude that
π‘₯
π‘₯
ℰ𝑑π‘₯ (πœ‰) ∈ 𝕃1𝒫π‘₯ (ℱ𝑑 )
for all
β†’ πœ‰ in 𝐿1𝒫π‘₯
𝕃1𝒫π‘₯ (ℱ𝑑 ) by
πœ‰ ∈ 𝕃1𝒫π‘₯ .
πœ‰ ∈ 𝕃1𝒫π‘₯ . Since 𝑉𝑑π‘₯ is the identity on 𝕃1𝒫π‘₯ (ℱ𝑑 ), Step (iii) implies
= 𝑉𝑑π‘₯ (ℰ𝑑π‘₯ (πœ‰)). Moreover, the representations (4.3) and (5.1) yield
(iv) Let
π‘₯
that ℰ𝑑 (πœ‰)
β€²
𝑉𝑑π‘₯ (ℰ𝑑π‘₯ (πœ‰)) = ess sup𝑃 𝐸 𝑃 [ℰ𝑑π‘₯ (πœ‰)βˆ£β„±π‘‘ ]
𝑃 β€² βˆˆπ’«
π‘₯ (ℱ𝑑 ,𝑃 )
]
β€²[
β€²
β‰₯ ess sup𝑃 𝐸 𝑃 𝐸 𝑃 [πœ‰ π‘₯ βˆ£π’’π‘‘ ]ℱ𝑑
𝑃 β€² βˆˆπ’«π‘₯ (ℱ𝑑 ,𝑃 )
β€²
= ess sup𝑃 𝐸 𝑃 [πœ‰ π‘₯ βˆ£β„±π‘‘ ]
𝑃 β€² βˆˆπ’«π‘₯ (ℱ𝑑 ,𝑃 )
= 𝑉𝑑π‘₯ (πœ‰) 𝑃 -a.s.
for all
𝑃 ∈ 𝒫π‘₯ .
We conclude that (5.3) holds true.
Since
dened.
β„° π‘₯ (πœ‰) is a càdlàg process, its value β„°πœπ‘₯ (πœ‰) at a stopping time 𝜏
The following result states the quasi-sure representation of
is well
β„°πœπ‘₯ (πœ‰)
and the quasi-sure version of the dynamic programming principle in its nal
form.
Theorem 5.2. Let 0 ≀ 𝜚 ≀ 𝜏
≀ 𝑇 be 𝔾-stopping times and πœ‰ ∈ 𝕃1𝒫π‘₯ . Then
β€²
β„°πœšπ‘₯ (πœ‰) = ess sup𝑃 𝐸 𝑃 [β„°πœπ‘₯ (πœ‰)βˆ£π’’πœš ]
𝑃 β€² βˆˆπ’«π‘₯ (π’’πœš ,𝑃 )
𝑃 -a.s.
for all 𝑃 ∈ 𝒫π‘₯
𝑃 -a.s.
for all 𝑃 ∈ 𝒫π‘₯ .
(5.4)
and in particular
β€²
β„°πœšπ‘₯ (πœ‰) = ess sup𝑃 𝐸 𝑃 [πœ‰ π‘₯ βˆ£π’’πœš ]
𝑃 β€² βˆˆπ’«π‘₯ (π’’πœš ,𝑃 )
Moreover, there exists for each 𝑃 ∈ 𝒫π‘₯ a sequence 𝑃𝑛 ∈ 𝒫π‘₯ (π’’πœš , 𝑃 ) such that
β„°πœšπ‘₯ (πœ‰) = lim 𝐸 𝑃𝑛 [πœ‰ π‘₯ βˆ£π’’πœš ]
π‘›β†’βˆž
with a 𝑃 -a.s. increasing limit.
22
𝑃 -a.s.
Proof.
In view of Lemmata 4.6 and 4.7, the result is derived exactly as in
[19, Theorem 4.9].
As in (3.3), the relation (5.4) can be seen as a semigroup property
β„°πœšπ‘₯ (πœ‰) = β„°πœšπ‘₯ (β„°πœπ‘₯ (πœ‰)),
β„°πœπ‘₯ (πœ‰) is in the domain 𝕃1𝒫π‘₯ of β„°πœšπ‘₯ . The latter is guaranteed by
π‘₯
Lemma 4.3 when 𝜏 is a deterministic time. However, one cannot expect β„°πœ (πœ‰)
at least when
to be quasi uniformly continuous (cf. Remark 4.1) for a general stopping time,
for which reason we prefer to express the right hand side as in (5.4).
6
Hamilton-Jacobi-Bellman 2BSDE
In this section, we characterize the value process
β„° π‘₯ (πœ‰)
as the solution of
a 2BSDE. To this end, we rst examine the properties of
𝑃 ∈ 𝒫π‘₯ .
𝐡
under a xed
The following result is in the spirit of [28, Section 8].
Proposition 6.1. Let
π‘₯ ∈ ℝ𝑑 , 𝜈 ∈ 𝒰 and 𝑃 := 𝑃 (0, π‘₯, 𝜈). There exists a
progressively measurable transformation 𝛽 : Ξ© β†’ Ξ© (depending on π‘₯, 𝜈) such
that π‘Š := 𝛽(𝐡) is a 𝑃 -Brownian motion and
𝑃
𝑃
𝔽 = π”½π‘Š .
(6.1)
Moreover, 𝐡 is the 𝑃 -a.s. unique strong solution of the SDE
β‹…
∫
𝐡=
πœ‡(𝑑, π‘₯ + 𝐡, πœˆπ‘‘ (π‘Š )) 𝑑𝑑 +
0
Proof.
β‹…
∫
𝜎(𝑑, π‘₯ + 𝐡, πœˆπ‘‘ (π‘Š )) π‘‘π‘Šπ‘‘
under 𝑃.
0
𝑋 := 𝑋(0, π‘₯, 𝜈).
Let
As in Lemma 3.4, Assumption 2.1 implies the
the existence of a progressively measurable transformation
𝛽:Ξ©β†’Ξ©
such
that
𝛽(𝑋 0 ) = 𝐡
Let
π‘Š := 𝛽(𝐡).
𝑃0 -a.s.
(6.2)
Then
(𝐡, 𝑋 0 )𝑃0 = (𝛽(𝑋 0 ), 𝑋 0 )𝑃0 = (𝛽(𝐡), 𝐡)𝑃 = (π‘Š, 𝐡)𝑃 ;
i.e., the distribution of
(π‘Š, 𝐡)
under
0
have 𝔽𝑋
𝑃0
=
𝑃.
(𝐡, 𝑋 0 )
In particular,
𝔽𝛽(𝑋 0 )
𝑃0
under
π‘Š
is
𝑃0 coincides with the distribution of
a 𝑃 -Brownian motion. Moreover, we
by Assumption 2.1 and therefore
𝔽𝐡
𝑃
𝑃
= 𝔽𝛽(𝐡) ,
which is (6.1). Note that
0
0
𝑋 = 𝑋 (𝐡) =
∫
∫
0
πœ‡(𝑑, π‘₯ + 𝑋 , πœˆπ‘‘ (𝐡)) 𝑑𝑑 +
23
𝜎(𝑑, π‘₯ + 𝑋 0 , πœˆπ‘‘ (𝐡)) 𝑑𝐡𝑑
𝑃0 . Let π‘Œ be the (unique, strong) solution of the analogous SDE
∫
∫
π‘Œ = πœ‡(𝑑, π‘₯ + π‘Œ, πœˆπ‘‘ (π‘Š )) 𝑑𝑑 + 𝜎(𝑑, π‘₯ + π‘Œ, πœˆπ‘‘ (π‘Š )) π‘‘π‘Šπ‘‘ under 𝑃.
under
Using the denition of
𝑃
and (6.2), we have that
(π‘Œ, π‘Š )𝑃 = (𝑋 0 (π‘Š ), π‘Š )𝑃
= (𝑋 0 (𝐡), 𝐡)𝑃0
= (𝑋 0 , 𝛽(𝑋 0 ))𝑃0
= (𝐡, 𝛽(𝐡))𝑃
= (𝐡, π‘Š )𝑃 .
In view of (6.1), it follows that
In the sequel, we denote by
π‘Œ =𝐡
𝑀 𝐡,𝑃
𝑃 -a.s.
holds
the local martingale part in the canon-
ical semimartingale decomposition of
𝐡
under
𝑃.
Corollary 6.2. Let 𝑃
𝑃
∈ 𝒫π‘₯ . Then the ltration 𝔽 is right-continuous. If,
in addition, 𝜎 is invertible, then (𝑀 𝐡,𝑃 , 𝑃 ) has the predictable representation
property.
𝑃
(𝔽 , 𝑃 )-local marunder 𝑃 , for some
The latter statement means that any right-continuous
tingale
𝔽
𝑃
𝑁
has a representation
-predictable process
Proof.
𝑁 = 𝑁0 +
∫
𝑍 𝑑𝑀 𝐡,𝑃
𝑍.
We have seen in Proposition 6.1 that
𝑃
is generated by a Browβˆ«β‹…
ˆ𝑑 π‘‘π‘Šπ‘‘ for
π‘Š , hence right-continuous, and that 𝑀 𝐡,𝑃 = 0 𝜎
𝜎
ˆ𝑑 := 𝜎(𝑑, π‘₯ + 𝐡, πœˆπ‘‘ (π‘Š )), where 𝜈 ∈ 𝒰 . By changing 𝜎
Λ† on a 𝑑𝑑 × π‘ƒ -nullset,
𝑃
we may assume that 𝜎
is 𝔽 -predictable. Using the Brownian representation
∫ Λ† βˆ’1
theorem and π‘Š =
𝜎
Λ† 𝑑𝑀 𝐡,𝑃 , we deduce that 𝑀 𝐡,𝑃 has the representa-
𝔽
nian motion
tion property.
The following formulation of 2BSDE is, of course, inspired by [27].
Denition 6.3.
πœ‰ ∈ 𝐿1𝒫π‘₯ and consider a pair (π‘Œ, 𝑍) of processes with
values in ℝ ×
π‘Œ is càdlàg 𝔾-adapted while 𝑍 is 𝔾-predictable
βˆ«π‘‡
2
and
0 βˆ£π‘π‘  ∣ π‘‘βŸ¨π΅βŸ©π‘  < ∞ 𝒫π‘₯ -q.s. Then (π‘Œ, 𝑍) is called a solution of the
Let
ℝ𝑑 such that
2BSDE (6.3) if there exists a family
processes satisfying
∫
π‘Œπ‘‘ = πœ‰ βˆ’
𝐸 𝑃 [βˆ£πΎπ‘‡π‘ƒ ∣] < ∞
(𝐾 𝑃 )𝑃 βˆˆπ’«π‘₯
of
𝔽
𝑃
-adapted increasing
such that
𝑇
𝑍𝑠 𝑑𝑀𝑠𝐡,𝑃 + 𝐾𝑇𝑃 βˆ’ 𝐾𝑑𝑃 ,
0 ≀ 𝑑 ≀ 𝑇,
𝑃 -a.s.
for all
𝑃 ∈ 𝒫π‘₯
𝑑
(6.3)
and such that the following minimality condition holds for all
ess inf 𝑃 𝐸
𝑃 β€² βˆˆπ’«π‘₯ (𝒒𝑑 ,𝑃 )
𝑃′
[
𝑃′
𝐾𝑇
𝑃′
]
βˆ’ 𝐾𝑑 𝒒𝑑 = 0 𝑃 -a.s.
24
for all
0 ≀ 𝑑 ≀ 𝑇:
𝑃 ∈ 𝒫π‘₯ .
(6.4)
Moreover, a càdlàg process π‘Œ is said to be of class (D,𝒫π‘₯ ) if the family
{π‘Œπœ }𝜏 is uniformly integrable under 𝑃 for all 𝑃 ∈ 𝒫π‘₯ , where 𝜏 runs through
all 𝔾-stopping times. The following is our main result.
Theorem 6.4. Assume that 𝜎 is invertible and let πœ‰ ∈ 𝕃1𝒫π‘₯ .
(i) There exists a (𝑑𝑑 × π’«π‘₯ -q.s. unique) 𝔾-predictable process 𝑍 πœ‰ such that
(
)βˆ’1
𝑍 πœ‰ = π‘‘βŸ¨π΅, π΅βŸ©π‘ƒ
π‘‘βŸ¨β„° π‘₯ (πœ‰), π΅βŸ©π‘ƒ
𝑃 -a.s.
for all 𝑃 ∈ 𝒫π‘₯ .
(6.5)
(ii) The pair (β„° π‘₯ (πœ‰), 𝑍 πœ‰ ) is the minimal solution of the 2BSDE (6.3); i.e.,
if (π‘Œ, 𝑍) is another solution, then β„° π‘₯ (πœ‰) ≀ π‘Œ 𝒫π‘₯ -q.s.
(iii) If (π‘Œ, 𝑍) is a solution of (6.3) such that π‘Œ is of class (D,𝒫π‘₯ ), then
(π‘Œ, 𝑍) = (β„° π‘₯ (πœ‰), 𝑍 πœ‰ ).
In particular, if πœ‰ ∈ 𝕃𝑝𝒫π‘₯ for some 𝑝 > 1, then (β„° π‘₯ (πœ‰), 𝑍 πœ‰ ) is the unique
solution of (6.3) in the class (D,𝒫π‘₯ ).
Proof.
Given two processes which are (càdlàg) semimartingales under all
𝑃 ∈ 𝒫π‘₯ ,
their quadratic covariation can be dened
𝒫π‘₯ -q.s.
by using the
integration-by-parts formula and Bichteler's pathwise stochastic integration
[1, Theorem 7.14]; therefore, the right hand side of (6.5) can be used as a
𝑍 πœ‰ . The details of the argument are as in [19, Proposition 4.10].
Let 𝑃 ∈ 𝒫π‘₯ . By Proposition 6.1, 𝐡 is an Itô process under 𝑃 ; in par𝑃
𝐡,𝑃 , π‘†βŸ©π‘ƒ 𝑃 -a.s. for any 𝑃 -semimartingale 𝑆 .
ticular, we have ⟨𝐡, π‘†βŸ© = βŸ¨π‘€
The Doob-Meyer theorem under 𝑃 and Corollary 6.2 then yield the decomdenition of
position
β„° π‘₯ (πœ‰) = β„°0π‘₯ (πœ‰) +
∫
𝑍 πœ‰ 𝑑𝑀 𝐡,𝑃 βˆ’ 𝐾 𝑃
𝑃 -a.s.
and we obtain (ii) and (iii) by following the arguments in [19, Theorem 4.15].
If
πœ‰ ∈ 𝕃𝑝𝒫π‘₯
for some
𝑝 ∈ (1, ∞), then β„° π‘₯ (πœ‰) is of class (D,𝒫π‘₯ ) as a consequence
of Jensen's inequality (cf. [19, Lemma 4.14]). Therefore, the last assertion
follows from the above.
We conclude by interpreting the canonical process
set of scenarios
𝒫π‘₯ ,
𝐡,
seen under the
as a model for drift and volatility uncertainty in the
Knightian sense.
Remark 6.5.
Consider the set-valued process
D𝑑 (πœ”) :=
{(
)
}
πœ‡(𝑑, πœ”, 𝑒), 𝜎(𝑑, πœ”, 𝑒) : 𝑒 ∈ π‘ˆ βŠ† ℝ𝑑 × β„π‘‘×𝑑 .
𝑃 ∈ 𝒫π‘₯ can be seen as a scenario in
𝐡 ) take values in D, 𝑃 -a.s. Then, the
In view of Proposition 6.1, each
which
the drift and the volatility (of
upper
π‘₯
expectation β„° (πœ‰) is the corresponding worst-case expectation (see [19] for
a connection to superhedging in nance). Note that
D
is a random process
although the coecients of our controlled SDE are non-random.
25
Indeed,
the path-dependence of the SDE translates to an
πœ” -dependence
in the weak
formulation that we are considering.
πœ‡ ≑ 0, we have constructed a sublinear expectation
𝐺-expectation of [17]. While the latter is dened by
specifying a set-valued process like D in the rst place, we have started here
from a controlled SDE under 𝑃0 . It seems that the present construction
In particular, for
similar to the random
is somewhat less technical that the one in [17]; in particular, we did not
work with the process
and [17].
π‘Ž
Λ† = π‘‘βŸ¨π΅βŸ©π‘‘ /𝑑𝑑
which played an important role in [26]
However, it seems that the Lipschitz conditions on
πœ‡
and
𝜎
are
essential, while [17] merely used a notion of uniform continuity.
References
[1] K. Bichteler. Stochastic integration and 𝐿𝑝 -theory of semimartingales. Ann.
Probab., 9(1):4989, 1981.
[2] R. Buckdahn and J. Ma. Stochastic viscosity solutions for nonlinear stochastic
partial dierential equations. I. Stochastic Process. Appl. , 93(2):181204, 2001.
[3] R. Buckdahn and J. Ma. Stochastic viscosity solutions for nonlinear stochastic
partial dierential equations. II. Stochastic Process. Appl. , 93(2):205228,
2001.
[4] R. Buckdahn and J. Ma. Pathwise stochastic control problems and stochastic
HJB equations. SIAM J. Control Optim. , 45(6):22242256, 2007.
[5] P. Cheridito, H. M. Soner, N. Touzi, and N. Victoir. Second-order backward
stochastic dierential equations and fully nonlinear parabolic PDEs. Comm.
Pure Appl. Math., 60(7):10811110, 2007.
[6] C. Dellacherie and P. A. Meyer. Probabilities and Potential B . North Holland,
Amsterdam, 1982.
[7] L. Denis, M. Hu, and S. Peng. Function spaces and capacity related to a
sublinear expectation: application to 𝐺-Brownian motion paths. Potential
Anal., 34(2):139161, 2011.
[8] L. Denis and C. Martini. A theoretical framework for the pricing of contingent
claims in the presence of model uncertainty. Ann. Appl. Probab., 16(2):827
852, 2006.
[9] N. El Karoui. Les aspects probabilistes du contrôle stochastique. In Ecole
d'été de probabilitées de Saint-Flour , volume 876 of Lecture Notes in Math.,
pages 73238, Springer, Berlin, 1981.
[10] R. J. Elliott. Stochastic Calculus and Applications . Springer, New York, 1982.
[11] J. Feldman and M. Smorodinsky. Simple examples of non-generating Girsanov
processes. In Séminaire de Probabilités XXXI , volume 1655 of Lecture Notes
in Math., pages 247251. Springer, Berlin, 1997.
[12] W. H. Fleming and H. M. Soner. Controlled Markov Processes and Viscosity
Solutions. Springer, New York, 2nd edition, 2006.
[13] Y. Hu, P. Imkeller, and M. Müller. Utility maximization in incomplete markets. Ann. Appl. Probab., 15(3):16911712, 2005.
[14] P.-L. Lions and P. Souganidis. Fully nonlinear stochastic partial dierential
equations. C. R. Acad. Sci. Paris Sér. I Math. , 326(9):10851092, 1998.
[15] P.-L. Lions and P. Souganidis. Fully nonlinear stochastic partial dierential
equations: non-smooth equations and applications. C. R. Acad. Sci. Paris
Sér. I Math., 327(8):735741, 1998.
26
[16] M. Mandelkern. On the uniform continuity of Tietze extensions. Arch. Math.,
55(4):387388, 1990.
[17] M. Nutz. Random 𝐺-expectations. Preprint arXiv:1009.2168v1 , 2010.
[18] M. Nutz. The Bellman equation for power utility maximization with semimartingales. Ann. Appl. Probab., 22(1):363406, 2012.
[19] M. Nutz and H. M. Soner. Superhedging and dynamic risk measures under
volatility uncertainty. Preprint arXiv:1011.2958v1 , 2010.
[20] E. Pardoux and S. Peng. Adapted solution of a backward stochastic dierential
equation. Systems Control Lett., 14(1):5561, 1990.
[21] S. Peng. Stochastic Hamilton-Jacobi-Bellman equations. SIAM J. Control
Optim., 30(2):284304, 1992.
[22] S. Peng. Filtration consistent nonlinear expectations and evaluations of contingent claims. Acta Math. Appl. Sin. Engl. Ser. , 20(2):191214, 2004.
[23] S. Peng. 𝐺-expectation, 𝐺-Brownian motion and related stochastic calculus
of Itô type. In Stochastic Analysis and Applications , volume 2 of Abel Symp.,
pages 541567, Springer, Berlin, 2007.
[24] S. Peng. Multi-dimensional 𝐺-Brownian motion and related stochastic calculus
under 𝐺-expectation. Stochastic Process. Appl. , 118(12):22232253, 2008.
[25] S. Peng. Note on viscosity solution of path-dependent PDE and 𝐺-martingales.
Preprint arXiv:1106.1144v1 , 2011.
[26] H. M. Soner, N. Touzi, and J. Zhang. Dual formulation of second order target
problems. To appear in Ann. Appl. Probab. , 2010.
[27] H. M. Soner, N. Touzi, and J. Zhang. Wellposedness of second order backward
SDEs. To appear in Probab. Theory Related Fields , 2010.
[28] H. M. Soner, N. Touzi, and J. Zhang. Quasi-sure stochastic analysis through
aggregation. Electron. J. Probab., 16(2):18441879, 2011.
[29] D. Stroock and S. R. S. Varadhan. Multidimensional Diusion Processes .
Springer, New York, 1979.
27