PDF

The Bellman Equation for Power Utility
Maximization with Semimartingales
Marcel Nutz
ETH Zurich, Department of Mathematics, 8092 Zurich, Switzerland
[email protected]
First Version: December 9, 2009. This Version: March 5, 2011.
Abstract
We study utility maximization for power utility random elds with
and without intermediate consumption in a general semimartingale
model with closed portfolio constraints. We show that any optimal
strategy leads to a solution of the corresponding Bellman equation.
The optimal strategies are described pointwise in terms of the opportunity process, which is characterized as the minimal solution of the
Bellman equation. We also give verication theorems for this equation.
Keywords power utility, Bellman equation, opportunity process, semimartingale
characteristics, BSDE.
AMS 2000 Subject Classications Primary 91B28; secondary 93E20, 60G44.
JEL Classication G11, C61.
Acknowledgements.
Financial support by Swiss National Science Founda-
tion Grant PDFM2-120424/1 is gratefully acknowledged. The author thanks
Christoph Czichowsky for fruitful discussions and Martin Schweizer, Nicholas
Westray and an anonymous referee for comments on an earlier version of the
manuscript.
1 Introduction
A classical problem of mathematical nance is the maximization of expected
utility obtained from consumption or from terminal wealth. This paper focuses on power utility functions and presents the corresponding dynamic
programming in a general constrained semimartingale framework. The homogeneity of these utility functions leads to a factorization of the value
process into a part depending on the current wealth and the so-called oppor-
𝐿. In our setting, the Bellman equation describes the drift
𝐿 and claries the local structure of our problem. Finding an optimal
𝑑
strategy boils down to maximizing a random function 𝑦 7β†’ 𝑔(πœ”, 𝑑, 𝑦) on ℝ
for every state πœ” and date 𝑑. This function is given in terms of the semimartingale characteristics of 𝐿 as well as the asset returns, and its maximum
tunity process
rate of
1
yields the drift rate of
𝐿.
The role of the opportunity process is to augment
the information contained in the return characteristics in order to have a
local sucient statistic for the global optimization problem.
We present three main results.
First, we show that if there exists an
optimal strategy for the utility maximization problem,
cess 𝐿 solves the Bellman equation
the opportunity pro-
and we provide a local description of the
optimal strategies. We state the Bellman equation in two forms, as an identity for the drift rate of
(BSDE) for
𝐿.
𝐿
and as a backward stochastic dierential equation
Second, we characterize the opportunity process as the
mal solution of this equation.
mini-
Finally, given some solution and an associated
strategy, one can ask whether the strategy is optimal and the solution is
the opportunity process. We present two dierent approaches which lead to
verication theorems
not comparable in strength unless the constraints are
convex.
The present dynamic programming approach should be seen as complementary to convex duality, which remains the only method to obtain
tence
exis-
of optimal strategies in general models; see Kramkov and Schacher-
mayer [21], Karatzas and šitkovi¢ [20], Karatzas and Kardaras [19]. However, convex duality alone oers limited insight into the optimal strategies
for incomplete markets. In some cases, the Bellman equation can be solved
directly by analytic methods; e.g., in the setting of Example 5.8 with continuous asset prices or in the Lévy process setting of Nutz [25]. In addition
to the existence, one then obtains a way to compute the optimal strategies
(at least numerically) and study their properties.
This paper is organized as follows. The next section species the optimization problem in detail, recalls the opportunity process and the martingale optimality principle, and xes the notation for the characteristics. We
also introduce set-valued processes describing the budget condition and state
the assumptions on the portfolio constraints. Section 3 derives the Bellman
equation, rst as a drift condition and then as a BSDE. It becomes more explicit as we specialize to the case of continuous asset prices. The denition of
a solution of the Bellman equation is given in Section 4, where we show the
minimality of the opportunity process. Section 5 deals with the verication
problem, which is converse to the derivation of the Bellman equation since
it requires the passage from the local maximization to the global optimization problem. We present an approach via the value process and a second
approach via a deator, which corresponds to the dual problem in a suitable setting. Appendix A belongs to Section 3 and contains the measurable
selections for the construction of the Bellman equation. It is complemented
by Appendix B, where we construct an alternative parametrization of the
market model by representative portfolios.
2
2 Preliminaries
π‘₯, 𝑦 ∈ ℝ, we denote π‘₯+ = max{π‘₯, 0}
𝑑
and π‘₯ ∧ 𝑦 = min{π‘₯, 𝑦}. We set 1/0 := ∞ where necessary. If 𝑧 ∈ ℝ
𝑖
⊀ its transpose, and
is a 𝑑-dimensional vector, 𝑧 is its 𝑖th coordinate, 𝑧
⊀
1/2
𝑑
βˆ£π‘§βˆ£ = (𝑧 𝑧)
the Euclidean norm. If 𝑋 is an ℝ -valued semimartingale
𝑑
and πœ‹ is an ℝ -valued predictable integrand, the vector stochastic integral is
∫
a scalar semimartingale with initial value zero and denoted by
πœ‹ 𝑑𝑋 or by
πœ‹ βˆ™ 𝑋 . The quadratic variation is the 𝑑×𝑑-matrix [𝑋] := [𝑋, 𝑋] and if π‘Œ is a
𝑖
𝑖
scalar semimartingale, [𝑋, π‘Œ ] is the 𝑑-vector with [𝑋, π‘Œ ] := [𝑋 , π‘Œ ]. When
The following notation is used.
If
the reference measure is understood, relations between measurable functions
hold almost everywhere unless otherwise mentioned. Our reference for any
unexplained notion from stochastic calculus is Jacod and Shiryaev [15].
2.1 The Optimization Problem
𝑇 ∈ (0, ∞) and a stochastic basis (Ξ©, β„±, 𝔽, 𝑃 ),
where the ltration 𝔽 = (ℱ𝑑 )π‘‘βˆˆ[0,𝑇 ] satises the usual assumptions of right
continuity and completeness as well as β„±0 = {βˆ…, Ξ©} 𝑃 -a.s. We consider an
ℝ𝑑 -valued càdlàg semimartingale 𝑅 with 𝑅0 = 0 representing the returns of 𝑑
We x the time horizon
risky assets. Their discounted prices are given by the stochastic exponential
𝑆 = β„°(𝑅) = (β„°(𝑅1 ), . . . , β„°(𝑅𝑑 )); in the nancial application, the components
of 𝑆 are assumed to be positive. Our agent also has a bank account at his
disposal; it does not pay interest.
π‘₯0
𝑑
is a predictable 𝑅-integrable ℝ -valued process
The agent is endowed with a deterministic initial capital
trading strategy
πœ‹π‘–
> 0. A
πœ‹ , where
indicates the fraction of wealth (or the portfolio proportion) invested
𝑖th risky asset.
∫ 𝑇 A consumption strategy is a nonnegative optional
𝑐 such that 0 𝑐𝑑 𝑑𝑑 < ∞ 𝑃 -a.s. We want to consider two cases.
Either consumption occurs only at the terminal time 𝑇 (utility from terminal
in the
process
wealth only); or there is intermediate consumption plus a bulk consumption
at the time horizon. To unify the notation, we introduce the measure
[0, 𝑇 ]
πœ‡
on
by
{
0
πœ‡(𝑑𝑑) :=
𝑑𝑑
in the case without intermediate consumption,
in the case with intermediate consumption.
πœ‡βˆ˜ := πœ‡+𝛿{𝑇 } , where 𝛿{𝑇 } is the unit Dirac measure at 𝑇 . The wealth
process 𝑋(πœ‹, 𝑐) corresponding to a pair (πœ‹, 𝑐) is dened by the equation
∫ 𝑑
∫ 𝑑
𝑋𝑑 (πœ‹, 𝑐) = π‘₯0 +
π‘‹π‘ βˆ’ (πœ‹, 𝑐)πœ‹π‘  𝑑𝑅𝑠 βˆ’
𝑐𝑠 πœ‡(𝑑𝑠), 0 ≀ 𝑑 ≀ 𝑇.
Let also
0
0
We dene the set of trading and consumption pairs
{
π’œ0 (π‘₯0 ) := (πœ‹, 𝑐) : 𝑋(πœ‹, 𝑐) > 0, π‘‹βˆ’ (πœ‹, 𝑐) > 0
3
and
}
𝑐𝑇 = 𝑋𝑇 (πœ‹, 𝑐) .
These are the strategies that satisfy the budget constraint. The convention
𝑐𝑇 = 𝑋𝑇 (πœ‹, 𝑐) means that all the remaining wealth is consumed at time
𝑇 . We consider also exogenous constraints imposed on the agent. For each
(πœ”, 𝑑) ∈ Ξ© × [0, 𝑇 ] we are given a set C𝑑 (πœ”) βŠ† ℝ𝑑 which contains the origin.
The set of (constrained)
admissible
strategies is
{
π’œ(π‘₯0 ) := (πœ‹, 𝑐) ∈ π’œ0 (π‘₯0 ) : πœ‹π‘‘ (πœ”) ∈ C𝑑 (πœ”) for
it is nonempty as
C
0 ∈ C𝑑 (πœ”).
all
}
(πœ”, 𝑑) ;
Further assumptions on the set-valued mapping
π‘₯0 and usually
π’œ for π’œ(π‘₯0 ). Abusing the notation, we write 𝑐 ∈ π’œ and call 𝑐 admissible
there exists πœ‹ such that (πœ‹, 𝑐) ∈ π’œ; an analogous convention is used for
will be introduced in Section 2.4. We x the initial capital
write
if
similar expressions.
We will often parametrize the consumption strategies as a fraction of
wealth. Let
(πœ‹, 𝑐) ∈ π’œ
and
𝑋 = 𝑋(πœ‹, 𝑐).
πœ… :=
is called the
Then
𝑐
𝑋
propensity to consume corresponding to (πœ‹, 𝑐).
This yields a one-
(πœ‹, 𝑐) ∈ π’œ and the pairs (πœ‹, πœ…) such
βˆ«π‘‡
πœ… is a nonnegative optional process satisfying 0 πœ…π‘  𝑑𝑠 < ∞
πœ…π‘‡ = 1 (see Nutz [26, Remark 2.1] for details). We shall abuse
to-one correspondence between the pairs
that
πœ‹βˆˆπ’œ
𝑃 -a.s.
and
and
the notation and identify a consumption strategy with the corresponding
(πœ‹, πœ…) ∈ π’œ. Note
(
)
𝑋(πœ‹, πœ…) = π‘₯0 β„° πœ‹ βˆ™ 𝑅 βˆ’ πœ… βˆ™ πœ‡ .
propensity to consume; e.g., we write
This simplies verifying that some pair
implies
π‘‹βˆ’ (πœ‹, πœ…) > 0
(πœ‹, πœ…)
that
is admissible as
𝑋(πœ‹, πœ…) > 0
(cf. [15, II.8a]).
The preferences of the agent are modeled by a time-additive random
Let 𝐷 be a càdlàg, adapted, strictly positive
]
𝐷𝑠 πœ‡βˆ˜ (𝑑𝑠) < ∞ and x 𝑝 ∈ (βˆ’βˆž, 0) βˆͺ (0, 1). We
utility function as follows.
process such that
𝐸
[βˆ«π‘‡
0
dene the power utility random eld
π‘ˆπ‘‘ (π‘₯) := 𝐷𝑑 𝑝1 π‘₯𝑝 ,
This is the general form of a
π‘₯ ∈ (0, ∞), 𝑑 ∈ [0, 𝑇 ].
𝑝-homogeneous
utility random eld such that
a constant consumption yields nite expected utility.
convex conjugate
Interpretations and
𝐷 are discussed in [26]. We
of π‘₯ 7β†’ π‘ˆπ‘‘ (π‘₯),
{
}
π‘ˆπ‘‘βˆ— (𝑦) = sup π‘ˆπ‘‘ (π‘₯) βˆ’ π‘₯𝑦 = βˆ’ 1π‘ž 𝑦 π‘ž 𝐷𝑑𝛽 ;
applications for the process
denote by
π‘ˆβˆ—
the
(2.1)
π‘₯>0
𝑝
π‘ž := π‘βˆ’1
∈ (βˆ’βˆž, 0) βˆͺ (0, 1) is the exponent conjugate to 𝑝
1
constant 𝛽 :=
1βˆ’π‘ > 0 is the relative risk tolerance of π‘ˆ . Note
here
4
and the
that we
exclude the well-studied logarithmic utility (e.g., Goll and Kallsen [11]) which
corresponds to
𝐸
𝑝 = 0.
expected utility
corresponding to a consumption strategy 𝑐 ∈ π’œ is
[ ∫The
]
∫
𝑇
∘ (𝑑𝑑) ; i.e., either 𝐸[π‘ˆ (𝑐 )] or 𝐸[ 𝑇 π‘ˆ (𝑐 ) 𝑑𝑑 + π‘ˆ (𝑐 )]. The
π‘ˆ
(𝑐
)
πœ‡
𝑑 𝑑
𝑑 𝑑
𝑇 𝑇
𝑇 𝑇
0
0
(value of the) utility maximization problem is said to be
𝑒(π‘₯0 ) := sup 𝐸
Note that this condition is void if
(πœ‹, 𝑐) ∈ π’œ(π‘₯0 )
is called
if
]
π‘ˆπ‘‘ (𝑐𝑑 ) πœ‡βˆ˜ (𝑑𝑑) < ∞.
(2.2)
0
π‘βˆˆπ’œ(π‘₯0 )
strategy
𝑇
[∫
nite
𝑝<0
optimal
if
π‘ˆ < 0. If (2.2) holds,
]
π‘ˆπ‘‘ (𝑐𝑑 ) πœ‡βˆ˜ (𝑑𝑑) = 𝑒(π‘₯0 ).
as then
𝐸
[βˆ«π‘‡
0
a
Finally, we introduce the following sets; they are of minor importance
and used only in the case
𝑝 < 0:
}
βˆ«π‘‡
π’œπ‘“ := (πœ‹, 𝑐) ∈ π’œ : 0 π‘ˆπ‘‘ (𝑐𝑑 ) πœ‡βˆ˜ (𝑑𝑑) > βˆ’βˆž ,
]
}
{
[∫ 𝑇
π’œπ‘“ 𝐸 := (πœ‹, 𝑐) ∈ π’œ : 𝐸 0 π‘ˆπ‘‘ (𝑐𝑑 ) πœ‡βˆ˜ (𝑑𝑑) > βˆ’βˆž .
{
Anticipating that (2.2) will be in force, the indices stand for nite and
nite expectation. Clearly
π’œπ‘“ 𝐸 βŠ† π’œπ‘“ βŠ† π’œ,
and equality holds if
𝑝 ∈ (0, 1).
2.2 Opportunity Process
We recall the opportunity process, a reduced form of the value process in the
language of control theory. We assume (2.2) in this section, which ensures
that the following process is nite.
By [26, Proposition 3.1, Remark 3.7]
there exists a unique càdlàg semimartingale
𝐿,
called
opportunity process,
such that
𝐿𝑑
1
𝑝
(
𝑋𝑑 (πœ‹, 𝑐)
)𝑝
= ess sup 𝐸
π‘Λœβˆˆπ’œ(πœ‹,𝑐,𝑑)
[∫
𝑑
𝑇
]
π‘ˆπ‘  (˜
𝑐𝑠 ) πœ‡βˆ˜ (𝑑𝑠)ℱ𝑑
(2.3)
{
}
(πœ‹, 𝑐) ∈ π’œ, where π’œ(πœ‹, 𝑐, 𝑑) := (˜
πœ‹ , π‘Λœ) ∈ π’œ : (˜
πœ‹ , π‘Λœ) = (πœ‹, 𝑐) on [0, 𝑑] .
1 𝑝
note that 𝐿𝑇 = 𝐷𝑇 and that 𝑒(π‘₯0 ) = 𝐿0 π‘₯0 is the value function
𝑝
for any
We
from (2.2). The following is contained in [26, Lemma 3.5].
Lemma 2.1. 𝐿 is a special semimartingale for all 𝑝. If 𝑝 ∈ (0, 1), then
𝐿, πΏβˆ’ > 0 up to evanescence. If 𝑝 < 0, the same holds provided that an
optimal strategy exists.
Proposition 2.2 ([26, Proposition 3.4]). Let (πœ‹, 𝑐) ∈ π’œπ‘“ 𝐸 . Then the process
(
)𝑝
𝐿𝑑 𝑝1 𝑋𝑑 (πœ‹, 𝑐) +
∫
𝑑
π‘ˆπ‘  (𝑐𝑠 ) πœ‡(𝑑𝑠),
𝑑 ∈ [0, 𝑇 ]
0
is a supermartingale; it is a martingale if and only if (πœ‹, 𝑐) is optimal.
This is the martingale optimality principle.
value of this process equals
(πœ‹, 𝑐) ∈ π’œ βˆ– π’œπ‘“ 𝐸 .
𝐸[
βˆ«π‘‡
0
π‘ˆπ‘‘ (𝑐𝑑 ) πœ‡βˆ˜ (𝑑𝑑)],
5
The expected terminal
hence the assertion fails for
2.3 Semimartingale Characteristics
In the remainder of this section we introduce tools which are necessary to
describe the optimization problem locally. The use of semimartingale characteristics and set-valued processes follows [11] and [19], which consider logarithmic utility and convex constraints. That problem diers from ours in
that it is myopic; i.e., the characteristics of
𝑅
are sucient to describe the
local problem and so there is no need for an opportunity process.
We refer to [15] for background regarding semimartingale characteristics
and random measures. Let
πœ‡π‘…
be the integer-valued random measure associ-
β„Ž : ℝ𝑑 β†’ ℝ𝑑 be a cut-o function; i.e., β„Ž is
𝑅
𝑅 𝑅
bounded and β„Ž(π‘₯) = π‘₯ in a neighborhood of π‘₯ = 0. Let (𝐡 , 𝐢 , 𝜈 ) be the
predictable characteristics of 𝑅 relative to β„Ž. The canonical representation
of 𝑅 (cf. [15, II.2.35]) is
ated with the jumps of
𝑅
and let
𝑅 = 𝐡 𝑅 + 𝑅𝑐 + β„Ž(π‘₯) βˆ— (πœ‡π‘… βˆ’ 𝜈 𝑅 ) + (π‘₯ βˆ’ β„Ž(π‘₯)) βˆ— πœ‡π‘… .
The nite variation process
(π‘₯ βˆ’ β„Ž(π‘₯)) βˆ— πœ‡π‘…
(2.4)
contains essentially the large
𝑅. The rest is the canonical decomposition of the semimartingale
¯ = 𝑅 βˆ’ (π‘₯ βˆ’ β„Ž(π‘₯)) βˆ— πœ‡π‘… , which has bounded jumps: 𝐡 𝑅 = 𝐡 𝑅 (β„Ž) is
𝑅
𝑐
predictable of nite variation, 𝑅 is a continuous local martingale, and β„Ž(π‘₯)βˆ—
𝑅
𝑅
(πœ‡ βˆ’ 𝜈 ) is a purely discontinuous local martingale.
As 𝐿 is a special semimartingale (Lemma 2.1), it has a canonical de𝐿
𝐿
𝐿 is predictable
composition 𝐿 = 𝐿0 + 𝐴 + 𝑀 . Here 𝐿0 is constant, 𝐴
𝐿
of nite variation and also called the drift of 𝐿, 𝑀
is a local martingale,
𝐿
𝐿
and 𝐴0 = 𝑀0 = 0. Analogous notation will be used for other special semi𝐿
𝐿 𝐿
martingales. It is then possible to consider the characteristics (𝐴 , 𝐢 , 𝜈 )
β€²
of 𝐿 with respect to the identity instead of a cut-o function. Writing π‘₯ for
the identity on ℝ, the canonical representation is
jumps of
𝐿 = 𝐿0 + 𝐴𝐿 + 𝐿𝑐 + π‘₯β€² βˆ— (πœ‡πΏ βˆ’ 𝜈 𝐿 );
see [15, II.2.38]. It will be convenient to use the joint characteristics of the
ℝ𝑑 × β„-valued process (𝑅, 𝐿). We denote a generic point in ℝ𝑑 × β„ by (π‘₯, π‘₯β€² )
𝑅,𝐿 , 𝐢 𝑅,𝐿 , 𝜈 𝑅,𝐿 ) be the characteristics of (𝑅, 𝐿) with respect to the
and let (𝐡
β€²
β€²
function (π‘₯, π‘₯ ) 7β†’ (β„Ž(π‘₯), π‘₯ ). More precisely, we choose good versions of
the characteristics so that they satisfy the properties given in [15, II.2.9]. For
(𝑑 + 1)-dimensional process (𝑅, 𝐿) we have the canonical representation
( ) ( ) ( 𝑅) ( 𝑐) (
)
(
)
𝑅
0
𝐡
𝑅
β„Ž(π‘₯)
π‘₯ βˆ’ β„Ž(π‘₯)
𝑅,𝐿
𝑅,𝐿
=
+
+
+
βˆ—(πœ‡ βˆ’πœˆ )+
βˆ—πœ‡π‘…,𝐿 .
𝐿
𝐿0
𝐴𝐿
𝐿𝑐
π‘₯β€²
0
the
We denote by
(𝑏𝑅,𝐿 , 𝑐𝑅,𝐿 , 𝐹 𝑅,𝐿 ; 𝐴)
the
dierential
characteristics with
respect to a predictable locally integrable increasing process
𝐴𝑑 := 𝑑 +
βˆ‘
𝑖
Var(𝐡 𝑅𝐿,𝑖 )𝑑 +
βˆ‘
𝐴;
e.g.,
(
)
Var(𝐢 𝑅𝐿,𝑖𝑗 )𝑑 + ∣(π‘₯, π‘₯β€² )∣2 ∧ 1 βˆ— πœˆπ‘‘π‘…,𝐿 .
𝑖,𝑗
6
βˆ™ 𝐴 = 𝐢 𝑅,𝐿 , and 𝐹 𝑅,𝐿
𝑏𝑅,𝐿 βˆ™ 𝐴 = 𝐡 𝑅,𝐿 , 𝑐𝑅,𝐿 (
)
𝑐𝑅
𝑐𝑅𝐿
= (𝑏𝑅 , π‘ŽπΏ )⊀ and 𝑐𝑅,𝐿 = (𝑐𝑅𝐿
, i.e., 𝑐𝑅𝐿
⊀
𝐿
)
𝑐
Then
𝑏𝑅,𝐿
(𝑐𝑅𝐿 )
βˆ™
𝐴=
𝐴 = 𝜈 𝑅,𝐿 .
is a 𝑑-vector
βˆ™
We write
satisfying
βŸ¨π‘…π‘ , 𝐿𝑐 ⟩. We will often use that
∫
(∣π‘₯∣2 + ∣π‘₯β€² ∣2 ) ∧ (1 + ∣π‘₯β€² ∣) 𝐹 𝑅,𝐿 (𝑑(π‘₯, π‘₯β€² )) < ∞
(2.5)
ℝ𝑑 ×ℝ
because
𝐿
is a special semimartingale (cf. [15, II.2.29]). Let
π‘Œ
be any scalar
π‘Œ π‘Œ
π‘Œ
semimartingale with dierential characteristics (𝑏 , 𝑐 , 𝐹 ) relative to
and a cut-o function
β„ŽΜ„.
π‘Œ
𝐴
We call
π‘Œ
∫
π‘Ž := 𝑏 +
(
)
π‘₯ βˆ’ β„ŽΜ„(π‘₯) 𝐹 π‘Œ (𝑑π‘₯)
the drift rate of π‘Œ whenever the integral is well dened with values in
[βˆ’βˆž, ∞], even if it is not nite. Note that π‘Žπ‘Œ does not depend on the
choice of β„ŽΜ„. If π‘Œ is special, the drift rate is nite and even 𝐴-integrable (and
𝐿
𝐿
𝐿
vice versa). As an example, π‘Ž is the drift rate of 𝐿 and π‘Ž βˆ™ 𝐴 = 𝐴 yields
the drift.
Remark 2.3.
Then its
drift rate
fact that
Assume π‘Œ is a nonpositive scalar semimartingale.
π‘Žπ‘Œ is well dened with values in [βˆ’βˆž, ∞). Indeed, the
π‘Œ = π‘Œβˆ’ + Ξ”π‘Œ ≀ 0 implies that π‘₯ ≀ βˆ’π‘Œβˆ’ , 𝐹 π‘Œ (𝑑π‘₯)-a.e.
If
a
π‘Œ
π‘Žπ‘Œ ∈ [βˆ’βˆž, 0], we call π‘Œ
π‘Œ need not be nite,
Here π‘Ž
is a scalar semimartingale with drift rate
semimartingale with nonpositive drift rate.
as in the case of a compound Poisson process with negative, non-integrable
jumps. We refer to Kallsen [17] for the concept of
by
𝐿(𝐴)
the set of
𝐴-integrable
𝜎 -localization. Denoting
β„±0 is trivial, we
processes and recalling that
conclude the following, e.g., from [19, Appendix 3].
Lemma 2.4. Let π‘Œ be a semimartingale with nonpositive drift rate.
(i) π‘Œ is a 𝜎-supermartingale ⇔ π‘Žπ‘Œ is nite ⇔ π‘Œ is 𝜎-locally of class (D).
(ii) π‘Œ is a local supermartingale ⇔ π‘Žπ‘Œ ∈ 𝐿(𝐴) ⇔ π‘Œ is locally of class (D).
(iii) If π‘Œ is uniformly bounded from below, it is a supermartingale.
2.4 Constraints and Degeneracies
We introduce some set-valued processes that will be used in the sequel, that
is, for each
(πœ”, 𝑑)
they describe a subset of
ℝ𝑑 .
We refer to Rockafellar [28]
and Aliprantis and Border [1, Ÿ18] for background.
We start by expressing the budget constraint in this fashion. The process
{
}
{
}
C𝑑0 (πœ”) := 𝑦 ∈ ℝ𝑑 : 𝐹𝑑𝑅 (πœ”) π‘₯ ∈ ℝ𝑑 : 𝑦 ⊀ π‘₯ < βˆ’1 = 0
was called the
natural constraints
in [19].
Clearly
C0
is closed, convex,
and contains the origin. Moreover, one can check (see [19, Ÿ3.3]) that it is
7
predictable
(C 0 )βˆ’1 (𝐺)
in the sense that for each closed
= {(πœ”, 𝑑) : C𝑑 (πœ”) ∩ 𝐺 βˆ•= βˆ…}
𝐺 βŠ† ℝ𝑑 ,
the lower inverse image
is predictable. (Here one can replace
closed by compact or by open; see [28, 1A].) A statement such as C
0
closed means that C𝑑 (πœ”) is closed for all
omit the arguments
(πœ”, 𝑑).
(πœ”, 𝑑);
0 is
moreover, we will often
We also consider the slightly smaller set-valued
process
{
}
{
}
C 0,βˆ— := 𝑦 ∈ ℝ𝑑 : 𝐹 𝑅 π‘₯ ∈ ℝ𝑑 : 𝑦 ⊀ π‘₯ ≀ βˆ’1 = 0 .
These processes relate to the budget constraint as follows.
Lemma 2.5. A process πœ‹ ∈ 𝐿(𝑅) satises β„°(πœ‹ βˆ™ 𝑅) β‰₯ 0 (> 0) up to evanescence if and only if πœ‹ ∈ C 0 (C 0,βˆ— ) 𝑃 βŠ— 𝐴-a.e.
Proof.
β„°(πœ‹ βˆ™ 𝑅) > 0 if and only if 1 + πœ‹ ⊀ Δ𝑅 > 0 ([15, II.8a]).
Writing 𝑉 (π‘₯) = 1{π‘₯: 1+πœ‹ ⊀ π‘₯≀0} (π‘₯), we have that (𝑃 βŠ— 𝐴){πœ‹ ∈
/ C 0,βˆ— } =
]
[
βˆ‘
𝐸[𝑉 (π‘₯) βˆ— πœˆπ‘‡π‘… ] = 𝐸[𝑉 (π‘₯) βˆ— πœ‡π‘…
𝑠≀𝑇 1{π‘₯: 1+πœ‹π‘ βŠ€ Δ𝑅𝑠 ≀0} . For the equiva𝑇] = 𝐸
0
lence with C , interchange strict and non-strict inequality signs.
Recall that
The process
C 0,βˆ—
C 0,βˆ—
is not closed in general (nor relatively open). Clearly
C 0 , and in fact
C 0 is the closure of C 0,βˆ— : for 𝑦 ∈ C𝑑0 (πœ”), the sequence
0,βˆ—
0,βˆ—
is in C𝑑 (πœ”) and converges to 𝑦 . This implies that C
0,βˆ—
cf. [1, 18.3]. We will not be able to work directly with C
βŠ†
{(1 βˆ’ 1/𝑛)𝑦}𝑛β‰₯1
is predictable;
because closedness is essential for the measurable selection arguments that
will be used.
We turn to the exogenous portfolio constraints; i.e., the set-valued process
C
containing the origin. We consider the following conditions:
(C1) C
(C2) C
(C3) If
is predictable.
is closed.
𝑝 ∈ (0, 1): There exists a (0, 1)-valued process πœ‚ such that
𝑦 ∈ (C ∩ C 0 ) βˆ– C 0,βˆ— =β‡’ πœ‚π‘¦ ∈ C for all πœ‚ ∈ (πœ‚, 1), 𝑃 βŠ— 𝐴-a.e.
Condition (C3) is clearly satised if
of a continuous process
𝑅,
C ∩ C 0 βŠ† C 0,βˆ— ,
which includes the case
C is convex or, more
𝑝 < 0, (C3) should be
and it is always satised if
generally, star-shaped with respect to the origin. If
read as always being satised.
We require (C3) to exclude a degenerate situation where, despite the
Inada condition
π‘ˆ β€² (0) = ∞,
it is actually desirable for the agent to have
a wealth process that vanishes in some states.
That situation, illustrated
in the subsequent example, would necessitate a more complicated notation
while it can arise only in cases that are of minor interest.
Example 2.6.
We assume that there is no intermediate consumption and
π‘₯0 = 1. Consider the one-period binomial model of a nancial market; i.e.,
𝑆 = β„°(𝑅) is a scalar process which is constant up to time 𝑇 , where it has
a single jump; say, 𝑃 [Δ𝑅𝑇 = βˆ’1] = 𝑝0 and 𝑃 [Δ𝑅𝑇 = 𝐾] = 1 βˆ’ 𝑝0 , where
8
𝑝0 ∈ (0, 1). The ltration is generated by 𝑅 and
C ≑ {0} βˆͺ {1}. Then 𝐸[π‘ˆ (𝑋𝑇 (πœ‹))] = π‘ˆ (1) if πœ‹π‘‡ = 0 and
𝐸[π‘ˆ (𝑋𝑇 (πœ‹))] = 𝑝0 π‘ˆ (0) + (1 βˆ’ 𝑝0 )π‘ˆ (1 + 𝐾) if πœ‹π‘‡ = 1. If π‘ˆ (0) > βˆ’βˆž, and if
𝐾 is large enough, πœ‹π‘‡ = 1 performs better despite the fact that its terminal
wealth vanishes with probability 𝑝0 > 0. Of course, this cannot happen if
π‘ˆ (0) = βˆ’βˆž, i.e., 𝑝 < 0.
𝐾 >0
is a constant and
we consider
By adjusting the constants in the example, one can also see that under
non-convex constraints, there is in general
no uniqueness
for the optimal
wealth processes (even if they are positive).
The nal set-valued process is related to linear dependencies of the assets.
As in [19], the predictable process of
null-investments
is
{
}
N := 𝑦 ∈ ℝ𝑑 : 𝑦 ⊀ 𝑏𝑅 = 0, 𝑦 ⊀ 𝑐𝑅 = 0, 𝐹 𝑅 {π‘₯ : 𝑦 ⊀ π‘₯ βˆ•= 0} = 0 .
ℝ𝑑 , hence closed, and provide the pointwise
description of the null-space of 𝐻 7β†’ 𝐻 βˆ™ 𝑅. That is, 𝐻 ∈ 𝐿(𝑅) satises
𝐻 βˆ™ 𝑅 ≑ 0 if and only if 𝐻 ∈ N 𝑃 βŠ— 𝐴-a.e. An investment with values in
N has no eect on the wealth process.
Its values are linear subspaces of
3 The Bellman Equation
We have now introduced the necessary notation to formulate our rst main
result. Two special cases of our Bellman equation can be found in the pioneering work of Mania and Tevzadze [23] and Hu
et al. [14].
These articles
consider models with continuous asset prices and we shall indicate the connections as we specialize to that case in Section 3.3. A related equation also
arises in the study of mean-variance hedging by Βƒerný and Kallsen [5] in
the context of locally square-integrable semimartingales, although they do
not use dynamic programming explicitly. Due to the quadratic setting, that
equation is more explicit than ours and the mathematical treatment is quite
dierent. Czichowsky and Schweizer [7] study a cone-constrained version of
the related Markowitz problem and there the equation is no longer explicit.
The Bellman equation highlights the local structure of our utility maximization problem. In addition, it has two main benets. First, it can be
used as an abstract tool to derive properties of the optimal strategies and the
opportunity process (e.g., Nutz [27]). Second, one can try to solve the equation directly in a given model and to deduce the optimal strategies. This is
the point of view taken in Section 5 and obviously requires the precise form
of the equation.
The following assumptions are in force for the entire Section 3.
Assumptions 3.1.
The value of the utility maximization problem is nite,
there exists an optimal strategy
(Λ†
πœ‹ , 𝑐ˆ) ∈ π’œ,
9
and
C
satises (C1)-(C3).
3.1 Bellman Equation in Joint Characteristics
Our rst main result is the Bellman equation stated as a description of
the drift rate of the opportunity process. We recall the conjugate function
π‘ˆπ‘‘βˆ— (𝑦) = βˆ’ 1π‘ž 𝑦 π‘ž 𝐷𝑑𝛽 .
Theorem 3.2. The drift rate π‘ŽπΏ of the opportunity process satises
π‘‘πœ‡
βˆ’π‘βˆ’1 π‘ŽπΏ = π‘ˆ βˆ— (πΏβˆ’ ) 𝑑𝐴
+ max 𝑔(𝑦),
(3.1)
π‘¦βˆˆC ∩C 0
where 𝑔 is the predictable random function
𝑔(𝑦) := πΏβˆ’ 𝑦
∫
+
⊀
(
𝑅
𝑏 +
𝑐𝑅𝐿
πΏβˆ’
+
(π‘βˆ’1) 𝑅
2 𝑐 𝑦
)
∫
π‘₯β€² 𝑦 ⊀ β„Ž(π‘₯) 𝐹 𝑅,𝐿 (𝑑(π‘₯, π‘₯β€² ))
+
ℝ𝑑 ×ℝ
{
}
(πΏβˆ’ + π‘₯β€² ) π‘βˆ’1 (1 + 𝑦 ⊀ π‘₯)𝑝 βˆ’ π‘βˆ’1 βˆ’ 𝑦 ⊀ β„Ž(π‘₯) 𝐹 𝑅,𝐿 (𝑑(π‘₯, π‘₯β€² )).
ℝ𝑑 ×ℝ
(3.2)
The unique (𝑃 βŠ— πœ‡βˆ˜ -a.e.) optimal propensity to consume is
πœ…
Λ†=
(𝐷)
1
1βˆ’π‘
.
(3.3)
πœ‹ βˆ— ∈ arg max 𝑔
(3.4)
𝐿
Any optimal trading strategy πœ‹βˆ— satises
C ∩C 0
and the corresponding optimal wealth process and consumption are given by
(
)
𝑋 βˆ— = π‘₯0 β„° πœ‹ βˆ— βˆ™ 𝑅 βˆ’ πœ…
Λ†βˆ™ πœ‡ ;
π‘βˆ— = 𝑋 βˆ— πœ…
Λ†.
We shall see in the proof that the maximization in (3.1) can be understood as a local version of the optimization problem. Indeed, recalling (2.1),
the right hand side of (3.1) is the maximum of a single function over certain
points
(π‘˜, 𝑦) ∈ ℝ+ × β„π‘‘
that correspond to the admissible controls
(πœ…, πœ‹).
Moreover, optimal controls are related to maximizers of this function, a characteristic feature of any dynamic programming equation. The maximum of
𝑔
is not explicit due to the jumps of
𝑅;
this simplies in the continuous case
considered in Section 3.3 below. Some mathematical comments are also in
order.
Remark 3.3.
(i) The random function
𝑔
is well dened on
C0
in the
extended sense (see Lemma A.2) and it does not depend on the choice
of the cut-o function
(ii) For
𝑝<0
β„Ž
by [15, II.2.25].
πœ‹ βˆ— ∈ 𝐿(𝑅) and πœ…
Λ†
βˆ—
πœ‹ takes values in C ∩ C 0
we have a more precise statement: Given
(πœ‹ βˆ— , πœ…
Λ†)
and maximizes 𝑔 .
βˆ— Λ† ).
triplet (𝐿, πœ‹ , πœ…
as in (3.3),
is optimal
if and only if
This will follow from Corollary 5.4 applied to the
10
(iii) For
𝑝 ∈ (0, 1),
partial results in this direction follow from Section 5.
The question is trivial for convex
(iv) If
C
arg maxC ∩C 0 𝑔
elements lies in N
is convex,
of any two
C
by the next item.
is unique in the sense that the dierence
(see Lemma A.3).
We split the proof of Theorem 3.2 into several steps; the plan is as follows.
Let
(πœ‹, πœ…) ∈ π’œπ‘“ 𝐸
𝑋 = 𝑋(πœ‹, πœ…). We recall from
∫
𝑍(πœ‹, πœ…) := 𝐿 𝑝1 𝑋 𝑝 + π‘ˆπ‘  (πœ…π‘  𝑋𝑠 ) πœ‡(𝑑𝑠)
and denote
that
is a supermartingale, and a martingale if and only if
Proposition 2.2
(πœ‹, πœ…) is optimal. Hence
(πœ‹, πœ…); the maximum
we shall calculate its drift rate and then maximize over
will be attained at any optimal strategy. This is fairly straightforward and
essentially the content of Lemma 3.7 below.
𝑑
maximize over a subset of ℝ for each
(πœ”, 𝑑)
In the Bellman equation, we
and not over a set of strategies.
This nal step is a measurable selection problem and its solution will be the
second part of the proof.
Lemma 3.4. Let (πœ‹, πœ…) ∈ π’œπ‘“ . The drift rate of 𝑍(πœ‹, πœ…) is
(
)
π‘‘πœ‡
π‘Žπ‘(πœ‹,πœ…) = 𝑋(πœ‹, πœ…)π‘βˆ’ π‘βˆ’1 π‘ŽπΏ + 𝑓 (πœ…) 𝑑𝐴
+ 𝑔(πœ‹) ∈ [βˆ’βˆž, ∞),
where 𝑓𝑑 (π‘˜) := π‘ˆπ‘‘ (π‘˜) βˆ’ πΏπ‘‘βˆ’ π‘˜ and 𝑔 is given by (3.2). Moreover, π‘Žπ‘(Λ†πœ‹,Λ†πœ…) = 0,
and π‘Žπ‘(πœ‹,πœ…) ∈ (βˆ’βˆž, 0] for (πœ‹, πœ…) ∈ π’œπ‘“ 𝐸 .
Proof. We can assume that the initial capital is π‘₯0 = 1. Let (πœ‹, πœ…) ∈ π’œπ‘“ ,
𝑍 := 𝑍(πœ‹, πœ…) is nite. We also set 𝑋 := 𝑋(πœ‹, πœ…). By Itô's
𝑋 𝑝 = β„°(πœ‹ βˆ™ 𝑅 βˆ’ πœ… βˆ™ πœ‡)𝑝 = β„°(π‘Œ ) with
{
}
⊀ 𝑅 βˆ™
⊀ 𝑝
⊀
π‘Œ = 𝑝(πœ‹ βˆ™ 𝑅 βˆ’ πœ… βˆ™ πœ‡) + 𝑝(π‘βˆ’1)
πœ‹
𝑐
πœ‹
𝐴
+
(1
+
πœ‹
π‘₯)
βˆ’
1
βˆ’
π‘πœ‹
π‘₯
βˆ— πœ‡π‘… .
2
then in particular
formula, we have
𝑍 and using 𝑋𝑠 = π‘‹π‘ βˆ’ πœ‡(𝑑𝑠)-a.e.
βˆ’π‘ βˆ™
π‘‹βˆ’
𝑍 = π‘βˆ’1 (πΏβˆ’πΏ0 +πΏβˆ’ βˆ™ π‘Œ +[𝐿, π‘Œ ])+π‘ˆ (πœ…) βˆ™ πœ‡.
Integrating by parts in the denition of
(path-by-path), we have
Here
[𝐿, π‘Œ ] = [𝐿𝑐 , π‘Œ 𝑐 ] +
βˆ‘
Ξ”πΏΞ”π‘Œ
{
}
= π‘πœ‹ ⊀ 𝑐𝑅𝐿 βˆ™ 𝐴 + 𝑝π‘₯β€² πœ‹ ⊀ π‘₯ βˆ— πœ‡π‘…,𝐿 + π‘₯β€² (1 + πœ‹ ⊀ π‘₯)𝑝 βˆ’ 1 βˆ’ π‘πœ‹ ⊀ π‘₯ βˆ— πœ‡π‘…,𝐿 .
Thus
βˆ’π‘ βˆ™
𝑍
π‘‹βˆ’
equals
⊀ 𝑅 βˆ™
⊀ 𝑅𝐿 βˆ™
𝐴
π‘βˆ’1 (𝐿 βˆ’ 𝐿0 ) + πΏβˆ’ πœ‹ βˆ™ 𝑅 + 𝑓 (πœ…) βˆ™ πœ‡ + πΏβˆ’ (π‘βˆ’1)
2 πœ‹ 𝑐 πœ‹ 𝐴+πœ‹ 𝑐
{
}
+ π‘₯β€² πœ‹ ⊀ π‘₯ βˆ— πœ‡π‘…,𝐿 + (πΏβˆ’ + π‘₯β€² ) π‘βˆ’1 (1 + πœ‹ ⊀ π‘₯)𝑝 βˆ’ π‘βˆ’1 βˆ’ πœ‹ ⊀ π‘₯ βˆ— πœ‡π‘…,𝐿 .
Writing
π‘₯ = β„Ž(π‘₯) + π‘₯ βˆ’ β„Ž(π‘₯)
and
¯ = 𝑅 βˆ’ (π‘₯ βˆ’ β„Ž(π‘₯)) βˆ— πœ‡π‘…
𝑅
as in (2.4),
βˆ’π‘ βˆ™
π‘‹βˆ’
𝑍=
(3.5)
(
⊀ 𝑐𝑅𝐿
(π‘βˆ’1) 𝑅 )
2 𝑐 πœ‹
βˆ’1
⊀
¯ + 𝑓 (πœ…) βˆ™ πœ‡ + πΏβˆ’ πœ‹
π‘βˆ’1 (𝐿 βˆ’ 𝐿0 ) + πΏβˆ’ πœ‹ βˆ™ 𝑅
πΏβˆ’ +
{
+ π‘₯β€² πœ‹ ⊀ β„Ž(π‘₯) βˆ— πœ‡π‘…,𝐿 + (πΏβˆ’ + π‘₯β€² ) π‘βˆ’1 (1 + πœ‹ ⊀ π‘₯)𝑝 βˆ’ 𝑝
11
𝐴
}
βˆ’ πœ‹ β„Ž(π‘₯) βˆ— πœ‡π‘…,𝐿 .
βˆ™
Since
πœ‹
need not be locally bounded, we use from now on a predictable cut-
Then the compensator
πœ‹ ⊀ β„Ž(π‘₯) is bounded; e.g., β„Ž(π‘₯) = π‘₯1{∣π‘₯βˆ£β‰€1}∩{βˆ£πœ‹βŠ€ π‘₯βˆ£β‰€1} .
β€² ⊀
𝑅,𝐿 exists, since 𝐿 is special.
of π‘₯ πœ‹ β„Ž(π‘₯) βˆ— πœ‡
(πœ‹, πœ…) ∈ π’œπ‘“ 𝐸 .
Then the compensator of the last integral in the
o function
Let
β„Ž
such that
right hand side of (3.5) also exists; indeed, all other terms in that equality
are special, since
𝑍
is a supermartingale.
The drift rate can now be read
from (3.5) and (2.4), and it is nonpositive by the supermartingale property.
The drift rate vanishes for the optimal
(Λ†
πœ‹, πœ…
Λ†)
by the martingale condition
from Proposition 2.2.
(πœ‹, πœ…) ∈ π’œπ‘“ βˆ– π’œπ‘“ 𝐸 . Note that necessarily 𝑝 < 0 (otherwise
𝑍 is well
=
Thus 𝑍 ≀ 0, so by Remark 2.3 the drift rate π‘Ž
dened with values in [βˆ’βˆž, ∞)alternatively, this can also be read from
𝑍
the integrals in (3.5) via (2.5). Using directly the denition of π‘Ž , we nd
𝑍
the same formula for π‘Ž is as above.
Now consider
π’œπ‘“
π’œπ‘“ 𝐸 ).
We do not have the supermartingale property for
𝑍(πœ‹,πœ…)
is not evident that π‘Ž
≀0
(πœ‹, πœ…) ∈ π’œπ‘“ βˆ– π’œπ‘“ 𝐸 , so it
in that case. However, we have the following
Lemma 3.5. Let (πœ‹, πœ…) ∈ π’œπ‘“ . Then π‘Žπ‘ (πœ‹, πœ…) ∈ [0, ∞] implies π‘Žπ‘ (πœ‹, πœ…) = 0.
Proof.
Denote
𝑍 = 𝑍(πœ‹, πœ…).
For
𝑝 > 0
we have
π’œπ‘“ = π’œπ‘“ 𝐸 and the
Then 𝑍 ≀ 0 and by
Let 𝑝 < 0.
π‘Žπ‘ ∈ [0, ∞] implies that 𝑍 is a submartingale . Hence
]
[βˆ«π‘‡
𝐸[𝑍𝑇 ] = 𝐸 0 π‘ˆπ‘‘ (πœ…π‘‘ 𝑋𝑑 (πœ‹, πœ…)) πœ‡βˆ˜ (𝑑𝑑) > βˆ’βˆž, that is, (πœ‹, πœ…) ∈ π’œπ‘“ 𝐸 . Now
𝑍
Lemma 3.4 yields π‘Ž (πœ‹, πœ…) ≀ 0.
claim is immediate from Lemma 3.4.
Lemma 2.4(iii),
We observe in Lemma 3.4 that the drift rate splits into separate functions
involving
πœ…
and
πœ‹,
respectively. For this reason, we can single out the
Proof of the consumption formula
.
(πœ‹, πœ…) ∈ π’œ. Note the followβˆ—
ing feature of our parametrization: we have (πœ‹, πœ… ) ∈ π’œ for any nonnegaβˆ«π‘‡ βˆ—
βˆ—
βˆ—
tive optional process πœ… such that
0 πœ…π‘  πœ‡(𝑑𝑠) < ∞ and πœ…π‘‡ = 1. Indeed,
𝑋(πœ‹, πœ…) = π‘₯0 β„°(πœ‹ βˆ™ 𝑅 βˆ’ πœ… βˆ™ πœ‡) is positive by assumption. As πœ‡ is continuous,
𝑋(πœ‹, πœ…βˆ— ) = π‘₯0 β„°(πœ‹ βˆ™ 𝑅 βˆ’ πœ…βˆ— βˆ™ πœ‡) is also positive.
In particular, let (Λ†
πœ‹, πœ…
Λ† ) be optimal, 𝛽 = (1 βˆ’ 𝑝)βˆ’1 and πœ…βˆ— = (𝐷/𝐿)𝛽 ; then
(Λ†
πœ‹ , πœ…βˆ— ) ∈ π’œ. In fact the paths of π‘ˆ (πœ…βˆ— 𝑋(Λ†
πœ‹ , πœ…βˆ— )) = π‘βˆ’1 𝐷𝛽𝑝+1 𝑋(Λ†
πœ‹ , πœ…βˆ— )𝑝 πΏβˆ’π›½π‘
are bounded 𝑃 -a.s. (because the processes are càdlàg; 𝐿, πΏβˆ’ > 0 and 𝛽𝑝+1 =
𝛽 > 0) so that (Λ†
πœ‹ , πœ…βˆ— ) ∈ π’œπ‘“ .
βˆ—
𝛽
Note that 𝑃 βŠ— πœ‡-a.e., we have πœ… = (𝐷/πΏβˆ’ ) = arg maxπ‘˜β‰₯0 𝑓 (π‘˜), hence
βˆ—
βˆ—
𝑓 (πœ… ) β‰₯ 𝑓 (Λ†
πœ…). Suppose (𝑃 βŠ— πœ‡){𝑓 (πœ… ) > 𝑓 (Λ†
πœ…)} > 0, then the formula from
βˆ—)
βˆ—
𝑍(Λ†
πœ‹
,Λ†
πœ…
)
𝑍(Λ†
πœ‹
,πœ…
Lemma 3.4 and π‘Ž
= 0 imply π‘Ž
β‰₯ 0 and (𝑃 βŠ—π΄){π‘Žπ‘(Λ†πœ‹,πœ… ) > 0} > 0,
a contradiction to Lemma 3.5. It follows that πœ…
Λ† = πœ…βˆ— 𝑃 βŠ— πœ‡-a.e. since 𝑓 has
(3.3)
Let
a unique maximum.
Remark 3.6.
The previous proof does not use the assumptions (C1)-(C3).
12
Lemma 3.7. Let πœ‹ be a predictable process with values in C
∩ C 0,βˆ— .
Then
{
}
(𝑃 βŠ— 𝐴) 𝑔(Λ†
πœ‹ ) < 𝑔(πœ‹) = 0.
Proof.
We argue by contradiction and assume
By redening
πœ‹,
we may assume that
πœ‹ = πœ‹
Λ†
(𝑃 βŠ— 𝐴){𝑔(Λ†
πœ‹ ) < 𝑔(πœ‹)} > 0.
on the complement of this
predictable set. Then
𝑔(Λ†
πœ‹ ) ≀ 𝑔(πœ‹)
and
(𝑃 βŠ— 𝐴){𝑔(Λ†
πœ‹ ) < 𝑔(πœ‹)} > 0.
(3.6)
πœ‹ is 𝜎 -bounded, we can nd a constant 𝐢 > 0 such that the process
πœ‹
˜ := πœ‹1βˆ£πœ‹βˆ£β‰€πΆ + πœ‹
Λ† 1βˆ£πœ‹βˆ£>𝐢 again satises (3.6); that is, we may assume that πœ‹
0,βˆ— , this implies (πœ‹, πœ…
is 𝑅-integrable. Since πœ‹ ∈ C ∩ C
Λ† ) ∈ π’œ (as observed
above, the consumption πœ…
Λ† plays no role here). The contradiction follows as
As
in the previous proof.
In view of Lemma 3.7, the main task will be to construct a
maximizing sequence for
measurable
𝑔.
Lemma 3.8. Under Assumptions 3.1, there exists a sequence
dictable C ∩ C 0,βˆ— -valued processes such that
of pre-
𝑃 βŠ— 𝐴-a.e.
lim sup 𝑔(πœ‹ 𝑛 ) = sup 𝑔
C ∩C 0
𝑛
(πœ‹ 𝑛 )
We defer the proof of this lemma to Appendix A, together with the study
of the properties of
𝑔.
Proof of Theorem 3.2.
πœ‹ 𝑛 yields
π‘Žπ‘(Λ†πœ‹,Λ†πœ…) =
πœ‹ =
0 =
𝑃 βŠ— πœ‡-a.e.
The theorem can then be proved as follows.
Let
πœ‹π‘›
be as in Lemma 3.8. Then Lemma 3.7 with
𝑔(Λ†
πœ‹ ) = supC ∩C 0 𝑔 , which
π‘‘πœ‡
π‘βˆ’1 π‘ŽπΏ + 𝑓 (Λ†
πœ…) 𝑑𝐴
+ 𝑔(Λ†
πœ‹ ).
is (3.4).
By Lemma 3.4 we have
This is (3.1) as
𝑓 (Λ†
πœ…) = π‘ˆ βˆ— (πΏβˆ’ )
due to (3.3).
3.2 Bellman Equation as BSDE
In this section we express the Bellman equation as a BSDE. The unique
orthogonal decomposition of the local martingale
𝑀𝐿
with respect to
𝑅
(cf. [15, III.4.24]) leads to the representation
𝐿 = 𝐿0 + 𝐴𝐿 + πœ‘πΏ βˆ™ 𝑅𝑐 + π‘Š 𝐿 βˆ— (πœ‡π‘… βˆ’ 𝜈 𝑅 ) + 𝑁 𝐿 ,
(3.7)
πœ‘πΏ ∈ 𝐿2π‘™π‘œπ‘ (𝑅𝑐 ), π‘Š 𝐿 ∈ πΊπ‘™π‘œπ‘ (πœ‡π‘… ), and 𝑁 𝐿 is
𝐿 𝑐
𝑐
𝑃
𝐿 ˜ = 0. The
a local martingale such that ⟨(𝑁 ) , 𝑅 ⟩ = 0 and 𝑀 𝑅 (Δ𝑁 βˆ£π’«)
πœ‡
𝐿
𝑅
last statement means that 𝐸[(𝑉 Δ𝑁 )βˆ—πœ‡π‘‡ ] = 0 for any suciently integrable
predictable function 𝑉 = 𝑉 (πœ”, 𝑑, π‘₯). We also introduce
∫
ˆ𝑑𝐿 :=
π‘Š
π‘Š 𝐿 (𝑑, π‘₯) 𝜈 𝑅 ({𝑑} × π‘‘π‘₯),
where, using the notation of [15],
ℝ𝑑
13
(
)
Λ† 𝐿 by denition of the
Ξ” π‘Š 𝐿 βˆ— (πœ‡π‘… βˆ’ 𝜈 𝑅 ) = π‘Š 𝐿 (Δ𝑅)1{Ξ”π‘…βˆ•=0} βˆ’ π‘Š
𝐿
𝑅
𝑅
purely discontinuous local martingale π‘Š βˆ— (πœ‡ βˆ’ 𝜈 ) and we can write
then
Λ† 𝐿 + Δ𝑁 𝐿 .
Δ𝐿 = Δ𝐴𝐿 + π‘Š 𝐿 (Δ𝑅)1{Ξ”π‘…βˆ•=0} βˆ’ π‘Š
We recall that Assumptions 3.1 are in force. Now (3.1) can be restated as
follows, the random function
𝑔
being the same as before but in new notation.
Corollary 3.9. The opportunity process 𝐿 and the processes dened by (3.7)
satisfy the BSDE
𝐿 = 𝐿0 βˆ’ π‘π‘ˆ βˆ— (πΏβˆ’ ) βˆ™ πœ‡ βˆ’ 𝑝 max 𝑔(𝑦) βˆ™ 𝐴 + πœ‘πΏ βˆ™ 𝑅𝑐 + π‘Š 𝐿 βˆ— (πœ‡π‘… βˆ’ 𝜈 𝑅 ) + 𝑁 𝐿
π‘¦βˆˆC ∩C 0
(3.8)
with terminal condition 𝐿𝑇 = 𝐷𝑇 , where 𝑔 is given by
𝑔(𝑦) :=
(
) ∫ (
(
)
(π‘βˆ’1) )
𝑅 πœ‘πΏ
⊀ 𝑅
Λ† 𝐿 𝑦 ⊀ β„Ž(π‘₯) 𝐹 𝑅 (𝑑π‘₯)
πΏβˆ’ 𝑦 𝑏 + 𝑐 πΏβˆ’ + 2 𝑦 +
Δ𝐴𝐿 + π‘Š 𝐿 (π‘₯) βˆ’ π‘Š
ℝ𝑑
∫
(
){
}
Λ† 𝐿 π‘βˆ’1 (1 + 𝑦 ⊀ π‘₯)𝑝 βˆ’ π‘βˆ’1 βˆ’ 𝑦 ⊀ β„Ž(π‘₯) 𝐹 𝑅 (𝑑π‘₯).
+
πΏβˆ’ + Δ𝐴𝐿 + π‘Š 𝐿 (π‘₯) βˆ’ π‘Š
ℝ𝑑
We observe that the orthogonal part
of
𝑔.
𝑁 𝐿 does not appear in the denition
In a suitable setting, it is linked to the dual problem; see Remark 5.18.
It is possible (but notationally more cumbersome) to prove a version of
Lemma 3.4 using
𝑔
as in Corollary 3.9 and the decomposition (3.7), thus
involving only the characteristics of
of
(𝑅, 𝐿).
𝑅
instead of the joint characteristics
Using this approach, we see that the increasing process
BSDE can be chosen based on
𝑅 and without reference to 𝐿.
𝐴
in the
This is desirable
if we want to consider other solutions of the equation, as in Section 4. One
consequence is that
𝐴
𝑅 is
π‘βˆ’1 𝐴𝐿 = βˆ’π‘“ (Λ†
πœ…) βˆ™ πœ‡ βˆ’ 𝑔(Λ†
πœ‹ ) βˆ™ 𝐴,
respect to 𝐴 + πœ‡, and we conclude:
can be chosen to be continuous if and only if
quasi left continuous (cf. [15, II.2.9]). Since
Var(𝐴𝐿 )
is absolutely continuous with
Remark 3.10.
If
𝑅
is quasi left continuous,
𝐴𝐿
is continuous.
𝑅 is quasi left continuous, 𝜈 𝑅 ({𝑑} × β„π‘‘ ) = 0 for all 𝑑 by [15, II.1.19],
Λ† 𝐿 = 0 and we have the simpler formula
hence π‘Š
(
) ∫
(
(π‘βˆ’1) )
⊀ 𝑅
𝑅 πœ‘πΏ
𝑔(𝑦) = πΏβˆ’ 𝑦 𝑏 + 𝑐 πΏβˆ’ + 2 𝑦 +
π‘Š 𝐿 (π‘₯)𝑦 ⊀ β„Ž(π‘₯) 𝐹 𝑅 (𝑑π‘₯)
𝑑
ℝ
∫
(
){
}
+
πΏβˆ’ + π‘Š 𝐿 (π‘₯) π‘βˆ’1 (1 + 𝑦 ⊀ π‘₯)𝑝 βˆ’ π‘βˆ’1 βˆ’ 𝑦 ⊀ β„Ž(π‘₯) 𝐹 𝑅 (𝑑π‘₯).
If
ℝ𝑑
14
3.3 The Case of Continuous Prices
In this section we specialize the previous results to the case where
𝑅
is a
continuous semimartingale and mild additional conditions are satised. As
usual in this setting, the martingale part of
than
𝑅𝑐 .
𝑅
will be denoted by
𝑀
rather
In addition to Assumptions 3.1, the following conditions are in
force for the present Section 3.3.
Assumptions 3.11.
(i)
(ii)
𝑅
is continuous,
𝑅=𝑀+
∫
π‘‘βŸ¨π‘€ βŸ©πœ†
for some
(iii) the orthogonal projection of
Note that
C 0,βˆ— = ℝ𝑑
πœ† ∈ 𝐿2π‘™π‘œπ‘ (𝑀 )
C
onto
N
structure condition ),
(
βŠ₯ is closed.
due to (i), in particular (C3) is void. When
𝑅
is
continuous, it necessarily satises (ii) when a no-arbitrage property holds;
see Schweizer [29]. By (i) and (ii) we can write the dierential characteristics
βˆ‘
𝐴𝑑 := 𝑑 + 𝑑𝑖=1 βŸ¨π‘€ 𝑖 βŸ©π‘‘ . It will be convenient to
𝑅 = 𝜎𝜎 ⊀ , where 𝜎 is a predictable matrix-valued process; hence
factorize 𝑐
𝜎𝜎 ⊀ 𝑑𝐴 = π‘‘βŸ¨π‘€ ⟩. Then (ii) implies N = ker 𝜎 ⊀ because 𝜎𝜎 ⊀ 𝑦 = 0 implies
(𝜎 ⊀ 𝑦)⊀ (𝜎 ⊀ 𝑦) = 0. Since 𝜎 ⊀ : ker(𝜎 ⊀ )βŠ₯ β†’ 𝜎 ⊀ ℝ𝑑 is a homeomorphism, we
of
𝑅
with respect to, e.g.,
see that (iii) is equivalent to
𝜎⊀C
is closed.
This condition depends on the semimartingale
closedness of
C
itself if
𝜎
has full rank.
𝑅.
It is equivalent to the
For certain constraint sets (e.g.,
closed polyhedral or compact) the condition is satised for
all matrices 𝜎, but
not so, e.g., for non-polyhedral cone constraints. We mention that violation
of (iii) leads to nonexistence of optimal strategies in simple examples (cf. [25,
Example 3.5]) and we refer to Czichowsky and Schweizer [8] for background.
Under (i), (3.7) is the more usual Kunita-Watanabe decomposition
𝐿 = 𝐿0 + 𝐴𝐿 + πœ‘πΏ βˆ™ 𝑀 + 𝑁 𝐿 ,
where
πœ‘πΏ ∈ 𝐿2π‘™π‘œπ‘ (𝑀 )
and
see Ansel and Stricker [2,
𝑁𝐿
cas
is a local martingale such that
[𝑀, 𝑁 𝐿 ] = 0;
βˆ… βˆ•= 𝐾 βŠ† ℝ𝑑 is a closed set,
𝑑𝐾 (π‘₯) = min{∣π‘₯ βˆ’ π‘¦βˆ£ : 𝑦 ∈ 𝐾},
3]. If
we denote
𝑑2𝐾 is
𝐾
the squared distance. We also dene the (set-valued) projection Ξ 
which
𝑑
maps π‘₯ ∈ ℝ to the points in 𝐾 with minimal distance to π‘₯,
the Euclidean distance to
𝐾
by
and
{
}
Π𝐾 (π‘₯) = 𝑦 ∈ 𝐾 : ∣π‘₯ βˆ’ π‘¦βˆ£ = 𝑑𝐾 (π‘₯) βˆ•= βˆ….
If
𝐾
is convex,
Π𝐾
is the usual (single-valued) Euclidean projection. In the
present continuous setting, the random function
𝑔
(
πœ‘πΏ
𝑔(𝑦) = πΏβˆ’ 𝑦 ⊀ 𝜎𝜎 ⊀ πœ† +
+
πΏβˆ’
and so the Bellman BSDE becomes more explicit.
15
simplies considerably:
π‘βˆ’1
2 𝑦
)
(3.9)
Corollary 3.12. Any optimal trading strategy πœ‹βˆ— satises
𝜎 ⊀ πœ‹ βˆ— ∈ Π𝜎
⊀C
{
(
πœ‘πΏ )}
.
𝜎 ⊀ (1 βˆ’ 𝑝)βˆ’1 πœ† +
πΏβˆ’
The opportunity process satises the BSDE
𝐿 = 𝐿0 βˆ’ π‘π‘ˆ βˆ— (πΏβˆ’ ) βˆ™ πœ‡ + 𝐹 (πΏβˆ’ , πœ‘πΏ ) βˆ™ 𝐴 + πœ‘πΏ βˆ™ 𝑀 + 𝑁 𝐿 ;
𝐿𝑇 = 𝐷𝑇 ,
where
𝐹 (πΏβˆ’ , πœ‘πΏ ) =
)
{
(
(
πœ‘πΏ )
1
2
⊀
βˆ’1
+
πœ†+
πΏβˆ’ 𝑝(1 βˆ’ 𝑝)π‘‘πœŽβŠ€ C 𝜎 (1 βˆ’ 𝑝)
2
πΏβˆ’
𝑝 ⊀
π‘βˆ’1 𝜎
(
}
πœ‘πΏ )2
πœ†+
.
πΏβˆ’
2
𝑝
If C is a convex cone, 𝐹 (πΏβˆ’ , πœ‘πΏ ) = 2(π‘βˆ’1)
πΏβˆ’ Π𝜎 C 𝜎 ⊀ πœ† + πΏπœ‘βˆ’ . If
(
(
∫
𝐿)
𝐿 )⊀
𝑝
πΏβˆ’ πœ† + πΏπœ‘βˆ’ π‘‘βŸ¨π‘€ ⟩ πœ† + πΏπœ‘βˆ’ and
C = ℝ𝑑 , then 𝐹 (πΏβˆ’ , πœ‘πΏ ) βˆ™ 𝐴 = 2(π‘βˆ’1)
(
𝐿)
the unique (mod. N ) optimal trading strategy is πœ‹βˆ— = (1 βˆ’ 𝑝)βˆ’1 πœ† + πΏπœ‘βˆ’ .
Proof.
⊀
{
𝐿
(
)}
(
𝐿 )}
⊀ {
𝜎 ⊀ (arg maxC 𝑔) = Π𝜎 C 𝜎 ⊀ 𝛽 πœ†+ πΏπœ‘βˆ’
βˆ—
by completing the square in (3.9), moreover, for any πœ‹ ∈ arg maxC 𝑔 ,
(
)}
{ (
(
πœ‘πΏ )
πœ‘πΏ )
πœ‘πΏ )⊀ ⊀ (
βˆ—
βˆ’1 2
⊀
1
𝑔(πœ‹ ) = 2 πΏβˆ’ 𝛽 πœ† +
βˆ’ 𝛽 π‘‘πœŽβŠ€ C 𝜎 𝛽 πœ† +
.
𝜎𝜎 πœ† +
πΏβˆ’
πΏβˆ’
πΏβˆ’
Let
𝛽 = (1βˆ’π‘)βˆ’1 .
We obtain
Ξ  := Π𝜎 C is singlevalued, positively homogeneous, and Ξ π‘₯ is orthogonal to π‘₯ βˆ’ Ξ π‘₯ for any
(
𝐿)
π‘₯ ∈ ℝ𝑑 . Writing Ξ¨ := 𝜎 ⊀ πœ† + πΏπœ‘βˆ’ we get 𝑔(πœ‹ βˆ— ) = πΏβˆ’ 𝛽(Ξ Ξ¨)⊀ (Ξ¨ βˆ’ 21 Ξ Ξ¨) =
(
(
πΏβˆ’ 21 𝛽 Ξ Ξ¨)⊀ Ξ Ξ¨). Finally, Ξ Ξ¨ = Ξ¨ if C = ℝ𝑑 . The result follows from
In the case where
C,
and hence
𝜎⊀C ,
is a convex cone,
⊀
Corollary 3.9.
Of course the consumption formula (3.3) and Remark 3.3 still apply. We
remark that the BSDE for the unconstrained case
𝐷 = 1)
C = ℝ𝑑
(and
πœ‡ = 0,
was previously obtained in [23] in a similar spirit. A variant of the
constrained BSDE for an Itô process model (and
πœ‡ = 0, 𝐷 = 1)
appears
in [14], where a converse approach is taken: the equation is derived only
formally and then existence results for BSDEs are employed together with a
verication argument. We shall extend that result in Section 5 (Example 5.8)
when we study verication.
𝐿 is
log(𝐿)
If
for
continuous, the BSDE of Corollary 3.12 simplies if it is stated
rather than
𝐿,
but in general the given form is more convenient
as the jumps are hidden in
𝑁 𝐿.
16
Remark 3.13.
(i) Continuity of
𝑅 does not imply that 𝐿 is continuous.
For
instance, in the Itô process model of Barndor-Nielsen and Shephard [3] with
Lévy driven coecients, the opportunity process is not continuous. See, e.g.,
Theorem 3.3 and the subsequent remark in Kallsen and Muhle-Karbe [18]. If
𝑅 satises the structure condition and the ltration 𝔽 is continuous, it clearly
follows that 𝐿 is continuous. Here 𝔽 is called continuous if all 𝔽-martingales
are continuous, as, e.g., for the Brownian ltration. In general, 𝐿 is related
to the predictable characteristics of the asset returns rather than their levels.
As an example, Lévy models have jumps but constant characteristics; here
𝐿
turns out to be a smooth function (see [25]).
(ii)
In the present setting we see that
𝐹
has quadratic growth in
πœ‘πΏ ,
so that the Bellman equation is a quadratic BSDE (see also Example 5.8).
In general,
𝐹
does not satisfy the bounds which are usually assumed in
the theory of such BSDEs.
Together with existence results for the utility
maximization problem (see the citations from the introduction), the Bellman
equation yields various examples of BSDEs with the opportunity process as
a solution. This includes terminal conditions
𝐷𝑇
which are integrable and
unbounded (see also [26, Remark 2.4]).
4 Minimality of the Opportunity Process
This section considers the Bellman equation as such, having possibly many
solutions, and we characterize the opportunity process as the minimal solution. As mentioned above, it seems more natural to use the BSDE formulation for this purpose (but see Remark 4.4). We rst have to clarify what we
mean by a solution of the BSDE. We consider
𝑅
and
𝐴
as given. Since the
nite variation part in the BSDE is predictable, a solution will certainly be
a special semimartingale. If
β„“
is any special semimartingale, there exists a
unique orthogonal decomposition [15, III.4.24]
β„“ = β„“0 + 𝐴ℓ + πœ‘β„“ βˆ™ 𝑅𝑐 + π‘Š β„“ βˆ— (πœ‡π‘… βˆ’ 𝜈 𝑅 ) + 𝑁 β„“ ,
(4.1)
using the same notation as in (3.7). These processes are essentially unique,
and so it suces to consider the left hand side of the BSDE for the notion
of a solution. (In BSDE theory, a solution would be, at least, a quadruple.)
We dene the random function
Since
β„“
𝑔ℓ
as in Corollary 3.9, with
𝐿
replaced by
β„“.
is special, we have
∫
(∣π‘₯∣2 + ∣π‘₯β€² ∣2 ) ∧ (1 + ∣π‘₯β€² ∣) 𝐹 𝑅,β„“ (𝑑(π‘₯, π‘₯β€² )) < ∞
(4.2)
ℝ𝑑 ×ℝ
and the arguments from Lemma A.2 show that
values in
ℝ βˆͺ {sign(𝑝)∞}. Hence we
𝐿 replaced by β„“; i.e.,
𝑔ℓ
is well dened on
C0
with
can consider (formally at rst) the
BSDE (3.8) with
β„“ = β„“0 βˆ’π‘π‘ˆ βˆ— (β„“βˆ’ ) βˆ™ πœ‡βˆ’π‘ max 𝑔 β„“ (𝑦) βˆ™ 𝐴+πœ‘β„“ βˆ™ 𝑅𝑐 +π‘Š β„“ βˆ—(πœ‡π‘… βˆ’πœˆ 𝑅 )+𝑁 β„“
π‘¦βˆˆC ∩C 0
17
(4.3)
with terminal condition
ℓ𝑇 = 𝐷𝑇 .
Denition 4.1. A càdlàg special semimartingale β„“ is called a solution of the
Bellman equation (4.3) if
βˆ™ β„“, β„“βˆ’ > 0,
βˆ™
C ∩ C 0,βˆ— -valued
supC ∩C 0 𝑔 β„“ < ∞,
there exists a
βˆ™ β„“
πœ‹
Λ‡ ∈ 𝐿(𝑅)
such that
ℓ𝑇 = 𝐷𝑇 .
βˆ’1
𝑝) . We call
𝑔 β„“ (Λ‡
πœ‹) =
and the processes from (4.1) satisfy (4.3) with
Moreover, we dene
(𝐷/β„“)𝛽 , where
πœ…
Λ‡ :=
β„“,
strategy associated with
If the process
β„“>0
process
πœ‹
Λ‡
𝛽 = (1 βˆ’
and for brevity, we also call
(β„“, πœ‹
Λ‡, πœ…
Λ‡)
(Λ‡
πœ‹, πœ…
Λ‡)
the
a solution.
is not unique, we choose and x one. The assumption
excludes pathological cases where
β„“
jumps to zero and becomes posi-
tive immediately afterwards and thereby ensures that
πœ…
Λ‡
is admissible. More
precisely, the following holds.
Remark 4.2.
(i)
(ii)
(Λ‡
πœ‹, πœ…
Λ‡) ∈
Let
supC ∩C 0 𝑔 β„“
(iii) If
π‘βˆˆ
(β„“, πœ‹
Λ‡, πœ…
Λ‡)
is a predictable,
(0, 1), 𝑔 β„“ is nite on
(iv) The condition
𝑝<0
be a solution of the Bellman equation.
π’œπ‘“ 𝐸 .
β„“>0
𝐴-integrable
C∩
process.
C 0.
is automatically satised if (a)
𝑝 ∈ (0, 1)
or if (b)
and there is no intermediate consumption and Assumptions 3.1
are satised.
Proof.
βˆ«π‘‡
πœ…
Λ‡ 𝑠 πœ‡(𝑑𝑠) < ∞ 𝑃 -a.s. since the paths of β„“ are bounded
βˆ«π‘‡
πœ…π‘‘ 𝑋𝑑 (Λ‡
πœ‹, πœ…
Λ‡ )) πœ‡(𝑑𝑑) < ∞ as in the proof
away from zero. Moreover,
0 π‘ˆπ‘‘ (Λ‡
of (3.3) (stated after Lemma 3.5). This shows (Λ‡
πœ‹, πœ…
Λ‡ ) ∈ π’œπ‘“ . The fact that
𝑓
𝐸
(Λ‡
πœ‹, πœ…
Λ‡ ) ∈ π’œ is contained in the proof of Lemma 4.9 below.
β„“
β„“ πœ‹ ). Hence sup
β„“ βˆ™ 𝐴 is
β„“
(ii) We have 0 = 𝑔 (0) ≀ supC ∩C 0 𝑔 = 𝑔 (Λ‡
C ∩C 0 𝑔
(i) We have
0
well dened, and it is nite because otherwise (4.3) could not hold.
(iii)
𝑔ℓ
<∞
Note that
𝑝>0
implies
𝑔 β„“ > βˆ’βˆž
by its denition and (4.2), while
by assumption.
𝑝 > 0, (4.3) states that 𝐴ℓ
β„“βˆ’ > 0 implies β„“ β‰₯ 0,
β„“ is a supermartingale by Lemma 2.4. Since ℓ𝑇 = 𝐷𝑇 > 0, the minimum
principle for nonnegative supermartingales shows β„“ > 0. Under (b) the
assertion is a consequence of Theorem 4.5 below (which shows β„“ β‰₯ 𝐿 > 0)
upon noting that the condition β„“ > 0 is not used in its proof when there is
(iv) If
is decreasing. As
no intermediate consumption.
It may seem debatable to make existence of the maximizer
πœ‹
Λ‡
part of the
denition of a solution. However, associating a control with the solution is
crucial for the following theory. Some justication is given by the following
result for the continuous case (where
C 0,βˆ— = ℝ𝑑 ).
18
Proposition 4.3. Let β„“ be any càdlàg special semimartingale such that
β„“, β„“βˆ’ > 0. Under Assumptions 3.11, (C1) and (C2), there exists a C ∩ C 0,βˆ— valued predictable process πœ‹Λ‡ such that 𝑔ℓ (Λ‡πœ‹) = supC ∩C 0 𝑔ℓ < ∞, and any
such process is 𝑅-integrable.
Proof.
As
𝑔ℓ
is analogous to (3.9), it is continuous and its supremum over
ℝ𝑑 is nite. By continuity of 𝑅 and the structure condition,
βˆ«π‘‡ ⊀
βˆ«π‘‡ ⊀ 2
only if
0 πœ‹ π‘‘βŸ¨π‘€ βŸ©πœ‹ = 0 ∣𝜎 πœ‹βˆ£ 𝑑𝐴 < ∞ 𝑃 -a.s.
πœ‹ ∈ 𝐿(𝑅) if and
C is compact, then Lemma A.4 yields a measurable
⊀
𝜎⊀ C 𝜎 ⊀ πœ“
selector πœ‹ for arg maxC 𝑔 . As in the proof of Corollary 3.12, 𝜎 πœ‹ ∈ Ξ 
(
)
∫
𝑇
πœ‘β„“
⊀ 2
for πœ“ := 𝛽 πœ† +
β„“βˆ’ , which satises 0 ∣𝜎 πœ“βˆ£ 𝑑𝐴 < ∞ by denition of πœ† and
πœ‘β„“ . We note that ∣𝜎 ⊀ πœ‹βˆ£ ≀ ∣𝜎 ⊀ πœ“βˆ£+∣𝜎 ⊀ πœ‹βˆ’πœŽ ⊀ πœ“βˆ£ ≀ 2∣𝜎 ⊀ πœ“βˆ£ due to the denition
of the projection and 0 ∈ C .
In the general case we approximate C by a sequence of compact con𝑛 := C ∩ {π‘₯ ∈ ℝ𝑑 : ∣π‘₯∣ ≀ 𝑛}, each of which yields a selector πœ‹ 𝑛
straints C
⊀ 𝑛
⊀
⊀ 𝑛
for arg maxC 𝑛 𝑔 . By the above, ∣𝜎 πœ‹ ∣ ≀ 2∣𝜎 πœ“βˆ£, so the sequence (𝜎 πœ‹ )𝑛
is bounded for xed (πœ”, 𝑑). A random index argument as in the proof of
Lemma A.4 yields a selector πœ— for a cluster point of this sequence. We have
πœ— ∈ 𝜎 ⊀ C by closedness of this set and we nd a selector πœ‹
Λ‡ for the preimage
⊀
βˆ’1
((𝜎 ) πœ—)∩C using [28, 1Q]. We have πœ‹
Λ‡ ∈ arg maxC 𝑔 as the sets C 𝑛 increase
βˆ«π‘‡ ⊀ 2
βˆ«π‘‡ ⊀ 2
to C , and
Λ‡ ∣ 𝑑𝐴 ≀ 2 0 ∣𝜎 πœ“βˆ£ 𝑑𝐴 < ∞ shows πœ‹
Λ‡ ∈ 𝐿(𝑅).
0 ∣𝜎 πœ‹
Assume rst that
Another example for the construction of
πœ‹
Λ‡
is given in [25, Ÿ5]. In general,
two ingredients are needed: Existence of a maximizer for xed
(πœ”, 𝑑)
will
typically require a compactness condition in the form of a no-arbitrage assumption (in the previous proof, this is the structure condition). Moreover,
a measurable selection is required; here the techniques from the appendices
may be useful.
Remark 4.4.
The BSDE formulation of the Bellman equation has the ad-
vantage that we can choose
all solutions.
𝐴
based on
𝑅
and speak about the class of
However, we do not want to write proofs in this cumber-
some notation. Once we x a solution
β„“
(and maybe
𝐿,
and nitely many
other semimartingales), we can choose a new reference process
𝐴˜ = 𝐴 + 𝐴′
β€²
(where 𝐴 is increasing), with respect to which our semimartingales admit
dierential characteristics; in particular we can use the joint characteristics
˜ . As we change 𝐴, all drift rates change in that they
(𝑏𝑅,β„“ , 𝑐𝑅,β„“ , 𝐹 𝑅,β„“ ; 𝐴)
˜
are multiplied by 𝑑𝐴/𝑑𝐴
, so any (in)equalities between them are preserved.
With this in mind, we shall use the joint characteristics of (𝑅, β„“) in the sequel without further comment and treat the two formulations of the Bellman
equation as equivalent.
Our denition of a solution of the Bellman equation is loose in terms of
integrability assumptions.
Even in the continuous case, it is unclear how
19
many solutions exist. The next result shows that we can always identify
by taking the smallest one; i.e.,
𝐿≀ℓ
for any solution
Theorem 4.5. Under Assumptions 3.1, the opportunity process
acterized as the minimal solution of the Bellman equation.
Remark 4.6.
𝐿
β„“.
𝐿
is char-
As a consequence, the Bellman equation has a bounded solu-
tion if and only if the opportunity process is bounded (and similarly for other
integrability properties). In conjunction with [26, Ÿ4.2] this yields examples
of quadratic BSDEs which have bounded terminal value (for
𝐷𝑇
bounded),
but no bounded solution.
The proof of Theorem 4.5 is based on the following result; it is the fundamental property of any Bellman equation.
Proposition 4.7. Let
any (πœ‹, πœ…) ∈ π’œπ‘“ ,
𝑍(πœ‹, πœ…) :=
(β„“, πœ‹
Λ‡, πœ…
Λ‡)
β„“ 𝑝1
(
be a solution of the Bellman equation. For
)𝑝
𝑋(πœ‹, πœ…) +
∫
(
)
π‘ˆπ‘  πœ…π‘  𝑋𝑠 (πœ‹, πœ…) πœ‡(𝑑𝑠)
(4.4)
is a semimartingale with nonpositive drift rate. Moreover, 𝑍(Λ‡πœ‹, πœ…Λ‡ ) is a local
martingale.
Proof.
Let
(πœ‹, πœ…) ∈ π’œπ‘“ .
Note that
𝑍 := 𝑍(πœ‹, πœ…)
satises
sign(𝑝)𝑍 β‰₯ 0,
𝑍
hence has a well dened drift rate π‘Ž by Remark 2.3. The drift rate can be
β„“
calculated as in Lemma 3.4: If 𝑓 is dened similarly to the function 𝑓 in
that lemma but with
𝐿
replaced by
β„“,
then
{
}
π‘‘πœ‡
π‘Žπ‘ = 𝑋(πœ‹, πœ…)π‘βˆ’ π‘βˆ’1 π‘Žβ„“ + 𝑓 β„“ (πœ…) 𝑑𝐴
+ 𝑔 β„“ (πœ‹)
{(
) π‘‘πœ‡
}
= 𝑋(πœ‹, πœ…)π‘βˆ’ 𝑓 β„“ (πœ…) βˆ’ 𝑓 β„“ (Λ‡
πœ…) 𝑑𝐴
+ 𝑔 β„“ (πœ‹) βˆ’ 𝑔 β„“ (Λ‡
πœ‹) .
This is nonpositive because
case
(πœ‹, πœ…) := (Λ‡
πœ‹, πœ…
Λ‡ ) we have
sign(𝑝)𝑍 β‰₯ 0.
πœ…
Λ‡ and πœ‹
Λ‡ maximize 𝑓 β„“ and 𝑔 β„“ . For the special
𝑍
π‘Ž = 0 and so 𝑍 is a 𝜎 -martingale, thus a local
martingale as
Remark 4.8.
In Proposition 4.7, semimartingale with nonpositive drift
rate can be replaced by 𝜎 -supermartingale if
𝑔ℓ
is nite on
C ∩ C 0.
Theorem 4.5 follows from the next lemma (which is actually stronger).
We recall that for
𝑝<0
the opportunity process
𝐿
can be dened without
further assumptions.
Lemma 4.9. Let β„“ be a solution of the Bellman equation. If 𝑝 < 0, then
𝐿 ≀ β„“. For 𝑝 ∈ (0, 1), the same holds if (2.2) is satised and there exists an
optimal strategy.
20
Proof. Let (β„“, πœ‹Λ‡ , πœ…Λ‡ ) be a solution and dene 𝑍(πœ‹, πœ…) as in (4.4).
Case 𝑝 < 0: We choose (πœ‹, πœ…) := (Λ‡πœ‹, πœ…Λ‡ ). As 𝑍(Λ‡πœ‹, πœ…Λ‡ ) is a negative
cal martingale by Proposition 4.7, it is a submartingale.
𝐸[𝑍𝑇 (Λ‡
πœ‹, πœ…
Λ‡ )] > βˆ’βˆž,
lo-
In particular,
𝐿𝑇 = 𝐷𝑇 , this is the statement that the
i.e., (Λ‡
πœ‹, πœ…
Λ‡ ) ∈ π’œπ‘“ 𝐸 this completes the proof of ReΛ‡ := 𝑋(Λ‡
Λ‡,
πœ‡βˆ˜ = πœ‡ + 𝛿{𝑇 } . With 𝑋
πœ‹, πœ…
Λ‡ ) and 𝑐ˇ := πœ…
ˇ𝑋
and using
expected utility is nite,
mark 4.2(i). Recall that
ℓ𝑇 = 𝐷𝑇 = 𝐿𝑇 , we deduce
∫ 𝑑
]
[
1 ˇ𝑝
ℓ𝑑 𝑝 𝑋𝑑 +
π‘ˆπ‘  (Λ‡
𝑐𝑠 ) πœ‡(𝑑𝑠) = 𝑍𝑑 (Λ‡
πœ‹, πœ…
Λ‡ ) ≀ 𝐸 𝑍𝑇 (Λ‡
πœ‹, πœ…
Λ‡ )ℱ𝑑
0
] ∫ 𝑑
[∫ 𝑇
∘
π‘ˆπ‘  (˜
𝑐𝑠 ) πœ‡ (𝑑𝑠)ℱ𝑑 +
π‘ˆπ‘  (Λ‡
𝑐𝑠 ) πœ‡(𝑑𝑠)
≀ ess supπ‘Λœβˆˆπ’œ(Λ‡πœ‹,ˇ𝑐,𝑑) 𝐸
𝑑
0
∫ 𝑑
ˇ𝑝 +
= 𝐿𝑑 𝑝1 𝑋
π‘ˆπ‘  (Λ‡
𝑐𝑠 ) πœ‡(𝑑𝑠),
𝑑
and using
0
Case 𝑝 ∈ (0, 1):
Then
We choose
1 ˇ𝑝
𝑝 𝑋𝑑
< 0, we have ℓ𝑑 β‰₯ 𝐿𝑑 .
(πœ‹, πœ…) := (Λ†
πœ‹, πœ…
Λ† ) to be an optimal strategy.
where the last equality holds by (2.3). As
𝑍(Λ†
πœ‹, πœ…
Λ† ) β‰₯ 0 is a supermartingale by Proposition 4.7 and Lemma 2.4(iii),
and we obtain
ˆ𝑝 +
ℓ𝑑 𝑝1 𝑋
𝑑
∫
0
𝑑
]
[
π‘ˆπ‘  (Λ†
𝑐𝑠 ) πœ‡(𝑑𝑠) = 𝑍𝑑 (Λ†
πœ‹, πœ…
Λ† ) β‰₯ 𝐸 𝑍𝑇 (Λ†
πœ‹, πœ…
Λ† )ℱ𝑑
∫ 𝑑
]
[∫ 𝑇
∘
1 ˆ𝑝
π‘ˆπ‘  (Λ†
𝑐𝑠 ) πœ‡(𝑑𝑠)
=𝐸
π‘ˆπ‘  (Λ†
𝑐𝑠 ) πœ‡ (𝑑𝑠)ℱ𝑑 = 𝐿𝑑 𝑝 𝑋𝑑 +
0
by the optimality of
(Λ†
πœ‹, πœ…
Λ†)
0
and (2.3). More precisely, we have used the fact
(Λ†
πœ‹, πœ…
Λ† ) is also conditionally optimal (see [26,
conclude ℓ𝑑 β‰₯ 𝐿𝑑 .
that
we
Remark 3.3]). As
1 ˆ𝑝
𝑝 𝑋𝑑
> 0,
5 Verication
Suppose that we have found a solution of the Bellman equation; then we want
to know whether it is the opportunity process and whether the associated
strategy is optimal.
In applications, it might not be clear
a priori
that
an optimal strategy exists or even that the utility maximization problem
is nite.
Therefore, we stress that in this section these properties are not
assumed. Also, we do not need the assumptions on
C
made in Section 2.4
they are not necessary because we start with a given solution.
Generally speaking, verication involves the candidate for an optimal
control,
(Λ‡
πœ‹, πœ…
Λ‡)
in our case, and all the competing ones.
It is often very
dicult to check a condition involving all these controls, so it is desirable to
have a verication theorem whose assumptions involve only
(Λ‡
πœ‹, πœ…
Λ‡ ).
We present two verication approaches. The rst one is via the value process and is classical for general dynamic programming: it uses little structure
21
of the given problem. For
𝑝 ∈ (0, 1),
it yields the desired result. However,
in a general setting, this is not the case for
𝑝 < 0.
The second approach
uses the concavity of the utility function. To fully exploit this and make the
verication conditions necessary, we will assume that
C
is convex. In this
case, we shall obtain the desired verication theorem for all values of
𝑝.
5.1 Verication via the Value Process
The basis of this approach is the following simple result; we state it separately
for better comparison with Lemma 5.10 below. In the entire section,
is dened by (4.4) whenever
β„“
𝑍(πœ‹, πœ…)
is given.
Lemma 5.1. Let β„“ be any positive càdlàg semimartingale with ℓ𝑇 = 𝐷𝑇 and
let (Λ‡πœ‹, πœ…Λ‡ ) ∈ π’œ. Assume that for all (πœ‹, πœ…) ∈ π’œπ‘“ 𝐸 , the process 𝑍(πœ‹, πœ…) is a
supermartingale. Then 𝑍(Λ‡πœ‹, πœ…Λ‡ ) is a martingale if and only if (2.2) holds and
(Λ‡
πœ‹, πœ…
Λ‡ ) is optimal and β„“ = 𝐿.
Proof. β‡’:
𝑍0 (πœ‹, πœ…) = β„“0 𝑝1 π‘₯𝑝0 does not depend on (πœ‹, πœ…) and
βˆ«π‘‡
∘
that 𝐸[𝑍𝑇 (πœ‹, πœ…)] = 𝐸[
0 π‘ˆπ‘‘ (πœ…π‘‘ (𝑋𝑑 (πœ‹, πœ…))) πœ‡ (𝑑𝑑)] is the expected utility corΛ‡
responding to (πœ‹, πœ…). With 𝑋 := 𝑋(Λ‡
πœ‹, πœ…
Λ‡ ), the (super)martingale condiβˆ«π‘‡
βˆ«π‘‡
∘
∘
Λ‡
π‘ˆ
(Λ‡
πœ…
𝑋
)
πœ‡
(𝑑𝑑)]
β‰₯
𝐸[
tion implies that 𝐸[
𝑑
𝑑
𝑑
0 π‘ˆπ‘‘ (πœ…π‘‘ 𝑋𝑑 (πœ‹, πœ…)) πœ‡ (𝑑𝑑)] for
0
𝑓
𝐸
𝑓
𝐸
all (πœ‹, πœ…) ∈ π’œ
. Since for (πœ‹, πœ…) ∈ π’œ βˆ– π’œ
the expected utility is βˆ’βˆž, this
shows that (Λ‡
πœ‹, πœ…
Λ‡ ) is optimal with 𝐸[𝑍𝑇 (Λ‡
πœ‹, πœ…
Λ‡ )] = 𝑍0 (Λ‡
πœ‹, πœ…
Λ‡ ) = β„“0 𝑝1 π‘₯𝑝0 < ∞. In
particular, the opportunity process 𝐿 is well dened. By Proposition 2.2,
∫
Λ‡ 𝑝 + π‘ˆπ‘  (Λ‡
𝐿 𝑝1 𝑋
𝑐𝑠 ) πœ‡(𝑑𝑠) is a martingale, and as its terminal value equals
Λ‡ > 0.
𝑍𝑇 (Λ‡
πœ‹, πœ…
Λ‡ ), we deduce β„“ = 𝐿 by comparison with (4.4), using 𝑋
Recall that
The converse is contained in Proposition 2.2.
We can now state our rst verication theorem.
Theorem 5.2. Let (β„“, πœ‹Λ‡ , πœ…Λ‡) be a solution of the Bellman equation.
(i) If 𝑝 ∈ (0, 1), the following are equivalent:
(a) 𝑍(Λ‡πœ‹, πœ…Λ‡ ) is of class (D),
(b) 𝑍(Λ‡πœ‹, πœ…Λ‡ ) is a martingale,
(c) (2.2) holds and (Λ‡πœ‹, πœ…Λ‡ ) is optimal and β„“ = 𝐿.
(ii) If 𝑝 < 0, the following are equivalent:
(a) 𝑍(πœ‹, πœ…) is of class (D) for all (πœ‹, πœ…) ∈ π’œπ‘“ 𝐸 ,
(b) 𝑍(πœ‹, πœ…) is a supermartingale for all (πœ‹, πœ…) ∈ π’œπ‘“ 𝐸 ,
(c) (Λ‡πœ‹, πœ…Λ‡ ) is optimal and β„“ = 𝐿.
Proof.
(πœ‹, πœ…) ∈ π’œπ‘“ , 𝑍(πœ‹, πœ…) is positive and π‘Žπ‘(πœ‹,πœ…) ≀ 0 by
Proposition 4.7, hence 𝑍(πœ‹, πœ…) is a supermartingale according to Lemma 2.4.
By Proposition 4.7, 𝑍(Λ‡
πœ‹, πœ…
Λ‡ ) is a local martingale, so it is a martingale if and
When
𝑝>0
and
only if it is of class (D). Lemma 5.1 implies the result.
22
If
𝑝 < 0, 𝑍(πœ‹, πœ…)
is negative.
Thus the local martingale
𝑍(Λ‡
πœ‹, πœ…
Λ‡)
is a
submartingale, and a martingale if and only if it is also a supermartingale.
Note that a class (D) semimartingale with nonpositive drift rate is a supermartingale. Conversely, any negative supermartingale
to the bounds
that if
β„“ = 𝐿,
0 β‰₯ 𝑍 β‰₯ 𝐸[𝑍𝑇 βˆ£π”½].
𝑝<0
is of class (D) due
then Proposition 2.2 yields (b).
Theorem 5.2 is as good as it gets for
for
𝑍
Lemma 5.1 implies the result after noting
𝑝 > 0, but as announced, the result
is not satisfactory. In particular settings, this can be improved.
Remark 5.3
< 0). (i) Assume we know a priori that if there
(Λ†
πœ‹, πœ…
Λ† ) ∈ π’œ, then
{
}
(Λ†
πœ‹, πœ…
Λ† ) ∈ π’œ(𝐷) := (πœ‹, πœ…) ∈ π’œ : 𝑋(πœ‹, πœ…)𝑝 is of class (D) .
(𝑝
is an
optimal strategy
π’œ(𝐷) . If
furthermore β„“ is bounded (which is not a strong assumption when 𝑝 < 0), the
class (D) condition in Theorem 5.2(ii) is automatically satised for (πœ‹, πœ…) ∈
π’œ(𝐷) . The verication then reduces to checking that (Λ‡
πœ‹, πœ…
Λ‡ ) ∈ π’œ(𝐷) .
In this case we can reduce our optimization problem to the class
(ii)
How can we establish the condition needed for (i)? One possibility
is to show that
𝐿
is uniformly bounded away from zero; then the condition
follows (see the argument in the next proof ).
when we try to apply this.
for
𝐿
Of course,
𝐿
is not known
However, [26, Ÿ4.2] gives veriable conditions
to be (bounded and) bounded away from zero.
C = ℝ𝑑 ,
They are stated for
𝑑
𝐿ℝ is the
𝑑
ℝ𝑑
opportunity process corresponding to C = ℝ , the actual 𝐿 satises 𝐿 β‰₯ 𝐿
the unconstrained case
but can be used nevertheless: if
because the supremum in (2.3) is taken over a smaller set in the constrained
case.
In the situation where
β„“
and
πΏβˆ’1
are bounded, we can also use the fol-
lowing result. Note also its use in Remark 3.3(ii) and recall that
1/0 := ∞.
Corollary 5.4. Let 𝑝 < 0 and let (β„“, πœ‹Λ‡ , πœ…Λ‡) be a solution of the Bellman
equation. Let 𝐿 be the opportunity process and assume that β„“/𝐿 is uniformly
bounded. Then (Λ‡πœ‹, πœ…Λ‡ ) is optimal and β„“ = 𝐿.
Proof.
(
𝐿 𝑝1
(πœ‹, πœ…) ∈ π’œπ‘“ 𝐸
π‘ˆπ‘  (πœ…π‘  𝑋𝑠 ) πœ‡(𝑑𝑠) is
Fix arbitrary
𝑋(πœ‹, πœ…)
)𝑝
+
∫
sition 2.2, hence of class (D). Since
and let
𝑋 = 𝑋(πœ‹, πœ…).
The process
a negative supermartingale by Propo-
∫
π‘ˆπ‘  (πœ…π‘  𝑋𝑠 ) πœ‡(𝑑𝑠) is decreasing and its
π’œπ‘“ 𝐸 ), 𝐿 𝑝1 𝑋 𝑝 is also of class (D).
of class (D), and then so is 𝑍(πœ‹, πœ…).
terminal value is integrable (denition of
The assumption yields that
β„“ 𝑝1 𝑋 𝑝
is
As bounded solutions are of special interest in BSDE theory, let us note
the following consequence.
23
Corollary 5.5. Let 𝑝 < 0. Under Assumptions 3.1 the following are equivalent:
(i) 𝐿 is bounded and bounded away from zero,
(ii) there exists a unique bounded solution of the Bellman equation, and
this solution is bounded away from zero.
One can note that in the setting of [26, Ÿ4.2], these conditions are further
equivalent to a reverse Hölder inequality for the market model.
We give an illustration of Theorem 5.2 also for the case
far, we have considered only the given exponent
many situations, there will exist some
the exponent
𝑝0
instead of
𝑝,
𝑝0 ∈ (𝑝, 1)
𝑝
𝑝 ∈ (0, 1).
Thus
and assumed (2.2).
In
such that, if we consider
the utility maximization problem is still nite.
Note that by Jensen's inequality this is a stronger assumption. We dene
for
π‘ž0 β‰₯ 1
the class of semimartingales
β„“
πΏπ‘ž0 (𝑃 ),
bounded in
B(π‘ž0 ) := {β„“ : sup𝜏 βˆ₯β„“πœ βˆ₯πΏπ‘ž0 (𝑃 ) < ∞},
where the supremum ranges over all stopping times
𝜏.
Corollary 5.6. Let 𝑝 ∈ (0, 1) and let there be a constant π‘˜1 > 0 such that
𝐷 β‰₯ π‘˜1 . Assume that the utility maximization problem is nite for some
𝑝0 ∈ (𝑝, 1) and let π‘ž0 β‰₯ 1 be such that π‘ž0 > 𝑝0 /(𝑝0 βˆ’ 𝑝). If (β„“, πœ‹
Λ‡, πœ…
Λ‡ ) is a
solution of the Bellman equation (for 𝑝) with β„“ ∈ B(π‘ž0 ), then β„“ = 𝐿 and
(Λ‡
πœ‹, πœ…
Λ‡ ) is optimal.
Proof.
Let
𝑋(Λ‡
πœ‹, πœ…
Λ‡ ).
β„“ ∈ B(π‘ž0 )
be a solution,
(Λ‡
πœ‹, πœ…
Λ‡)
the associated strategy, and
Λ‡ =
𝑋
By Theorem 5.2 and an argument as in the previous proof, it suces
to show that
ˇ𝑝
ℓ𝑋
is of class (D). Let
For every stopping time
𝜏,
𝛿>1
be such that
𝛿/π‘ž0 + 𝛿𝑝/𝑝0 = 1.
Hölder's inequality yields
Λ‡ πœπ‘ )𝛿 ] = 𝐸[(β„“π‘žπœ0 )𝛿/π‘ž0 (𝑋
Λ‡ πœπ‘0 )𝛿𝑝/𝑝0 ] ≀ 𝐸[β„“π‘žπœ0 ]𝛿/π‘ž0 𝐸[𝑋
Λ‡ πœπ‘0 ]𝛿𝑝/𝑝0 .
𝐸[(β„“πœ 𝑋
We show that this is bounded uniformly in
is bounded in
𝐿𝛿 (𝑃 )
Λ‡ πœπ‘ : 𝜏
𝜏 ; then {β„“πœ 𝑋
and hence uniformly integrable.
stopping time}
Indeed,
𝐸[β„“π‘žπœ0 ]
is
bounded by assumption. The set of wealth processes corresponding to admis-
Λ‡ πœπ‘0 ] ≀ 𝑒(𝑝0 ) (π‘₯0 ),
𝐸[𝐷𝑇 𝑝10 𝑋
problem with exponent 𝑝0 .
sible strategies is stable under stopping. Therefore
the value function for the utility maximization
The result follows as
Remark 5.7.
𝐷𝑇 β‰₯ π‘˜ 1 .
In [26, Example 4.6] we give a condition which implies that
all 𝑝0 ∈ (0, 1).
𝐿 ∈ B(𝑝0 /𝑝) if 𝐷
the utility maximization problem is nite for
given such a
𝑝0 ∈ (𝑝, 1),
one can show that
Conversely,
is uniformly
bounded from above (see [27, Corollary 4.2]).
Example 5.8.
We apply our results in an Itô model with bounded mean
variance tradeo process together with an existence result for BSDEs. For
24
the case of utility from terminal wealth only, we retrieve (a minor generalization of ) the pioneering result of [14, Ÿ3]; the case with intermediate consumption is new. Let
(π‘š
β‰₯ 𝑑)
π‘Š
and assume that
be an
𝔽
π‘š-dimensional
π‘Š.
is generated by
standard Brownian motion
We consider
𝑑𝑅𝑑 = 𝑏𝑑 𝑑𝑑 + πœŽπ‘‘ π‘‘π‘Šπ‘‘ ,
ℝ𝑑×π‘š -valued with
everywhere full rank; moreover, we consider constraints C satisfying (C1)
and (C2). We are in the situation of Assumptions 3.3 with 𝑑𝑀 = 𝜎 π‘‘π‘Š and
πœ† = (𝜎𝜎 ⊀ )βˆ’1 𝑏. The process πœƒ := 𝜎 ⊀ πœ† is called market price of risk. We
assume that there are constants π‘˜π‘– > 0 such that
∫ 𝑇
βˆ£πœƒπ‘  ∣2 𝑑𝑠 ≀ π‘˜3 .
0 < π‘˜1 ≀ 𝐷 ≀ π‘˜2 and
where
𝑏
is predictable
ℝ𝑑 -valued
and
𝜎
is predictable
0
The latter condition is called
bounded mean-variance tradeo.
Note that
𝑑𝑄/𝑑𝑃 = β„°(βˆ’πœ† 𝑀 )𝑇 = β„°(βˆ’πœƒ π‘Š )𝑇 denes a local martingale measure
for β„°(𝑅). By [26, Ÿ4.2] the utility maximization problem is nite for all 𝑝
and the opportunity process 𝐿 is bounded and bounded away from zero. It
βˆ™
βˆ™
is continuous due to Remark 3.13(i).
As suggested above, we write the Bellman BSDE for
π΄π‘Œ
πœ‘π‘Œ
π‘Œ := log(𝐿)
rather
than 𝐿 in this setting. If π‘Œ =
+
𝑀 + 𝑁 π‘Œ is the Kunita-Watanabe
⊀
π‘Œ
βŠ₯
βŠ₯ βˆ™ π‘Š = π‘π‘Œ
decomposition, we write 𝑍 := 𝜎 πœ‘ and choose 𝑍 such that 𝑍
βˆ™
by Brownian representation. The orthogonality of the decomposition implies
πœŽβŠ€π‘ βŠ₯ = 0
consumption and
(with
𝑍 ⊀ 𝑍 βŠ₯ = 0. We write 𝛿 = 1 if there is intermediate
𝛿 = 0 otherwise. Then Itô's formula and Corollary 3.12
and that
𝐴𝑑 := 𝑑)
yield the BSDE
π‘‘π‘Œ = 𝑓 (π‘Œ, 𝑍, 𝑍 βŠ₯ ) 𝑑𝑑 + (𝑍 + 𝑍 βŠ₯ ) π‘‘π‘Š ;
π‘Œπ‘‡ = log(𝐷𝑇 )
(5.1)
with
(
)
𝑓 (π‘Œ, 𝑍, 𝑍 βŠ₯ ) = 21 𝑝(1 βˆ’ 𝑝) 𝑑2𝜎⊀ C 𝛽(πœƒ + 𝑍) + 2π‘ž βˆ£πœƒ + π‘βˆ£2
(
)
+ 𝛿(𝑝 βˆ’ 1)𝐷𝛽 exp (π‘ž βˆ’ 1)π‘Œ βˆ’ 12 (βˆ£π‘βˆ£2 + βˆ£π‘ βŠ₯ ∣2 ).
𝛽 = (1 βˆ’ 𝑝)βˆ’1
π‘ž = 𝑝/(𝑝 βˆ’ 1); the dependence on (πœ”, 𝑑) is suppressed
2
in the notation. Using the orthogonality relations and 𝑝(1 βˆ’ 𝑝)𝛽 = βˆ’π‘ž , one
βŠ₯
βŠ₯
βŠ₯
˜
˜
can check that 𝑓 (π‘Œ, 𝑍, 𝑍 ) = 𝑓 (π‘Œ, 𝑍 +𝑍 , 0) =: 𝑓 (π‘Œ, 𝑍), where 𝑍 := 𝑍 +𝑍 .
2
2
As 0 ∈ C , we have 𝑑 ⊀ (π‘₯) ≀ ∣π‘₯∣ . Hence there exist a constant 𝐢 > 0 and
𝜎 C
an increasing continuous function πœ™ such that
(
)
βˆ£π‘“ (𝑦, π‘§Λœ)∣ ≀ 𝐢 βˆ£πœƒβˆ£2 + πœ™(𝑦) + ∣˜
𝑧 ∣2 .
Here
and
The following monotonicity property handles the exponential nonlinearity
𝑝 βˆ’ 1 < 0 and π‘ž βˆ’ 1 < 0,
[
]
βˆ’π‘¦ 𝑓 (𝑦, π‘§Λœ) βˆ’ 𝑓 (0, π‘§Λœ) ≀ 0.
caused by the consumption: As
25
Thus we have Briand and Hu's [4, Condition (A.1)] after noting that they
call
βˆ’π‘“
what we call
and [4, Lemma 2] states the existence of a bounded
β„“ := exp(π‘Œ ) is the op(Λ‡
πœ‹, πœ…
Λ‡ ) by πœ…
Λ‡ := (𝐷/β„“)𝛽
and Proposition 4.3; then we have a solution (β„“, πœ‹
Λ‡, πœ…
Λ‡ ) of the Bellman equation in the sense of Denition 4.1. For 𝑝 < 0 (𝑝 ∈ (0, 1)), Corollary 5.4
(Corollary 5.6) yields β„“ = 𝐿 and the optimality of (Λ‡
πœ‹, πœ…
Λ‡ ). In fact, the same
verication argument applies if we replace πœ‹
Λ‡ by any other predictable C βˆ—
⊀ βˆ—
𝜎 ⊀ C {𝛽(πœƒ + 𝑍)}; recall from Proposition 4.3
valued πœ‹ such that 𝜎 πœ‹ ∈ Ξ 
βˆ—
that πœ‹ ∈ 𝐿(𝑅) holds automatically. To conclude: we have that
solution
π‘Œ
𝑓,
to the BSDE (5.1).
Let us check that
portunity process. We dene an associated strategy
𝐿 = exp(π‘Œ )
is the opportunity process
(πœ‹ βˆ— , πœ…
Λ†)
and the set of optimal strategies equals the set of all
βˆ™ πœ…
Λ†=
βˆ™ πœ‹βˆ—
(𝐷/𝐿)𝛽
is predictable,
C -valued
and
𝜎 ⊀ πœ‹ βˆ— ∈ Π𝜎
⊀C
{𝛽(πœƒ + 𝑍)} 𝑃 βŠ— 𝑑𝑑-a.e.
One can remark that the previous arguments show
π‘Œ
such that
πœ‡βˆ˜ -a.e.,
π‘Œ β€² = log(𝐿)
whenever
β€² is a solution of the BSDE (5.1) which is uniformly bounded from above.
Hence we have proved uniqueness for (5.1) in this class of solutions, which is
not immediate from BSDE theory. One can also note that, in contrast to [14],
we did not use the theory of
𝐡𝑀 𝑂
martingales in this example. Finally, we
remark that the existence of an optimal strategy can also be obtained by
convex duality, under the additional assumption that
C
is convex.
We close this section with a formula intended for future applications.
Remark 5.9.
(β„“, πœ‹
Λ‡, πœ…
Λ‡ ) be a solution of the Bellman equation. Sometimes
𝑍(Λ‡
πœ‹, πœ…
Λ‡ ) is of class (D).
⊀
Let β„Ž be a predictable cut-o function such that πœ‹
Λ‡ β„Ž(π‘₯) is bounded, e.g.,
β„Ž(π‘₯) = π‘₯1{∣π‘₯βˆ£β‰€1}∩{βˆ£Λ‡πœ‹βŠ€ π‘₯βˆ£β‰€1} , and dene Ξ¨ to be the local martingale
Let
exponential formulas can be used to verify that
βˆ™ 𝑀 β„“ + 𝑝ˇ
β„“βˆ’1
πœ‹ βˆ™ 𝑅𝑐 + 𝑝ˇ
πœ‹ ⊀ β„Ž(π‘₯) βˆ— (πœ‡π‘… βˆ’ 𝜈 𝑅 ) + 𝑝(π‘₯β€² /β„“βˆ’ )Λ‡
πœ‹ ⊀ β„Ž(π‘₯) βˆ— (πœ‡π‘…,β„“ βˆ’ 𝜈 𝑅,β„“ )
βˆ’
{
}
+ (1 + π‘₯β€² /β„“βˆ’ ) (1 + πœ‹
Λ‡ ⊀ π‘₯)𝑝 βˆ’ 1 βˆ’ 𝑝ˇ
πœ‹ ⊀ β„Ž(π‘₯) βˆ— (πœ‡π‘…,β„“ βˆ’ 𝜈 𝑅,β„“ ).
Then
β„°(Ξ¨) > 0,
Proof.
Let
and if
β„°(Ξ¨)
𝑍 = 𝑍(Λ‡
πœ‹, πœ…
Λ‡ ).
is of class (D), then
𝑍(Λ‡
πœ‹, πœ…
Λ‡)
By a calculation as in the proof of Lemma 3.4 and
the local martingale condition from Proposition 4.7,
Hence
𝑍 = 𝑍0 β„°(Ξ¨)
is also of class (D).
Λ‡ 𝑝 )βˆ’1 βˆ™ 𝑍 = β„“βˆ’ βˆ™ Ξ¨.
( 𝑝1 𝑋
βˆ’
in the case without intermediate consumption.
For
𝑍 is of
1 ˇ𝑝
π‘βˆ’1 = β„“ /𝐷
class (D) whenever β„“ 𝑋
is. Writing the denition of πœ…
Λ‡
as πœ…
Λ‡
βˆ’
𝑝
∫
ˇ𝑝 = 𝑍 βˆ’ πœ…
Λ‡ 𝑝 π‘‘πœ‡ = (β„“βˆ’ 1 𝑋
Λ‡ 𝑝 ) βˆ™ (Ξ¨ βˆ’ πœ…
βˆ™ πœ‡), hence
πœ‡-a.e., we have β„“ 𝑝1 𝑋
Λ‡ β„“βˆ’ 𝑝1 𝑋
Λ‡
𝑝 βˆ’
Λ‡ 𝑝 = 𝑍0 β„°(Ξ¨ βˆ’ πœ…
β„“ 𝑝1 𝑋
Λ‡ βˆ™ πœ‡) = 𝑍0 β„°(Ξ¨) exp(βˆ’Λ‡
πœ… βˆ™ πœ‡). It remains to note that
exp(βˆ’Λ‡
πœ… βˆ™ πœ‡) ≀ 1.
the general case, we have seen in the proof of Corollary 5.4 that
26
5.2 Verication via Deator
The goal of this section is a verication theorem which involves only the candidate for the optimal strategy and holds for general semimartingale models.
Our plan is as follows. Let
(β„“, πœ‹
Λ‡, πœ…
Λ‡)
C
and assume for the moment that
has a maximum at
πœ‹
Λ‡,
be a solution of the Bellman equation
is convex. As the concave function
the directional derivatives at
πœ‹
Λ‡
𝑔ℓ
in all directions should
be nonpositive (if they can be dened). A calculation will show that, at the
level of processes, this yields a supermartingale property which is well known
from duality theory and allows for verication.
In the case of non-convex
constraints, the directional derivatives need not be dened in any sense. Nevertheless, the formally corresponding quantities yield the expected result. To
make the rst order conditions necessary, we later specialize to convex
C.
As
in the previous section, we rst state a basic result; it is essentially classical.
Lemma 5.10. Let β„“ be any positive càdlàg semimartingale with ℓ𝑇 = 𝐷𝑇 .
Suppose there exists (Λ‡πœ‹, πœ…Λ‡ ) ∈ π’œ with πœ…Λ‡ = (𝐷/β„“)𝛽 and let 𝑋ˇ := 𝑋(Λ‡πœ‹, πœ…Λ‡ ).
Assume π‘Œ := ℓ𝑋ˇ π‘βˆ’1 has the property that for all (πœ‹, πœ…) ∈ π’œ,
∫
Ξ“(πœ‹, πœ…) := 𝑋(πœ‹, πœ…)π‘Œ +
πœ…π‘  𝑋𝑠 (πœ‹, πœ…)π‘Œπ‘  πœ‡(𝑑𝑠)
is a supermartingale. Then Ξ“(Λ‡πœ‹, πœ…Λ‡ ) is a martingale if and only if
and (Λ‡πœ‹, πœ…Λ‡ ) is optimal and β„“ = 𝐿.
Proof. β‡’:
(2.2)
holds
Λ‡ . Note
(πœ‹, πœ…) ∈ π’œ and denote 𝑐 = πœ…π‘‹(πœ‹, πœ…) and 𝑐ˇ = πœ…
ˇ𝑋
π‘βˆ’1
π‘βˆ’1
π‘βˆ’1
Λ‡
Λ‡
the partial derivative βˆ‚π‘ˆ (Λ‡
𝑐) = 𝐷ˇ
πœ… 𝑋
= ℓ𝑋
= π‘Œ . Concavity of π‘ˆ
implies π‘ˆ (𝑐) βˆ’ π‘ˆ (Λ‡
𝑐) ≀ βˆ‚π‘ˆ (Λ‡
𝑐)(𝑐 βˆ’ 𝑐ˇ) = π‘Œ (𝑐 βˆ’ 𝑐ˇ), hence
𝐸
[∫
𝑇
Let
]
[∫
π‘ˆπ‘  (𝑐𝑠 ) πœ‡βˆ˜ (𝑑𝑠) βˆ’ 𝐸
0
𝑇
𝑇
]
[∫
π‘ˆπ‘  (Λ‡
𝑐𝑠 ) πœ‡βˆ˜ (𝑑𝑠) ≀ 𝐸
0
]
π‘Œπ‘  (𝑐𝑠 βˆ’ 𝑐ˇ𝑠 ) πœ‡βˆ˜ (𝑑𝑠)
0
= 𝐸[Γ𝑇 (πœ‹, πœ…)] βˆ’ 𝐸[Γ𝑇 (Λ‡
πœ‹, πœ…
Λ‡ )].
Let
Ξ“(Λ‡
πœ‹, πœ…
Λ‡)
be a martingale; then
Ξ“0 (πœ‹, πœ…) = Ξ“0 (Λ‡
πœ‹, πœ…
Λ‡)
and the supermartin-
(πœ‹, πœ…) was arbitrary,
[βˆ«π‘‡
]
(Λ‡
πœ‹, πœ…
Λ‡ ) is optimal with expected utility 𝐸 0 π‘ˆπ‘  (Λ‡
𝑐𝑠 ) πœ‡βˆ˜ (𝑑𝑠) = 𝐸[ 𝑝1 Γ𝑇 (Λ‡
πœ‹, πœ…
Λ‡ )] =
1
1 𝑝
πœ‹, πœ…
Λ‡ ) = 𝑝 π‘₯0 β„“0 < ∞. The rest is as in the proof of Lemma 5.1.
𝑝 Ξ“0 (Λ‡
gale property imply that the last line is nonpositive. As
The process
π‘Œ
is a supermartingale deator in the language of [19]. We
refer to [26] for the connection of the opportunity process with convex duality, which in fact suggests Lemma 5.10. Note that unlike
previous section,
Ξ“(πœ‹, πœ…)
is positive for all values of
𝑍(πœ‹, πœ…)
from the
𝑝.
Our next goal is to link the supermartingale property to local rst order
conditions. Let
𝑦, 𝑦ˇ ∈ C ∩C 0 (we will plug in πœ‹
Λ‡ for 𝑦ˇ). The formal directional
𝑦ˇ in the direction of 𝑦 is (𝑦 βˆ’ 𝑦ˇ)⊀ βˆ‡π‘” β„“ (Λ‡
𝑦 ) = 𝐺ℓ (𝑦, 𝑦ˇ), where,
β„“
derivative of 𝑔 at
27
by formal dierentiation under the integral sign (cf. (3.2))
𝐺ℓ (𝑦, 𝑦ˇ) :=
(5.2)
(
𝑅ℓ
) ∫
𝑅
+ (𝑝 βˆ’ 1)𝑐 𝑦ˇ +
(𝑦 βˆ’ 𝑦ˇ)⊀ π‘₯β€² β„Ž(π‘₯) 𝐹 𝑅,β„“ (𝑑(π‘₯, π‘₯β€² ))
β„“βˆ’ (𝑦 βˆ’ 𝑦ˇ)⊀ 𝑏𝑅 + π‘β„“βˆ’
ℝ𝑑 ×ℝ
∫
{
}
+
(β„“βˆ’ + π‘₯β€² ) (1 + π‘¦Λ‡βŠ€ π‘₯)π‘βˆ’1 (𝑦 βˆ’ 𝑦ˇ)⊀ π‘₯ βˆ’ (𝑦 βˆ’ 𝑦ˇ)⊀ β„Ž(π‘₯) 𝐹 𝑅,β„“ (𝑑(π‘₯, π‘₯β€² )).
ℝ𝑑 ×ℝ
We take this expression as the denition of
𝐺ℓ (𝑦, 𝑦ˇ) whenever the last integral
is well dened (the rst one is nite by (4.2)). The dierentiation cannot be
justied in general, but see the subsequent section.
Lemma 5.11. Let 𝑦 ∈ C 0 and 𝑦ˇ ∈ C 0,βˆ— ∩ {𝑔ℓ > βˆ’βˆž}. Then 𝐺ℓ (𝑦, 𝑦ˇ) is well
dened with values in (βˆ’βˆž, ∞] and 𝐺ℓ (β‹…, 𝑦ˇ) is lower semicontinuous on C 0 .
Proof.
(𝑦 βˆ’ 𝑦ˇ)⊀ π‘₯ = 1 + 𝑦 ⊀ π‘₯ βˆ’ (1 + π‘¦Λ‡βŠ€ π‘₯), we can express 𝐺ℓ (𝑦, 𝑦ˇ) as
(
) ∫
⊀ 𝑅
𝑅
𝑐𝑅ℓ
β„“βˆ’ (𝑦 βˆ’ 𝑦ˇ) 𝑏 + β„“βˆ’ + (𝑝 βˆ’ 1)𝑐 𝑦ˇ +
(𝑦 βˆ’ 𝑦ˇ)⊀ π‘₯β€² β„Ž(π‘₯) 𝐹 𝑅,β„“ (𝑑(π‘₯, π‘₯β€² ))
Writing
ℝ𝑑 ×ℝ
∫
π‘¦βŠ€π‘₯
}
1+
⊀
βˆ’
1
βˆ’
(𝑦
+
(𝑝
βˆ’
1)Λ‡
𝑦
)
β„Ž(π‘₯)
𝐹 𝑅,β„“ (𝑑(π‘₯, π‘₯β€² ))
(1 + π‘¦Λ‡βŠ€ π‘₯)1βˆ’π‘
ℝ𝑑 ×ℝ
∫
{
}
βˆ’
(β„“βˆ’ + π‘₯β€² ) (1 + π‘¦Λ‡βŠ€ π‘₯)𝑝 βˆ’ 1 βˆ’ 𝑝ˇ
𝑦 ⊀ β„Ž(π‘₯) 𝐹 𝑅,β„“ (𝑑(π‘₯, π‘₯β€² )).
+
{
(β„“βˆ’ + π‘₯β€² )
ℝ𝑑 ×ℝ
𝑦 by (4.2). The last integral
𝑔 β„“ (Λ‡
𝑦 ), cf. (3.2), and it is nite if 𝑔 β„“ (Λ‡
𝑦 ) > βˆ’βˆž
The rst integral is nite and continuous in
above occurs in the denition of
and equals
+∞
otherwise. Finally, consider the second integral above and
⊀
1+𝑦 π‘₯
πœ“ = πœ“(𝑦, 𝑦ˇ, π‘₯, π‘₯β€² ). The Taylor expansion (1+Λ‡
=
𝑦 ⊀ π‘₯)1βˆ’π‘
)⊀
(π‘βˆ’1) (
⊀
⊀
3
1∫ + (𝑦 + (𝑝 βˆ’ 1)Λ‡
𝑦 ) π‘₯ + 2 2𝑦 + (𝑝 βˆ’ 2)Λ‡
𝑦 π‘₯ π‘₯ 𝑦ˇ + π‘œ(∣π‘₯∣ ) shows that
𝑅,β„“
is well dened and nite. It also shows that given a
{∣π‘₯∣+∣π‘₯β€² βˆ£β‰€1} πœ“ 𝑑𝐹
∫
𝑑
𝑅,β„“ is continuous
compact 𝐾 βŠ‚ ℝ , there is πœ€ > 0 such that
{∣π‘₯∣+∣π‘₯β€² βˆ£β‰€πœ€} πœ“ 𝑑𝐹
in 𝑦 ∈ 𝐾 (and also in 𝑦
Λ‡ ∈ 𝐾 ). The details are as in Lemma A.2. Moreover,
0
β€²
for 𝑦 ∈ C we have the lower bound πœ“ β‰₯ (β„“βˆ’ +π‘₯ ){βˆ’1βˆ’(𝑦 +(π‘βˆ’1)Λ‡
𝑦 )⊀ β„Ž(π‘₯)},
𝑅,β„“
β€²
which is 𝐹
-integrable on {∣π‘₯∣ + ∣π‘₯ ∣ > πœ€} for any πœ€ > 0, again by (4.2).
call its integrand
The result now follows by Fatou's lemma.
We can now connect the local rst order conditions for
𝑔ℓ
and the global
supermartingale property: it turns out that the formal derivative
mines the sign of the drift rate of
Ξ“
𝐺ℓ
deter-
(cf. (5.3) below), which leads to the
following proposition. Here and in the sequel, we denote
Λ‡ = 𝑋(Λ‡
𝑋
πœ‹, πœ…
Λ‡ ).
Proposition 5.12. Let (β„“, πœ‹Λ‡ , πœ…Λ‡) be a solution∫ of the Bellman equation and
Then Ξ“(πœ‹, πœ…) := ℓ𝑋ˇ π‘βˆ’1 𝑋(πœ‹, πœ…) + πœ…π‘  ℓ𝑠 𝑋ˇ π‘ π‘βˆ’1 𝑋𝑠 (πœ‹, πœ…) πœ‡(𝑑𝑠) is a
supermartingale (local martingale) if and only if 𝐺ℓ (πœ‹, πœ‹Λ‡ ) ≀ 0 (= 0).
(πœ‹, πœ…) ∈ π’œ.
28
Proof.
¯ = 𝑅 βˆ’ (π‘₯ βˆ’ β„Ž(π‘₯)) βˆ— πœ‡π‘… as in (2.4). We abbreviate πœ‹
𝑅
¯ :=
(𝑝 βˆ’ 1)Λ‡
πœ‹ + πœ‹ and similarly πœ…
¯ := (𝑝 βˆ’ 1)Λ‡
πœ… + πœ…. We defer to Lemma C.1 a
( π‘βˆ’1
)βˆ’1 ( π‘βˆ’1
)
Λ‡
Λ‡
βˆ™ ℓ𝑋
calculation showing that 𝑋
𝑋(πœ‹, πœ…) equals
βˆ’ π‘‹βˆ’ (πœ‹, πœ…)
Dene
(
)⊀
¯ βˆ’ β„“βˆ’ πœ…
β„“ βˆ’ β„“0 + β„“βˆ’ πœ‹
¯βˆ™π‘…
¯ βˆ™ πœ‡ + β„“βˆ’ (𝑝 βˆ’ 1) π‘βˆ’2
Λ‡ + πœ‹ 𝑐𝑅 πœ‹
Λ‡ βˆ™π΄+πœ‹
¯ ⊀ 𝑐𝑅ℓ βˆ™ 𝐴
2 πœ‹
{
}
+πœ‹
¯ ⊀π‘₯β€² β„Ž(π‘₯) βˆ— πœ‡π‘…,β„“ + (β„“βˆ’ + π‘₯β€² ) (1 + πœ‹
Λ‡ ⊀ π‘₯)π‘βˆ’1 (1 + πœ‹ ⊀ π‘₯) βˆ’ 1 βˆ’ πœ‹
¯ ⊀ β„Ž(π‘₯) βˆ— πœ‡π‘…,β„“ .
β„Ž such
(β„“, πœ‹
Λ‡, πœ…
Λ‡ ) is a
Here we use a predictable cut-o function
e.g.,
β„Ž(π‘₯) = π‘₯1{∣π‘₯βˆ£β‰€1}∩{∣¯πœ‹βŠ€ π‘₯βˆ£β‰€1} .
Since
that
πœ‹
¯ ⊀ β„Ž(π‘₯)
is bounded;
solution, the drift of
β„“
is
𝐴ℓ = βˆ’π‘π‘ˆ βˆ— (β„“βˆ’ ) βˆ™ πœ‡ βˆ’ 𝑝𝑔 β„“ (Λ‡
πœ‹ ) βˆ™ 𝐴 = (𝑝 βˆ’ 1)β„“βˆ’ πœ…
Λ‡ βˆ™ πœ‡ βˆ’ 𝑝𝑔 β„“ (Λ‡
πœ‹ ) βˆ™ 𝐴.
By Remark 2.3,
(βˆ’βˆž, ∞].
Ξ“ := Ξ“(πœ‹, πœ…)
has a well dened drift rate
π‘ŽΞ“
with values in
From the two formulas above and (2.4) we deduce
Λ‡ π‘βˆ’1 𝑋(πœ‹, πœ…)βˆ’ 𝐺ℓ (πœ‹, πœ‹
π‘ŽΞ“ = 𝑋
Λ‡ ).
βˆ’
(5.3)
Λ‡ π‘βˆ’1 𝑋(πœ‹, πœ…)βˆ’ > 0 by admissibility. If Ξ“ is a supermartingale,
𝑋
βˆ’
π‘ŽΞ“ ≀ 0, and the converse holds by Lemma 2.4 in view of Ξ“ β‰₯ 0.
Here
then
We obtain our second verication theorem from Proposition 5.12 and
Lemma 5.10.
Theorem 5.13. Let (β„“, πœ‹Λ‡ , πœ…Λ‡) be a solution of the Bellman equation. Assume
that 𝑃 βŠ— 𝐴-a.e., 𝐺ℓ (𝑦, πœ‹Λ‡ ) ∈ [βˆ’βˆž, 0] for all 𝑦 ∈ C ∩ C 0,βˆ— . Then
ˇ𝑝 +
Ξ“(Λ‡
πœ‹, πœ…
Λ‡ ) := ℓ𝑋
∫
Λ‡ 𝑠𝑝 πœ‡(𝑑𝑠)
πœ…
Λ‡ 𝑠 ℓ𝑠 𝑋
is a local martingale. It is a martingale if and only if
is optimal and β„“ = 𝐿 is the opportunity process.
If
C
(2.2)
holds and (Λ‡πœ‹, πœ…Λ‡ )
is not convex, one can imagine situations where the directional
derivative of
𝑔ℓ
at the maximum is positivei.e., the assumption on
𝐺ℓ (𝑦, πœ‹
Λ‡)
is sucient but not necessary. This changes in the subsequent section.
5.2.1 The Convex-Constrained Case
We assume in this section that
C
C ∩ C 0 is also convex. Our
β„“
condition on 𝐺 in Theorem 5.13 is
is convex; then
aim is to show that the nonnegativity
automatically satised in this case. We start with an elementary but crucial
observation about dierentiation under the integral sign.
Lemma 5.14. Consider two distinct points 𝑦0 and 𝑦ˇ in ℝ𝑑 and let 𝐢 =
{πœ‚π‘¦0 + (1 βˆ’ πœ‚)Λ‡
𝑦 : 0 ≀ πœ‚ ≀ 1}. Let 𝜌 be a function on Ξ£ × πΆ , where Ξ£ is
some Borel space with measure 𝜈 , such that π‘₯ 7β†’ 𝜌(π‘₯, 𝑦) is 𝜈 -measurable,
29
∫
𝜌+ (π‘₯, β‹…) 𝜈(𝑑π‘₯) < ∞
directional derivative
on 𝐢 , and 𝑦 7β†’ 𝜌(π‘₯, 𝑦) is concave. In particular, the
(
)
𝜌 π‘₯, 𝑦ˇ + πœ€(𝑦 βˆ’ 𝑦ˇ) βˆ’ 𝜌(π‘₯, 𝑦ˇ)
𝐷𝑦ˇ,𝑦 𝜌(π‘₯, β‹…) := lim
πœ€β†’0+
πœ€
exists in (βˆ’βˆž, ∞] for all π‘¦βˆ« ∈ 𝐢 . Let 𝛼 be another concave function on 𝐢 .
Dene 𝛾(𝑦) := 𝛼(𝑦) + 𝜌(π‘₯, 𝑦) 𝜈(𝑑π‘₯) and assume that 𝛾(𝑦0 ) > βˆ’βˆž and
that 𝛾(ˇ𝑦) = max𝐢 𝛾 < ∞. Then for all 𝑦 ∈ 𝐢 ,
∫
𝐷𝑦ˇ,𝑦 𝛾 = 𝐷𝑦ˇ,𝑦 𝛼 +
𝐷𝑦ˇ,𝑦 𝜌(π‘₯, β‹…) 𝜈(𝑑π‘₯) ∈ (βˆ’βˆž, 0]
(5.4)
and in particular 𝐷𝑦ˇ,𝑦 𝜌(π‘₯, β‹…) < ∞ 𝜈(𝑑π‘₯)-a.e.
Proof.
𝛾 ∫> βˆ’βˆž on 𝐢 . Let 𝑣 =
𝑦)
(𝑦 βˆ’ 𝑦ˇ) and πœ€ >
=
+ 𝜌(π‘₯,ˇ𝑦+πœ€π‘£)βˆ’πœŒ(π‘₯,Λ‡
𝜈(𝑑π‘₯).
πœ€
By concavity, these quotients increase monotonically as πœ€ ↓ 0, in particular
their limits exist. The left hand side is nonpositive as 𝑦
Λ‡ is a maximum and
Note that
𝛾
is concave, hence we also have
𝑦)
0, then 𝛾(ˇ𝑦+πœ€π‘£)βˆ’π›Ύ(Λ‡
πœ€
𝛼(Λ‡
𝑦 +πœ€π‘£)βˆ’π›Ό(Λ‡
𝑦)
πœ€
monotone convergence yields (5.4).
For completeness, let us mention that if
where the left hand side of (5.4) is
βˆ’βˆž
𝛾(𝑦0 ) = βˆ’βˆž, there are examples
but the right hand side is nite;
we shall deal with this case separately. We deduce the following version of
Theorem 5.13; as discussed, it involves only the control
(Λ‡
πœ‹, πœ…
Λ‡ ).
Theorem 5.15. Let (β„“, πœ‹Λ‡ , πœ…Λ‡) be a solution of the Bellman
equation and as∫
sume that C is convex. Then Ξ“(Λ‡πœ‹, πœ…Λ‡ ) := ℓ𝑋ˇ 𝑝 + πœ…Λ‡ 𝑠 ℓ𝑠 𝑋ˇ 𝑠𝑝 πœ‡(𝑑𝑠) is a local
martingale. It is a martingale if and only if (2.2) holds and (Λ‡πœ‹, πœ…Λ‡ ) is optimal
and β„“ = 𝐿.
Proof.
𝐺ℓ (𝑦, πœ‹
Λ‡ ) ∈ [βˆ’βˆž, 0] for
𝐺ℓ was dened
β„“
Lemma 5.14 yields 𝐺 (𝑦, πœ‹
Λ‡) ≀ 0
β„“
proof for 𝑝 ∈ (0, 1) as 𝑔 is then
To apply Theorem 5.13, we have to check that
π‘¦βˆˆC ∩
C 0,βˆ— . Recall that
πœ‹
Λ‡
β„“
is a maximizer for 𝑔 and that
by dierentiation under the integral sign.
𝑦 ∈ {𝑔 β„“ > βˆ’βˆž}. This ends the
β„“
nite. If 𝑝 < 0, the denition of 𝑔 and Remark A.7 show that the set
βˆͺ
β„“
{𝑔 > βˆ’βˆž} contains the set πœ‚βˆˆ[0,1) πœ‚(C ∩ C 0 ) which, in turn, is clearly
0,βˆ— . Hence {𝑔 β„“ > βˆ’βˆž} is dense in C ∩ C 0,βˆ— and we obtain
dense in C ∩ C
β„“
𝐺 (𝑦, πœ‹
Λ‡ ) ∈ [βˆ’βˆž, 0] for all 𝑦 ∈ C ∩ C 0,βˆ— using the lower semicontinuity from
whenever
Lemma 5.11.
Remark 5.16.
(i)
Ξ“(Λ‡
πœ‹, πœ…
Λ‡ ) = 𝑝𝑍(Λ‡
πœ‹, πœ…
Λ‡ ) if 𝑍 is
can be used also for Ξ“(Λ‡
πœ‹, πœ…
Λ‡ ).
We note that
in (4.4). In particular, Remark 5.9
dened as
(ii) Muhle-Karbe [24] considers certain one-dimensional (unconstrained)
ane models and introduces a sucient optimality condition in the form of
an algebraic inequality (see [24, Theorem 4.20(3)]). This condition can be
seen as a special case of the statement that
in particular, we have shown its necessity.
30
𝐺𝐿 (𝑦, πœ‹
Λ‡ ) ∈ [βˆ’βˆž, 0]
for
𝑦 ∈ C 0,βˆ— ;
Of course, all our verication results can be seen as a uniqueness result
for the Bellman equation. As an example, Theorem 5.15 yields:
Corollary 5.17. If C is convex, there is at most one solution of the Bellman
equation in the class of solutions (β„“, πœ‹Λ‡ , πœ…Λ‡ ) such that Ξ“(Λ‡πœ‹, πœ…Λ‡ ) is of class (D).
Similarly, one can give corollaries for the other results. We close with a
comment concerning convex duality.
Remark 5.18.
(i)
A major insight in [21] was that the dual domain for
utility maximization (here with
C = ℝ𝑑 )
should be a set of supermartin-
gales rather than (local) martingales when the price process has jumps. A
one-period example for
log-utility
[21, Example 5.1'] showed that the su-
permartingale solving the dual problem can indeed have nonvanishing drift.
In that example it is clear that this arises when the budget constraint becomes binding.
phenomenon.
For general models and
log-utility,
[11] comments on this
The calculations of this section yield an instructive local
picture also for power utility.
𝐿 and the optimal strat(Λ†
πœ‹, πœ…
Λ† ) solve the Bellman equation. Assume that C is convex and let
Λ† = 𝑋(Λ†
Λ† π‘βˆ’1 , which was the solution to the dual
𝑋
πœ‹, πœ…
Λ† ). Consider π‘ŒΛ† = 𝐿𝑋
Λ† β„°(πœ‹ βˆ™ 𝑅) is a supermartingale for
problem in [26]. We have shown that π‘Œ
Λ†
every πœ‹ ∈ π’œ; i.e., π‘Œ is a supermartingale deator. Choosing πœ‹ = 0, we see
Λ† is itself a supermartingale, and by (5.3) its drift rate satises
that π‘Œ
Under Assumptions 3.1, the opportunity process
egy
Λ†
⊀
Λ† π‘βˆ’1 𝐺𝐿 (0, πœ‹
Λ† π‘βˆ’1 πœ‹
π‘Žπ‘Œ = 𝑋
Λ† ) = βˆ’π‘‹
πœ‹ ).
βˆ’
βˆ’ Λ† βˆ‡π‘”(Λ†
π‘ŒΛ† is a local martingale if and only if πœ‹
Λ† ⊀ βˆ‡π‘”(Λ†
πœ‹ ) = 0. One can say
⊀
that βˆ’Λ†
πœ‹ βˆ‡π‘”(Λ†
πœ‹ ) < 0 means that the constraints are binding, whereas in an
Λ† has nonvanishunconstrained case the gradient of 𝑔 would vanish; i.e., π‘Œ
ing drift rate at a given (πœ”, 𝑑) whenever the constraints are binding. Even
𝑑
0 in the maximization of
if C = ℝ , we still have the budget constraint C
𝑔 . If in addition 𝑅 is continuous, C 0 = ℝ𝑑 and we are truly in an unconΛ† is a local martingale; indeed, in the setting of
strained situation. Then π‘Œ
Hence
Corollary 3.12 we calculate
(
1
π‘ŒΛ† = 𝑦0 β„° βˆ’ πœ† βˆ™ 𝑀 +
πΏβˆ’
Note how
𝑁 𝐿,
the martingale part of
βˆ™
𝐿
)
𝑁𝐿 ,
𝑦0 := 𝐿0 π‘₯π‘βˆ’1
0 .
orthogonal to
𝑅,
yields the solution
to the dual problem.
(ii) From the proof of Proposition 5.12 we have that the general formula
for the local martingale part of
π‘€π‘Œ
Λ†
π‘ŒΛ†
is
(
¯
Λ† π‘βˆ’1 βˆ™ 𝑀 𝐿 + πΏβˆ’ (𝑝 βˆ’ 1)Λ†
πœ‹ ⊀ π‘₯β€² β„Ž(π‘₯) βˆ— (πœ‡π‘…,𝐿 βˆ’ 𝜈 𝑅,𝐿 )
=𝑋
πœ‹ βˆ™ 𝑀 𝑅 + (𝑝 βˆ’ 1)Λ†
βˆ’
)
{
}
β€²
⊀ π‘βˆ’1
⊀
𝑅,𝐿
𝑅,𝐿
+ (πΏβˆ’ + π‘₯ ) (1 + πœ‹
Λ† π‘₯)
βˆ’ 1 βˆ’ (𝑝 βˆ’ 1)Λ†
πœ‹ β„Ž(π‘₯) βˆ— (πœ‡
βˆ’πœˆ ) .
31
This is relevant in the problem of
π‘ž -optimal
equivalent martingale measures ;
cf. Goll and Rüschendorf [12] for a general perspective. Let
𝑒(π‘₯0 ) < ∞, 𝐷 ≑
1, πœ‡ = 0, C =
M of equivalent local martingale
𝑆 = β„°(𝑅) is nonempty. Given π‘ž = 𝑝/(𝑝 βˆ’ 1) ∈ (βˆ’βˆž, 0) βˆͺ (0, 1)
βˆ’1 (𝑑𝑄/𝑑𝑃 )π‘ž ] is nite and
conjugate to 𝑝, 𝑄 ∈ M is called π‘ž -optimal if 𝐸[βˆ’π‘ž
minimal over M . If π‘ž < 0, i.e., 𝑝 ∈ (0, 1), then 𝑒(π‘₯0 ) < ∞ is equivalent to
βˆ’1 (𝑑𝑄/𝑑𝑃 )π‘ž ] < ∞; moreover,
the existence of some 𝑄 ∈ M such that 𝐸[βˆ’π‘ž
ℝ𝑑 , and assume that the set
measures for
Assumptions 3.1 are satised (see Kramkov and Schachermayer [21, 22]).
Using [21, Theorem 2.2(iv)] we conclude that
(a) the
π‘ž -optimal martingale measure exists if and only if π‘Žπ‘Œ ≑ 0 and 𝑀 π‘Œ
Λ†
Λ†
is a true martingale;
1 + 𝑦0βˆ’1 𝑀 π‘Œ
(b) in that case,
Λ†
is its
𝑃 -density
process.
This generalizes earlier results of [12] as well as of Grandits [13], Jeanblanc
et al. [16] and Choulli and Stricker [6].
A Proof of Lemma 3.8: A Measurable Maximizing
Sequence
The main goal of this appendix is to construct a measurable maximizing
sequence for the random function
𝑔
(cf. Lemma 3.8). The entire section is
under Assumptions 3.1. Before beginning the proof, we discuss the properties
of
𝑔;
recall that
𝑔(𝑦) := πΏβˆ’ 𝑦
∫
+
⊀
(
𝑅
𝑏 +
𝑐𝑅𝐿
πΏβˆ’
+
(π‘βˆ’1) 𝑅
2 𝑐 𝑦
)
∫
+
π‘₯β€² 𝑦 ⊀ β„Ž(π‘₯) 𝐹 𝑅,𝐿 (𝑑(π‘₯, π‘₯β€² ))
ℝ𝑑 ×ℝ
{
}
(πΏβˆ’ + π‘₯β€² ) π‘βˆ’1 (1 + 𝑦 ⊀ π‘₯)𝑝 βˆ’ π‘βˆ’1 βˆ’ 𝑦 ⊀ β„Ž(π‘₯) 𝐹 𝑅,𝐿 (𝑑(π‘₯, π‘₯β€² )).
ℝ𝑑 ×ℝ
(A.1)
Lemma A.1.
Proof.
πΏβˆ’ + π‘₯β€²
is strictly positive 𝐹 𝐿 (𝑑π‘₯β€² )-a.e.
We have
[
]
(𝑃 βŠ— 𝜈 𝐿 ){πΏβˆ’ + π‘₯β€² ≀ 0} = 𝐸 1{πΏβˆ’ +π‘₯β€² ≀0} βˆ— πœˆπ‘‡πΏ
[
]
= 𝐸 1{πΏβˆ’ +π‘₯β€² ≀0} βˆ— πœ‡πΏ
𝑇
[βˆ‘
]
=𝐸
1{𝐿𝑠 ≀0} 1{Δ𝐿𝑠 βˆ•=0} ,
𝑠≀𝑇
which vanishes as
𝐿>0
by Lemma 2.1.
(πœ”, 𝑑) and let 𝑙 := πΏπ‘‘βˆ’ (πœ”). Furthermore, let 𝐹 be any Lévy measure
𝑅,𝐿
𝑑+1
on ℝ
which is equivalent to 𝐹𝑑
(πœ”) and satises (2.5). Equivalence
Fix
32
C𝑑0 (πœ”), C𝑑0,βˆ— (πœ”), and N𝑑 (πœ”) are the same if dened with respect
𝑅
to 𝐹 instead of 𝐹 . Given πœ€ > 0, let
∫
{
}
𝐹
πΌπœ€ (𝑦) :=
(𝑙 + π‘₯β€² ) π‘βˆ’1 (1 + 𝑦 ⊀ π‘₯)𝑝 βˆ’ π‘βˆ’1 βˆ’ 𝑦 ⊀ β„Ž(π‘₯) 𝐹 (𝑑(π‘₯, π‘₯β€² )),
{∣π‘₯∣+∣π‘₯β€² βˆ£β‰€πœ€}
∫
{
}
𝐹
(𝑙 + π‘₯β€² ) π‘βˆ’1 (1 + 𝑦 ⊀ π‘₯)𝑝 βˆ’ π‘βˆ’1 βˆ’ 𝑦 ⊀ β„Ž(π‘₯) 𝐹 (𝑑(π‘₯, π‘₯β€² )),
𝐼>πœ€
(𝑦) :=
implies that
{∣π‘₯∣+∣π‘₯β€² ∣>πœ€}
so that
𝐹
𝐼 𝐹 (𝑦) := πΌπœ€πΉ (𝑦) + 𝐼>πœ€
(𝑦)
is the last integral in (A.1) when
𝐹 = 𝐹𝑑𝑅,𝐿 (πœ”).
We know from the proof
𝐹 𝑅,𝐿 (πœ‹) is well dened and nite for any
of Lemma 3.4 that 𝐼
course, when
𝐹 , 𝐼𝐹
general
𝑝 > 0,
πœ‹ ∈ π’œπ‘“ 𝐸
this is essentially due to the assumption (2.2)).
(of
For
has the following properties.
Lemma A.2. Consider a sequence 𝑦𝑛 β†’ π‘¦βˆž in C 0 .
(i) For any 𝑦 ∈ C 0 , the integral 𝐼 𝐹 (𝑦) is well dened in ℝ βˆͺ {sign(𝑝)∞}.
(ii) For πœ€ ≀ (2 sup𝑛 βˆ£π‘¦π‘› ∣)βˆ’1 we have πΌπœ€πΉ (𝑦𝑛 ) β†’ πΌπœ€πΉ (π‘¦βˆž ).
(iii) If 𝑝 ∈ (0, 1), 𝐼 𝐹 is l.s.c., that is, lim inf 𝑛 𝐼 𝐹 (𝑦𝑛 ) β‰₯ 𝐼 𝐹 (π‘¦βˆž ).
(iv) If 𝑝 < 0, 𝐼 𝐹 is u.s.c., that is, lim sup𝑛 𝐼 𝐹 (𝑦𝑛 ) ≀ 𝐼 𝐹 (π‘¦βˆž ). Moreover,
𝑦 ∈ C 0 βˆ– C 0,βˆ— implies 𝐼 𝐹 (𝑦) = βˆ’βˆž.
Proof.
The rst item follows from the subsequent considerations.
(ii) We may assume that β„Ž is the identity on {∣π‘₯∣ ≀ πœ€}; then on this set
π‘βˆ’1 (1 + 𝑦 ⊀ π‘₯)𝑝 βˆ’ π‘βˆ’1 βˆ’ 𝑦 ⊀ β„Ž(π‘₯) =: πœ“(𝑧)βˆ£π‘§=π‘¦βŠ€ π‘₯ , where the function πœ“ is smooth
on {βˆ£π‘§βˆ£ ≀ 1/2} βŠ† ℝ satisfying
πœ“(𝑧) = π‘βˆ’1 (1 + 𝑧)𝑝 βˆ’ π‘βˆ’1 βˆ’ 𝑧 =
π‘βˆ’1 2
2 𝑧
+ π‘œ(βˆ£π‘§βˆ£3 )
˜
πœ“(𝑧) = 𝑧 2 πœ“(𝑧)
with a function
˜
πœ“ that is continuous and in particular bounded on {βˆ£π‘§βˆ£ ≀ 1/2}.
β€² 2
2
As a Lévy measure, 𝐹 integrates (∣π‘₯ ∣ + ∣π‘₯∣ ) on compacts; in particular,
β€²
2
β€²
𝐺(𝑑(π‘₯, π‘₯ )) := ∣π‘₯∣ 𝐹 (𝑑(π‘₯, π‘₯ )) denes a nite measure on {∣π‘₯∣ + ∣π‘₯β€² ∣ ≀ πœ€}.
𝐹
βˆ’1 , and dominated converHence πΌπœ€ (𝑦) is well dened and nite for βˆ£π‘¦βˆ£ ≀ (2πœ€)
∫
𝐹
β€²
⊀
β€²
˜
gence shows that πΌπœ€ (𝑦) =
{∣π‘₯∣+∣π‘₯β€² βˆ£β‰€πœ€} (𝑙+π‘₯ )πœ“(𝑦 π‘₯) 𝐺(𝑑(π‘₯, π‘₯ )) is continuous
βˆ’1 }.
in 𝑦 on {βˆ£π‘¦βˆ£ ≀ (2πœ€)
𝐹 is bounded
(iii) For βˆ£π‘¦βˆ£ bounded by a constant 𝐢 , the integrand in 𝐼
β€²
β€²
β€²
from below by 𝐢 + ∣π‘₯ ∣ for some constant 𝐢 depending on 𝑦 only through
𝐢 . We choose πœ€ as before. As 𝐢 β€² + ∣π‘₯β€² ∣ is 𝐹 -integrable on {∣π‘₯∣ + ∣π‘₯β€² ∣ > πœ€}
𝐹
by (2.5), 𝐼 (𝑦) is well dened in ℝ βˆͺ {∞} and l.s.c. by Fatou's lemma.
because
1+𝑧
is bounded away from
0.
Thus
(iv) The rst part follows as in (iii), now the integrand is bounded from
𝐢 β€² + ∣π‘₯β€² ∣. If 𝑦 ∈ C 0 βˆ– C 0,βˆ— , Lemma
βˆ’βˆž on a set of positive 𝐹 -measure.
above by
equals
33
A.1 shows that the integrand
Lemma A.3. The function 𝑔 is concave. If C is convex, 𝑔 has at most one
maximum on C ∩ C 0 , modulo N .
Proof. We rst remark that the assertion is not trivial because 𝑔 need not
be strictly concave on
N
βŠ₯ , for example, the process
𝑅𝑑 = 𝑑(1, . . . , 1)⊀
was
not excluded.
𝑔(𝑦) = 𝐻𝑦 + 𝐽(𝑦), where 𝐻𝑦 = πΏβˆ’ 𝑦 ⊀ 𝑏𝑅 +
⊀ 𝑅
𝐹 𝑅,𝐿 (𝑦) is
+
𝐽(𝑦) = (π‘βˆ’1)
2 πΏβˆ’ 𝑦 𝑐 𝑦 + 𝐼
concave. We may assume that β„Ž(π‘₯) = π‘₯1{∣π‘₯βˆ£β‰€1} .
0 be such that 𝑔(𝑦 ) = 𝑔(𝑦 ) = sup 𝑔 =: 𝑔 βˆ— < ∞,
Let 𝑦1 , 𝑦2 ∈ C ∩ C
1
2
βˆ—
our aim is to show 𝑦1 βˆ’ 𝑦2 ∈ N . By concavity, 𝑔 = 𝑔((𝑦1 + 𝑦2 )/2)) =
[𝑔(𝑦1 ) + 𝑔(𝑦2 )]/2, which implies 𝐽((𝑦1 + 𝑦2 )/2)) = [𝐽(𝑦1 ) + 𝐽(𝑦2 )]/2 due to
the linearity of 𝐻 . Using the denition of 𝐽 , this shows that 𝐽 is constant
on the line segment connecting 𝑦1 and 𝑦2 . A rst consequence is that 𝑦1 βˆ’ 𝑦2
}
⊀ 𝑅
𝑅
⊀
lies in the set {𝑦 : 𝑦 𝑐 = 0, 𝐹 {π‘₯ : 𝑦 π‘₯ βˆ•= 0} = 0 and a second is that
𝐻𝑦1 = 𝐻𝑦2 . It remains to show (𝑦1 βˆ’ 𝑦2 )⊀ 𝑏𝑅 = 0 to have 𝑦1 βˆ’ 𝑦2 ∈ N .
𝑅
⊀
𝑅,𝐿 {π‘₯ : 𝑦 ⊀ β„Ž(π‘₯) βˆ•= 0} = 0.
Note that 𝐹 {π‘₯ : 𝑦 π‘₯ βˆ•= 0} = 0 implies 𝐹
⊀
𝑅
⊀
𝑅𝐿
Moreover, 𝑦 𝑐
= 0 implies 𝑦 𝑐
= 0 due to the absolute continuity
𝑐,𝑖
𝑐
𝑐,𝑖
βŸ¨π‘… , 𝐿 ⟩ β‰ͺ βŸ¨π‘… ⟩ which follows from the Kunita-Watanabe
inequality.
∫ β€²
Therefore, the rst consequence above implies
π‘₯ (𝑦1 βˆ’ 𝑦2 )⊀ β„Ž(π‘₯) 𝐹 𝑅,𝐿 = 0
⊀ 𝑅𝐿 = 0, and now the second consequence and the denition
and (𝑦1 βˆ’ 𝑦2 ) 𝑐
⊀ 𝑅
⊀ 𝑅 = 0 as
of 𝐻 yield 0 = 𝐻(𝑦1 βˆ’ 𝑦2 ) = πΏβˆ’ (𝑦1 βˆ’ 𝑦2 ) 𝑏 . Thus (𝑦1 βˆ’ 𝑦2 ) 𝑏
πΏβˆ’ > 0 and this ends the proof.
Note that
𝑦 ⊀ 𝑐𝑅𝐿
∫
𝑔
is of the form
π‘₯β€² 𝑦 ⊀ β„Ž(π‘₯) 𝐹 𝑅,𝐿 is linear and
We can now move toward the main goal of this section. Clearly we need
some variant of the Measurable Maximum Theorem (see, e.g., [1, 18.19],
[19, Theorem 9.5], [28, 2K]). We state a version that is tailored to our needs
and has a simple proof; the technique is used also in Proposition 4.3.
Lemma A.4. Let𝑑 D be a predictable set-valued process with nonempty compact values in 2ℝ . Let 𝑓 (𝑦) = 𝑓 (πœ”, 𝑑, 𝑦) be a proper function on D with
values in ℝ βˆͺ {βˆ’βˆž} such that
(i) 𝑓 (πœ‘) is predictable whenever πœ‘ is a D -valued predictable process,
(ii) 𝑦 7β†’ 𝑓 (𝑦) is upper semicontinuous on D for xed (πœ”, 𝑑).
Then there exists a D -valued predictable process πœ‹ such that 𝑓 (πœ‹) = maxD 𝑓 .
Proof. We start with the Castaing representation [28, 1B] of D : there exist
D -valued predictable processes (πœ‘π‘› )𝑛β‰₯1 such that {πœ‘π‘› : 𝑛 β‰₯ 1} = D for each
(πœ”, 𝑑). By (i), 𝑓 βˆ— := max𝑛 𝑓 (πœ‘π‘› ) is predictable, and 𝑓 βˆ— = maxD 𝑓 by (ii).
βˆ—
𝑛
Fix π‘˜ β‰₯ 1 and let Λ𝑛 := {𝑓 βˆ’ 𝑓 (πœ‘π‘› ) ≀ 1/π‘˜}, Ξ› := Λ𝑛 βˆ– (Ξ›1 βˆͺ β‹… β‹… β‹… βˆͺ Ξ›π‘›βˆ’1 ).
βˆ‘
π‘˜
βˆ—
π‘˜
π‘˜
Dene πœ‹ :=
𝑛 πœ‘π‘› 1Λ𝑛 , then 𝑓 βˆ’ 𝑓 (πœ‹ ) ≀ 1/π‘˜ and πœ‹ ∈ D .
π‘˜
It remains to select a cluster point: By compactness, (πœ‹ )π‘˜β‰₯1 is bounded
for each (πœ”, 𝑑), so there is a convergent subsequence along random indices
πœπ‘˜ . More precisely, there exists a strictly increasing sequence of integerβˆ—
valued predictable processes πœπ‘˜ = {πœπ‘˜ (πœ”, 𝑑)} and a predictable process πœ‹
34
such that
𝜏 (πœ”,𝑑)
(πœ”) = πœ‹π‘‘βˆ— (πœ”) for all (πœ”, 𝑑). See, e.g., the proof of Föllmer
βˆ—
βˆ—
Lemma 1.63]. We have 𝑓 = 𝑓 (πœ‹ ) by (ii).
limπ‘˜ πœ‹π‘‘ π‘˜
and Schied [10,
𝑔
Our random function
satises property (i) of Lemma A.4 because the
characteristics are predictable (recall the denition [15, II.1.6]). We also note
that the intersection of closed predictable processes is predictable [28, 1M].
The sign of
𝑝
𝑔 ; we start
: ∣π‘₯∣ ≀ π‘Ÿ}.
is important as it switches the semicontinuity of
(ℝ𝑑 )
= {π‘₯ ∈
ℝ𝑑
𝑝<0
and denote
π΅π‘Ÿ
Proof of Lemma 3.8 for 𝑝 < 0.
In this case
𝑔 is u.s.c. on C ∩C 0 (Lemma A.2).
with the immediate case
C0
D(𝑛) := C ∩
∩ 𝐡𝑛
Lemma A.4 yields a predictable process
∈ arg maxD(𝑛) 𝑔 for each 𝑛 β‰₯ 1, and clearly lim𝑛 𝑔(πœ‹ 𝑛 ) = supC ∩C 0 𝑔 . As
𝑔(πœ‹ 𝑛 ) β‰₯ 𝑔(0) = 0, we have πœ‹ 𝑛 ∈ C 0,βˆ— by Lemma A.2.
Let
(ℝ𝑑 ).
πœ‹π‘›
A.1 Measurable Maximizing Sequence for 𝑝 ∈ (0, 1)
Fix
𝑝 ∈ (0, 1).
Since the continuity properties of
𝑔
are not clear, we will use
an approximating sequence of continuous functions. (See also Appendix B,
where an alternative approach is discussed and the continuity is claried
under an additional assumption on
C .)
We will approximate
𝑔
using Lévy
measures with enhanced integrability, a method suggested by [19] in a similar
problem. This preserves monotonicity properties that will be useful to pass
to the limit.
satises the following
𝑅 is locally bounded, or more generally if 𝐹 𝑅,𝐿
condition. We start with xed (πœ”, 𝑑).
Denition A.5.
𝐹
All this is not necessary if
Let
be a Lévy measure on
𝐹 𝑅,𝐿 and satises (2.5). (i) We say that
∫
(ii) The
𝐹
is
ℝ𝑑+1
which is equivalent to
𝑝-suitable
if
(1 + ∣π‘₯β€² ∣)(1 + ∣π‘₯∣)𝑝 1{∣π‘₯∣>1} 𝐹 (𝑑(π‘₯, π‘₯β€² )) < ∞.
𝑝-suitable
approximating sequence for 𝐹
Lévy measures dened by
𝑑𝐹𝑛 /𝑑𝐹 = 𝑓𝑛 ,
is the sequence
(𝐹𝑛 )𝑛β‰₯1
of
where
𝑓𝑛 (π‘₯) = 1{∣π‘₯βˆ£β‰€1} + π‘’βˆ’βˆ£π‘₯∣/𝑛 1{∣π‘₯∣>1} .
It is easy to see that each
addition being
quence
𝑓𝑛
𝑝-suitable
𝐹𝑛
in (ii) shares the properties of
because
(1 +
is increasing, monotone convergence shows that
for any measurable function
𝑉 β‰₯0
which is dened as in (A.1) but with
∫
ℝ𝑑+1 . We denote by
𝐹 𝑅,𝐿 replaced by 𝐹 .
on
𝐹,
while in
∣π‘₯∣)𝑝 π‘’βˆ’βˆ£π‘₯∣/𝑛 is bounded. As the se∫
𝑉 𝑑𝐹𝑛 ↑ 𝑉 𝑑𝐹
𝑔 𝐹 the function
Lemma A.6. If 𝐹 is 𝑝-suitable, 𝑔𝐹 is real-valued and continuous on C 0 .
Proof. Pick 𝑦𝑛 β†’ 𝑦 in C 0 . The only term in (A.1) for which continuity
𝐹 , where we choose πœ€ as in
𝐼 𝐹 = πΌπœ€πΉ + 𝐼>πœ€
𝐹
𝐹
Lemma A.2. We have πΌπœ€ (𝑦𝑛 ) β†’ πΌπœ€ (𝑦) by that lemma. When 𝐹 is 𝑝-suitable,
𝐹
the continuity of 𝐼>πœ€ follows from the dominated convergence theorem.
is not evident, is the integral
35
Remark A.7.
Dene the set
βˆͺ
(C ∩ C 0 )β‹„ :=
πœ‚(C ∩ C 0 ).
πœ‚βˆˆ[0,1)
1 + 𝑦 ⊀ π‘₯ is 𝐹 𝑅 (𝑑π‘₯)-essentially bounded
away from zero. Indeed, 𝑦 = πœ‚π‘¦0 with πœ‚ ∈ [0, 1) and 𝐹 𝑅 {𝑦0⊀ π‘₯ β‰₯ βˆ’1} = 0,
⊀
𝑅
0 β‹„
0,βˆ— . If C is
hence 1 + 𝑦 π‘₯ β‰₯ 1 βˆ’ πœ‚ , 𝐹 -a.e. In particular, (C ∩ C ) βŠ† C
0 β‹„
star-shaped with respect to the origin, we also have (C ∩ C ) βŠ† C .
Its elements
𝑦
have the property that
We introduce the compact-valued process
D(π‘Ÿ) := C ∩ C 0 ∩ π΅π‘Ÿ (ℝ𝑑 ).
Lemma A.8. Let 𝐹 be 𝑝-suitable. Under (C3), arg maxD(π‘Ÿ) 𝑔𝐹 βŠ† C 0,βˆ— .
More generally, this holds whenever 𝐹 is a Lévy measure equivalent to
𝑅,𝐿
𝐹
satisfying (2.5) and 𝑔𝐹 is nite-valued.
Proof.
Assume that
as in the denition
derivative
𝐷𝑦ˇ,𝑦0 𝑔
𝑦ˇ ∈ C 0 βˆ– C 0,βˆ— is a maximum of 𝑔 𝐹 . Let πœ‚ ∈ (πœ‚, 1) be
of (C3) and 𝑦0 := πœ‚ 𝑦
Λ‡. By Lemma 5.14, the directional
can be calculated by dierentiating under the integral sign.
For the integrand of
𝐼𝐹
we have
{
}
{
}
𝐷𝑦ˇ,𝑦0 π‘βˆ’1 (1 + 𝑦 ⊀ π‘₯)𝑝 βˆ’ π‘βˆ’1 βˆ’ 𝑦 ⊀ β„Ž(π‘₯) = (1 βˆ’ πœ‚) (1 + π‘¦Λ‡βŠ€ π‘₯)π‘βˆ’1 π‘¦Λ‡βŠ€ π‘₯ βˆ’ π‘¦Λ‡βŠ€ β„Ž(π‘₯) .
But this is innite on a set of positive measure as
𝐹 {Λ‡
π‘¦βŠ€π‘₯
Let
= βˆ’1} > 0,
𝐹
𝑦ˇ ∈ C 0 βˆ– C 0,βˆ—
means that
contradicting the last assertion of Lemma 5.14.
be a Lévy measure on
ℝ𝑑+1
which is equivalent to
𝐹 𝑅,𝐿
and sat-
ises (2.5). The crucial step is
Lemma A.9. Let (𝐹𝑛 ) be the 𝑝-suitable approximating sequence for 𝐹 and
x π‘Ÿ > 0. For each 𝑛, arg maxD(π‘Ÿ) 𝑔𝐹𝑛 βˆ•= βˆ…, and for any π‘¦π‘›βˆ— ∈ arg maxD(π‘Ÿ) 𝑔𝐹𝑛
it holds that lim sup𝑛 𝑔𝐹 (π‘¦π‘›βˆ— ) = supD(π‘Ÿ) 𝑔𝐹 .
Proof.
We rst show that
𝐼 𝐹𝑛 (𝑦) β†’ 𝐼 𝐹 (𝑦) for any 𝑦 ∈ C 0 .
(A.2)
{ βˆ’1
}
∫
𝐹
β€²
Recall that 𝐼 𝑛 (𝑦) = (𝑙+π‘₯ ) 𝑝
(1+𝑦 ⊀ π‘₯)𝑝 βˆ’π‘βˆ’1 βˆ’π‘¦ ⊀ β„Ž(π‘₯) 𝑓𝑛 (π‘₯) 𝐹 (𝑑(π‘₯, π‘₯β€² )),
where 𝑓𝑛 is nonnegative and increasing in 𝑛. As 𝑓𝑛 = 1 in a neighbor𝐹𝑛
hood of the origin, we need to consider only 𝐼>πœ€ (for πœ€ = 1, say). Its
integrand is bounded below, simultaneously for all 𝑛, by a negative conβ€²
stant times (1 + ∣π‘₯ ∣), which is 𝐹 -integrable on the relevant domain. As
{
β€²
(𝑓𝑛 ) is increasing, we can apply monotone
convergence on the set (π‘₯, π‘₯ ) :
}
βˆ’1
⊀
𝑝
βˆ’1
⊀
𝑝 (1 + 𝑦 π‘₯) βˆ’ 𝑝 βˆ’ 𝑦 β„Ž(π‘₯) β‰₯ 0 and dominated convergence on the
complement to deduce (A.2).
36
π‘¦π‘›βˆ— ∈ arg maxD(π‘Ÿ) 𝑔 𝐹𝑛 is clear by compactness of D(π‘Ÿ) and
𝑔 𝐹𝑛 (Lemma A.6). Let 𝑦 ∈ D(π‘Ÿ) be arbitrary. By denition of
Existence of
continuity of
π‘¦π‘›βˆ—
and (A.2),
lim sup 𝑔 𝐹𝑛 (π‘¦π‘›βˆ— ) β‰₯ lim sup 𝑔 𝐹𝑛 (𝑦) = 𝑔 𝐹 (𝑦).
𝑛
𝑛
lim sup𝑛 𝑔 𝐹 (π‘¦π‘›βˆ— ) β‰₯ lim sup𝑛 𝑔 𝐹𝑛 (π‘¦π‘›βˆ— ). We can split the integral
𝐹
𝐼 𝑛 (𝑦) into a sum of three terms: The integral over {∣π‘₯∣ ≀ 1} is the same as
𝐹
for 𝐼 , since 𝑓𝑛 = 1 on this set. We can assume that the cut-o β„Ž vanishes
We show
outside
{∣π‘₯∣ ≀ 1}.
The second term is then
∫
(𝑙 + π‘₯β€² )π‘βˆ’1 (1 + 𝑦 ⊀ π‘₯)𝑝 𝑓𝑛 𝑑𝐹,
{∣π‘₯∣>1}
here the integrand is nonnegative and hence increasing in
the third term is
∫
𝑛,
for all
𝑦;
and
(𝑙 + π‘₯β€² )(βˆ’π‘βˆ’1 )𝑓𝑛 𝑑𝐹,
{∣π‘₯∣>1}
which is decreasing in
𝑛
but converges to
∫
{∣π‘₯∣>1} (𝑙
+ π‘₯β€² )(βˆ’π‘βˆ’1 ) 𝑑𝐹 .
Thus
𝑔 𝐹 (π‘¦π‘›βˆ— ) β‰₯ 𝑔 𝐹𝑛 (π‘¦π‘›βˆ— ) βˆ’ πœ€π‘›
∫
πœ€π‘› := {∣π‘₯∣>1} (𝑙 + π‘₯β€² )(βˆ’π‘βˆ’1 )(𝑓𝑛 βˆ’ 1) 𝑑𝐹 ↓ 0. Together,
supD(π‘Ÿ) 𝑔 𝐹 β‰₯ lim sup𝑛 𝑔 𝐹 (π‘¦π‘›βˆ— ) β‰₯ lim sup𝑛 𝑔 𝐹𝑛 (π‘¦π‘›βˆ— ) β‰₯ supD(π‘Ÿ) 𝑔 𝐹 .
with the sequence
conclude
Proof of Lemma 3.8 for 𝑝 ∈ (0, 1).
Fix
π‘Ÿ > 0.
we
By Lemma A.4 we can nd
𝑛,π‘Ÿ
𝑛,π‘Ÿ for arg max
𝐹𝑛
measurable selectors πœ‹
D(π‘Ÿ) 𝑔 ; i.e., πœ‹π‘‘ (πœ”) plays the role
𝑛 := πœ‹ 𝑛,𝑛 and noting D(𝑛) ↑ C ∩ C 0 ,
βˆ—
of 𝑦𝑛 in Lemma A.9. Taking πœ‹
𝑛 are C ∩ C 0 -valued predictable processes such
Lemma A.9 shows that πœ‹
lim sup𝑛 𝑔(πœ‹ 𝑛 ) = supC ∩C 0 𝑔 𝑃 βŠ— 𝐴-a.e.
0,βˆ— .
values in C
that
Lemma A.8 shows that
πœ‹π‘›
takes
B Parametrization by Representative Portfolios
This appendix introduces an equivalent transformation of the model
(𝑅, C )
with specic properties (Theorem B.3); the main idea is to substitute the
given assets by wealth processes that represent the investment opportunities
of the model. While the result is of independent interest, the main conclusion
in our context is that the approximation technique from Appendix A.1 for the
𝑝 ∈ (0, 1) can be avoided, at
C : If the utility maximization
case
least under slightly stronger assumptions
on
problem is nite, the corresponding Lévy
measure in the transformed model is
the corresponding function
𝑔
𝑝-suitable (cf. Denition A.5) and hence
is continuous. This is not only an alternative
argument to prove Lemma 3.8. In applications, continuity can be useful to
37
construct a maximizer for
not know
a priori
𝑔
(rather than a maximizing sequence) if one does
that there exists an optimal strategy. A static version of
our construction was carried out for the case of Lévy processes in [25, Ÿ4].
In this appendix we use the following assumptions on the set-valued
process
(C1)
(C2)
(C4)
C
of constraints:
C
is predictable.
C
is closed.
C
is star-shaped with respect to the origin:
πœ‚C βŠ† C
for all
πœ‚ ∈ [0, 1].
Since we already obtained a proof of Lemma 3.8, we do not strive for minimal conditions here. Clearly (C4) implies condition (C3) from Section 2.4,
but its main implication is that we can select a bounded (hence
𝑅-integrable)
process in the subsequent lemma. The following result is the construction of
the
𝑗 th representative portfolio, a portfolio
𝑗 th asset whenever this is feasible.
with the property that it invests
in the
Lemma B.1. Fix 1 ≀ 𝑗 ≀ 𝑑 and let 𝐻 𝑗 = {π‘₯ ∈ ℝ𝑑 : π‘₯𝑗 βˆ•= 0}. There exists
a bounded predictable C ∩ C 0,βˆ— -valued process πœ™ satisfying
{
}
{πœ™π‘— = 0} = C ∩ C 0,βˆ— ∩ 𝐻 𝑗 = βˆ… .
Proof.
𝑗
𝐡{1 = 𝐡1 (ℝ𝑑 ) be the} closed
unit ball and 𝐻 := 𝐻 . Condition
{
}
0,βˆ— ∩ 𝐻 = βˆ… = C ∩ 𝐡 ∩ C 0,βˆ— ∩ 𝐻 = βˆ… , hence we may
(C4) implies C ∩ C
1
𝑑
𝑗
βˆ’1 }
substitute C by C ∩ 𝐡1 . Dene the closed sets π»π‘˜ = {π‘₯ ∈ ℝ : ∣π‘₯ ∣ β‰₯ π‘˜
βˆͺ
0
for π‘˜ β‰₯ 1, then
π‘˜ π»π‘˜ = 𝐻 . Moreover, let Dπ‘˜ = C ∩ C ∩ π»π‘˜ . This is a
compact-valued predictable process, so there exists a predictable process πœ™π‘˜
𝑗
such that πœ™π‘˜ ∈ Dπ‘˜ (hence πœ™π‘˜ βˆ•= 0) on the set Ξ›π‘˜ := {Dπ‘˜ βˆ•= βˆ…} and πœ™π‘˜ = 0 on
βˆ‘
π‘˜
β€²
the complement. Dene Ξ› := Ξ›π‘˜ βˆ– (Ξ›1 βˆͺ β‹… β‹… β‹… βˆͺ Ξ›π‘˜βˆ’1 ) and πœ™ :=
π‘˜ πœ™}π‘˜ 1Ξ›π‘˜ .
{
} {
β€²
′𝑗
0
0,βˆ— ∩ 𝐻 = βˆ… ; the
Then βˆ£πœ™ ∣ ≀ 1 and {πœ™ = 0} = C ∩ C ∩ 𝐻 = βˆ… = C ∩ C
Let
second equality uses (C4) and Remark A.7. These two facts also show that
πœ™ := 12 πœ™β€²
has the same property while in addition being
Remark B.2.
diameter of
C
C ∩ C 0,βˆ— -valued.
The previous proof also applies if instead of (C4), e.g., the
is uniformly bounded and
C 0 = C 0,βˆ— .
Ξ¦ is a 𝑑×𝑑-matrix with columns πœ™1 , . . . , πœ™π‘‘ ∈ 𝐿(𝑅), the matrix stochas˜ = Ξ¦ βˆ™ 𝑅 is the ℝ𝑑 -valued process given by 𝑅
Λœπ‘— = πœ™π‘— βˆ™ 𝑅. If
tic integral 𝑅
𝑑
πœ“ ∈ 𝐿(Ξ¦ βˆ™ 𝑅) is ℝ -valued, then Ξ¦πœ“ ∈ 𝐿(𝑅) and
If
πœ“ βˆ™ (Ξ¦ βˆ™ 𝑅) = (Ξ¦πœ“) βˆ™ 𝑅.
If
D
(B.1)
is a set-valued process which is predictable, closed and contains the ori-
gin, then the preimage
Ξ¦βˆ’1 D
shares these properties (cf. [28, 1Q]). Convexity
and star-shape are also preserved.
We obtain the following model if we sequentially replace the given assets
by representative portfolios; here
1≀𝑗≀𝑑
𝑖
(i.e., 𝑒𝑗
𝑒𝑗
denotes the
= 𝛿𝑖𝑗 ).
38
𝑗 th
unit vector in
ℝ𝑑
for
Theorem B.3. There exists a predictable ℝ𝑑×𝑑 -valued uniformly bounded
process Ξ¦ such that the nancial market model with returns
˜ := Ξ¦ βˆ™ 𝑅
𝑅
and constraints C˜ := Ξ¦βˆ’1 C has the following properties: for all 1 ≀ 𝑗 ≀ 𝑑,
(i) Ξ”π‘…Λœπ‘— > βˆ’1 (positive prices),
(ii) 𝑒𝑗 ∈ C˜ ∩ C˜0,βˆ— , where C˜0,βˆ— = Ξ¦βˆ’1 C 0,βˆ— (entire wealth can be invested in
each asset),
˜ C˜) admits the same wealth processes as (𝑅, C ).
(iii) the model (𝑅,
Proof. We treat the components one by one. Let 𝑗 = 1 and let πœ™ = πœ™(1)
𝑅1 by the process πœ™ βˆ™ 𝑅,
Ξ¦ = Ξ¦(1) is the 𝑑 × π‘‘-matrix
⎞
be as in Lemma B.1. We replace the rst asset
equivalently, we replace
𝑅
or
Ξ¦ 𝑅, where
βŽ› 1
πœ™
⎜ πœ™2 1
⎟
⎜
⎟
Φ=⎜ .
⎟.
..
⎝ ..
⎠
.
𝑑
πœ™
1
by
βˆ™
Ξ¦βˆ’1 C 0 and we replace C by Ξ¦βˆ’1 C . Note
βˆ’1 (C ∩ C 0,βˆ— ) because Φ𝑒 = πœ™ ∈ C ∩ C 0,βˆ— by construction.
that 𝑒1 ∈ Ξ¦
1
0,βˆ— -valued process πœ‹ ∈ 𝐿(𝑅) there exists
We show that for every C ∩ C
πœ“ predictable such that Ξ¦πœ“ = πœ‹ . In view of (B.1), this will imply that the
The new natural constraints are
new model admits the same wealth processes as the old one.
On the set
{πœ™1 βˆ•= 0} = {Ξ¦ is invertible} we take πœ“ = Ξ¦βˆ’1 πœ‹ and on the complement we
1
𝑗
𝑗
choose πœ“ ≑ 0 and πœ“ = πœ‹ for 𝑗 β‰₯ 2; this is the same as inverting Ξ¦ on its
1
1
image. Note that {πœ™ = 0} βŠ† {πœ‹ = 0} by the choice of πœ™.
We proceed with the second component of the new model in the same
We obtain matrices Ξ¦(𝑗) for
Ξ¦Μ‚ = Ξ¦(1) β‹… β‹… β‹… Ξ¦(𝑑). Then Ξ¦Μ‚ has the required properties.
βˆ’1 (C ∩C 0,βˆ— ).
Indeed, the construction and Ξ¦(𝑖)𝑒𝑗 = 𝑒𝑗 for 𝑖 βˆ•= 𝑗 imply 𝑒𝑗 ∈ Ξ¦Μ‚
way, and then continue until the last one.
1≀𝑗 ≀𝑑
and set
This is (ii), and (i) is a consequence of (ii).
Coming back to the utility maximization problem, note that property
(iii) implies that the value functions and the opportunity processes for the
models
(𝑅, C )
and
˜ C˜)
(𝑅,
the sequel. Furthermore, if
coincide up to evanescence; we identify them in
π‘”Λœ denotes
the analogue of
𝑔
in the model
˜ C˜),
(𝑅,
cf. (A.1), we have the relation
π‘”Λœ(𝑦) = 𝑔(Φ𝑦),
𝑦 ∈ C˜0 .
π‘”Λœ is equivalent to nding one for 𝑔 and if (˜
πœ‹ , πœ…) is
˜ C˜) then (Φ˜
(𝑅,
πœ‹ , πœ…) is optimal for (𝑅, C ). In fact,
˜ C˜), in particular
interest carry over from (𝑅, C ) to (𝑅,
Finding a maximizer for
an optimal strategy for
most properties of
any no-arbitrage property that is dened via the set of admissible (positive)
wealth processes.
39
Remark B.4.
A classical no-arbitrage condition dened in a slightly dier-
ent way is that there exist a probability measure
a
π‘„β‰ˆπ‘ƒ
β„°(𝑅) is
˜
In this case, β„°(𝑅) is even
under which
𝜎 -martingale; cf. Delbaen and Schachermayer [9].
𝑄, as it is a 𝜎 -martingale with positive components.
a local martingale under
Property (ii) from Theorem B.3 is useful to apply the following result.
Lemma B.5. Let 𝑝 ∈ (0, 1) and assume 𝑒𝑗 ∈ C ∩ C 0,βˆ— for 1 ≀ 𝑗 ≀ 𝑑. Then
𝑒(π‘₯0 ) < ∞ implies that 𝐹 𝑅,𝐿 is 𝑝-suitable. If, in
∫ addition, there exists a
constant π‘˜1 such that 𝐷 β‰₯ π‘˜1 > 0, it follows that {∣π‘₯∣>1} ∣π‘₯βˆ£π‘ 𝐹 𝑅 (𝑑π‘₯) < ∞.
Proof. As 𝑝 > 0 and 𝑒(π‘₯0 ) < ∞, 𝐿 is well dened and 𝐿, πΏβˆ’ > 0 by
Section 2.2. No further properties were used to establish Lemma 3.4, whose
formula shows that
𝑔(πœ‹)
is nite
particular, from the denition of
π‘βˆ’1
πœ‹ ⊀ β„Ž(π‘₯)
𝑃 βŠ— 𝐴-a.e.
πœ‹ ∈{ π’œ = π’œπ‘“ 𝐸 . In
(πΏβˆ’ +π‘₯β€² ) π‘βˆ’1 (1+πœ‹ ⊀ π‘₯)𝑝 βˆ’
for all
𝑔 , it follows that
∫
βˆ’
𝐹 𝑅,𝐿 (𝑑(π‘₯, π‘₯β€² )) is nite. If 𝐷 β‰₯ π‘˜1∫, {[26, Lemma 3.5] shows
β€²
𝐿
β€²
that 𝐿 β‰₯ π‘˜1 , hence πΏβˆ’ + π‘₯ β‰₯ π‘˜1 𝐹 (𝑑π‘₯ )-a.e. and
π‘βˆ’1 (1 + πœ‹ ⊀ π‘₯)𝑝 βˆ’ π‘βˆ’1 βˆ’
}
πœ‹ ⊀ β„Ž(π‘₯) 𝐹 𝑅 (𝑑π‘₯) < ∞. We choose πœ‹ = 𝑒𝑗 (and πœ… arbitrary) for 1 ≀ 𝑗 ≀ 𝑑 to
}
deduce the result.
𝑒(π‘₯0 ) < ∞ does not imply any properties of 𝑅;
0,βˆ— = {0}. The transformation
for instance, in the trivial cases C = {0} or C
0,βˆ—
changes the geometry of C and C
such that Theorem B.3(ii) holds, and
In general, the condition
then the situation is dierent.
Corollary B.6. Let 𝑝 ∈ (0, 1) and 𝑒(π‘₯0 ) < ∞. In the model
˜
Theorem B.3, 𝐹 𝑅,𝐿
is 𝑝-suitable and hence π‘”Λœ is continuous.
˜ C˜)
(𝑅,
of
(𝑅, C )
𝑝-suitable approximating sequences. In some
cases, Lemma B.5 applies directly in (𝑅, C ). In particular, if the asset prices
𝑗
are strictly positive (Δ𝑅 > βˆ’1 for 1 ≀ 𝑗 ≀ 𝑑), then the positive orthant of
𝑑
0,βˆ—
ℝ is contained in C
and the condition of Lemma B.5 is satised as soon
as 𝑒𝑗 ∈ C for 1 ≀ 𝑗 ≀ 𝑑.
Therefore, to prove Lemma 3.8 under (C4), we may substitute
by
˜ C˜)
(𝑅,
and avoid the use of
C Omitted Calculation
This appendix contains a calculation which was omitted in the proof of
Proposition 5.12.
Lemma C.1. Let (β„“, πœ‹Λ‡ , πœ…Λ‡) be a solution of the Bellman equation, (πœ‹, πœ…) ∈ π’œ,
Λ‡ := 𝑋(Λ‡
¯ = 𝑅 βˆ’ (π‘₯ βˆ’ β„Ž(π‘₯)) βˆ— πœ‡π‘… as well as
𝑋 := 𝑋(πœ‹, πœ…) and 𝑋
πœ‹, πœ…
Λ‡ ). Dene 𝑅
Λ‡ π‘βˆ’1 𝑋 satises
πœ‹
¯ := (𝑝 βˆ’ 1)Λ‡
πœ‹ + πœ‹ and πœ…
¯ := (𝑝 βˆ’ 1)Λ‡
πœ… + πœ…. Then πœ‰ := ℓ𝑋
(
Λ‡ π‘βˆ’1 π‘‹βˆ’
𝑋
βˆ’
)βˆ’1
βˆ™
πœ‰=
(
)⊀
¯ βˆ’ β„“βˆ’ πœ…
β„“ βˆ’ β„“0 + β„“βˆ’ πœ‹
¯βˆ™π‘…
¯ βˆ™ πœ‡ + β„“βˆ’ (𝑝 βˆ’ 1) π‘βˆ’2
Λ‡ + πœ‹ 𝑐𝑅 πœ‹
Λ‡ βˆ™π΄+πœ‹
¯ ⊀ 𝑐𝑅ℓ βˆ™ 𝐴
2 πœ‹
{
}
+πœ‹
¯ ⊀π‘₯β€² β„Ž(π‘₯) βˆ— πœ‡π‘…,β„“ + (β„“βˆ’ + π‘₯β€² ) (1 + πœ‹
Λ‡ ⊀ π‘₯)π‘βˆ’1 (1 + πœ‹ ⊀ π‘₯) βˆ’ 1 βˆ’ πœ‹
¯ ⊀ β„Ž(π‘₯) βˆ— πœ‡π‘…,β„“ .
40
Proof.
We may assume
π‘₯0 = 1 .
This calculation is similar to the one in the
proof of Lemma 3.4 and therefore we shall be brief. By Itô's formula we have
Λ‡ π‘βˆ’1 = β„°(𝜁)
𝑋
Thus
for
𝜁 = (𝑝 βˆ’ 1)(Λ‡
πœ‹ βˆ™π‘…βˆ’πœ…
Λ‡ βˆ™ πœ‡) + (π‘βˆ’1)(π‘βˆ’2)
πœ‹
Λ‡ ⊀ 𝑐𝑅 πœ‹
Λ‡βˆ™π΄
2
{
}
+ (1 + πœ‹
Λ‡ ⊀ π‘₯)π‘βˆ’1 βˆ’ 1 βˆ’ (𝑝 βˆ’ 1)Λ‡
πœ‹ ⊀ π‘₯ βˆ— πœ‡π‘… .
(
)
Λ‡ π‘βˆ’1 𝑋 = β„° 𝜁 + πœ‹ βˆ™ 𝑅 βˆ’ πœ… βˆ™ πœ‡ + [𝜁, πœ‹ βˆ™ 𝑅] =: β„°(Ξ¨) with
𝑋
[𝑅, 𝜁] = [𝑅𝑐 , 𝜁 𝑐 ] +
βˆ‘
Ξ”π‘…Ξ”πœ
= (𝑝 βˆ’ 1)𝑐 πœ‹
Λ‡ βˆ™ 𝐴 + (𝑝 βˆ’ 1)Λ‡
πœ‹ ⊀ π‘₯π‘₯ βˆ— πœ‡π‘…
{
}
+ π‘₯ (1 + πœ‹
Λ‡ ⊀ π‘₯)π‘βˆ’1 βˆ’ 1 βˆ’ πœ‹
Λ‡ ⊀ π‘₯ βˆ— πœ‡π‘…
𝑅
and recombining the terms yields
(
)⊀
Ξ¨=πœ‹
¯ βˆ™π‘…βˆ’πœ…
¯ βˆ™ πœ‡ + (𝑝 βˆ’ 1) π‘βˆ’2
Λ‡ + πœ‹ 𝑐𝑅 πœ‹
Λ‡βˆ™π΄
2 πœ‹
{
}
+ (1 + πœ‹
Λ‡ ⊀ π‘₯)π‘βˆ’1 (1 + πœ‹ ⊀ π‘₯) βˆ’ 1 βˆ’ πœ‹
¯ ⊀ π‘₯ βˆ— πœ‡π‘… .
Then
(
Λ‡ π‘βˆ’1 π‘‹βˆ’
𝑋
βˆ’
)βˆ’1
βˆ™
πœ‰ = β„“ βˆ’ β„“0 + β„“βˆ’ βˆ™ Ξ¨ + [β„“, Ξ¨],
[β„“, Ξ¨] = [ℓ𝑐 , Ψ𝑐 ] +
βˆ‘
where
ΔℓΔΨ
=πœ‹
¯ ⊀ 𝑐𝑅ℓ βˆ™ 𝐴 + πœ‹
¯ ⊀ π‘₯β€² π‘₯ βˆ— πœ‡π‘…,β„“
{
}
+ π‘₯β€² (1 + πœ‹
Λ‡ ⊀ π‘₯)π‘βˆ’1 (1 + πœ‹ ⊀ π‘₯) βˆ’ 1 βˆ’ πœ‹
¯ ⊀ π‘₯ βˆ— πœ‡π‘…,β„“ .
We arrive at
(
Λ‡ π‘βˆ’1 π‘‹βˆ’
𝑋
βˆ’
)βˆ’1
βˆ™
πœ‰=
(
)⊀ 𝑅
β„“ βˆ’ β„“0 + β„“βˆ’ πœ‹
¯ βˆ™ 𝑅 βˆ’ β„“βˆ’ πœ…
¯ βˆ™ πœ‡ + β„“βˆ’ (𝑝 βˆ’ 1) π‘βˆ’2
πœ‹
Λ‡
+
πœ‹
𝑐 πœ‹
Λ‡ βˆ™π΄+πœ‹
¯ ⊀ 𝑐𝑅ℓ βˆ™ 𝐴
2
{
}
+πœ‹
¯ ⊀ π‘₯β€² π‘₯ βˆ— πœ‡π‘…,β„“ + (β„“βˆ’ + π‘₯β€² ) (1 + πœ‹
Λ‡ ⊀ π‘₯)π‘βˆ’1 (1 + πœ‹ ⊀ π‘₯) βˆ’ 1 βˆ’ πœ‹
¯ ⊀ π‘₯ βˆ— πœ‡π‘…,β„“ .
The result follows by writing
π‘₯ = β„Ž(π‘₯) + π‘₯ βˆ’ β„Ž(π‘₯).
References
[1] C. D. Aliprantis and K. C. Border. Innite Dimensional Analysis: A Hitchhiker's Guide. Springer, Berlin, 3rd edition, 2006.
[2] J. P. Ansel and C. Stricker. Décomposition de Kunita-Watanabe. In Séminaire
de Probabilités XXVII, volume 1557 of Lecture Notes in Math., pages 3032,
Springer, Berlin, 1993.
[3] O. E. Barndor-Nielsen and N. Shephard. Non-Gaussian Ornstein-Uhlenbeckbased models and some of their uses in nancial economics. J. R. Stat. Soc.
Ser. B Stat. Methodol., 63(2):167241, 2001.
41
[4] P. Briand and Y. Hu. Quadratic BSDEs with convex generators and unbounded terminal conditions. Probab. Theory Related Fields, 141:543567,
2008.
[5] A. Βƒerný and J. Kallsen. On the structure of general mean-variance hedging
strategies. Ann. Probab., 35(4):14791531, 2007.
[6] T. Choulli and C. Stricker. Comparing the minimal Hellinger martingale measure of order π‘ž to the π‘ž -optimal martingale measure. Stochastic Process. Appl.,
119(4):13681385, 2009.
[7] C. Czichowsky and M. Schweizer. On the Markowitz problem under convex
cone constraints. In preparation.
[8] C. Czichowsky and M. Schweizer. Closedness in the semimartingale topology
for spaces of stochastic integrals with constrained integrands. In Séminaire
de Probabilités XLIII, volume 2006 of Lecture Notes in Math., pages 413436,
Springer, Berlin, 2011.
[9] F. Delbaen and W. Schachermayer. The fundamental theorem of asset pricing
for unbounded stochastic processes. Math. Ann., 312:215250, 1998.
[10] H. Föllmer and A. Schied. Stochastic Finance: An Introduction in Discrete
Time. W. de Gruyter, Berlin, 2nd edition, 2004.
[11] T. Goll and J. Kallsen. A complete explicit solution to the log-optimal portfolio
problem. Ann. Appl. Probab., 13(2):774799, 2003.
[12] T. Goll and L. Rüschendorf. Minimax and minimal distance martingale
measures and their relationship to portfolio optimization. Finance Stoch.,
5(4):557581, 2001.
[13] P. Grandits. On martingale measures for stochastic processes with independent
increments. Theory Probab. Appl., 44(1):3950, 2000.
[14] Y. Hu, P. Imkeller, and M. Müller. Utility maximization in incomplete markets. Ann. Appl. Probab., 15(3):16911712, 2005.
[15] J. Jacod and A. N. Shiryaev. Limit Theorems for Stochastic Processes.
Springer, Berlin, 2nd edition, 2003.
[16] M. Jeanblanc, S. Klöppel, and Y. Miyahara. Minimal 𝑓 π‘ž -martingale measures
for exponential Lévy processes. Ann. Appl. Probab., 17(5/6):16151638, 2007.
[17] J. Kallsen. 𝜎 -localization and 𝜎 -martingales. Theory Probab. Appl., 48(1):152
163, 2004.
[18] J. Kallsen and J. Muhle-Karbe. Utility maximization in ane stochastic
volatility models. Int. J. Theor. Appl. Finance, 13(3):459477, 2010.
[19] I. Karatzas and C. Kardaras. The numéraire portfolio in semimartingale nancial models. Finance Stoch., 11(4):447493, 2007.
[20] I. Karatzas and G. šitkovi¢. Optimal consumption from investment and
random endowment in incomplete semimartingale markets. Ann. Probab.,
31(4):18211858, 2003.
[21] D. Kramkov and W. Schachermayer. The asymptotic elasticity of utility functions and optimal investment in incomplete markets. Ann. Appl. Probab.,
9(3):904950, 1999.
[22] D. Kramkov and W. Schachermayer. Necessary and sucient conditions in the
problem of optimal investment in incomplete markets. Ann. Appl. Probab.,
13(4):15041516, 2003.
[23] M. Mania and R. Tevzadze. A unied characterization of the π‘ž -optimal and
42
[24]
[25]
[26]
[27]
[28]
[29]
minimal entropy martingale measures. Georgian Math. J., 10(2):289310,
2003.
J. Muhle-Karbe. On Utility-Based Investment, Pricing and Hedging in Incomplete Markets. PhD thesis, TU München, 2009.
M. Nutz. Power utility maximization in constrained exponential Lévy models.
To appear in Math. Finance, 2009.
M. Nutz. The opportunity process for optimal consumption and investment
with power utility. Math. Financ. Econ., 3(3):139159, 2010.
M. Nutz. Risk aversion asymptotics for power utility maximization. To appear
in Probab. Theory Related Fields, 2010.
R. T. Rockafellar. Integral functionals, normal integrands and measurable
selections. In Nonlinear Operators and the Calculus of Variations, volume 543
of Lecture Notes in Math., pages 157207, Springer, Berlin, 1976.
M. Schweizer. On the minimal martingale measure and the Föllmer-Schweizer
decomposition. Stochastic Anal. and Appl., 13:573599, 1995.
43