Working Paper Series _______________________________________________________________________________________________________________________ National Centre of Competence in Research Financial Valuation and Risk Management Working Paper No. 588 The Mathematical Structure of Horizon-Dependence in Optimal Portfolio Choice Tahir Choulli Martin Schweizer First version: October 2009 Current version: October 2009 This research has been carried out within the NCCR FINRISK project on “Mathematical Methods in Financial Risk Management” ___________________________________________________________________________________________________________ The mathematical structure of horizon-dependence in optimal portfolio choice Tahir Choulli University of Alberta Martin Schweizer ∗ ETH Zürich Department of Mathematical Departement Mathematik and Statistical Sciences 632 Central Academic Building ETH-Zentrum, HG G 51.2 Edmonton, AB T6G 2G1 CH – 8092 Zürich Canada [email protected] Switzerland [email protected] Abstract: For every T ≥ 0, we consider the classical problem of maximising expected utility from terminal wealth for the time horizon T . This yields a collection ¡ ¢ b = (X bT )T ≥0 of optimal final wealths and a family ϑ̂ = ϑ̂(T, ·) of optimal X T ≥0 b has investment strategies, one each for every time horizon T . We prove that X e which is a semimartingale, and that ϑ̂ has a version which is of a version X finite variation with respect to the horizon parameter T . With the help of a Key words: new stochastic Fubini theorem, we also establish a novel decomposition of the e semimartingale X. optimal portfolio choice, horizon-dependence, utility maximisation, semimartingales, stochastic integrals, stochastic Fubini theorem, mathematical finance MSC 2000 Classification Numbers: 60G48, 60H05, 60G07, 28B05, 91B28 This version: 26.10.2009 ∗ corresponding author 0. Introduction Optimal investment problems are ubiquitous. They come up as soon as people start thinking about their future lifestyles in some structured way and try to achieve from their available resources a subjectively optimal outcome. Literally hundreds of papers have studied the ¡ many facets of such questions, ranging from psychological aspects see Jaggia/Thosar (2000) ¢ for a randomly chosen recent sample over economic issues to purely mathematical problems. But despite the vast amount of literature, one aspect of key importance seems not well understood to date: How exactly does the planning horizon affect the optimal investment decision? Everyone agrees that there is some effect, and awareness of this fundamental issue can be traced back in the scientific literature at least to Hakansson (1969) or Yaari (1965), to name just two early and well-known references. But a general quantitative result has remained elusive. In this paper, we provide a rigorous mathematical answer to the above question in a very general setting. Our financial market is modelled by a general semimartingale, and we consider for each T ≥ 0 the problem of maximising expected utility from terminal wealth for b = (X bT )T ≥0 of optimal final wealths and the time horizon T . This leads to a collection X ¡ ¢ a family ϑ̂ = ϑ̂(T, ·) T ≥0 of optimal investment strategies, one each for every time horizon b and we show how the field ϑ̂ behaves as a T . We elucidate the structure of the process X function of T , thus exhibiting precisely the mathematical structure of horizon-dependence in optimal portfolio choice. To the best of our knowledge, the approach in this paper is entirely new, and the main results have no genuine precursors in the literature. (For completeness, we should mention that the case of logarithmic utility has been solved in full generality; see Goll/Kallsen (2003). But this case is extremely special and non-typical — logarithmic utility is well known to induce myopic behaviour, which means that there is no dependence on the time horizon at all.) More detailed discussions on related work are provided later in Section 3 when we have a precise mathematical problem formulation and the appropriate terminology at our disposal. We are also well aware that many new questions (as well as potential applications) arise from our work here, and several of these will be addressed in forthcoming papers. The paper is structured as follows. Section 1 provides the mathematical formulation of the problem, explains the three main results, and gives an overview of the ideas and methods b has a version X e which needed for their proofs. Section 2 proves the first main result, that X is a semimartingale. Section 3 proves the second main result, that ϑ̂ has a version which is of finite variation with respect to the horizon parameter T . Section 5 establishes as third main e This is based on a result a remarkable and novel decomposition of the semimartingale X. new stochastic Fubini theorem, which is of independent mathematical interest and presented in Section 4. 1 1. Problem formulation and overview of results In this section, we mathematically explain the basic problem studied in this paper. We introduce some notation, provide intuition, and give an overview of the main results. Let (Ω, F , P ) be a probability space with a filtration IF = (Ft )t≥0 on which we have an IRd -valued semimartingale S = (St )t≥0 . We consider IRd -valued predictable processes R ϑ = (ϑt )t≥0 which are S-integrable in the sense that the stochastic integral ϑ . S = ϑ dS as a process is well defined (and hence a real-valued semimartingale). For a fixed constant x > 0, such a ϑ is called (an) admissible (strategy) up to time T ∈ [0, ∞] if for some a ≥ 0, we have P -almost surely (1.1) Xtx,ϑ := x + ϑ . St = x + Zt 0 ϑu dSu ≥ −a for 0 ≤ t ≤ T . The usual interpretation in mathematical finance is as follows. We consider a financial market with (one riskless and) d risky assets evolving over time; so the random variable St describes (discounted) asset prices at time t. It is well known that semimartingales are the most general possible models for this if some very basic economic postulates on the financial market should be satisfied. The assets can be traded dynamically over time; such activities by an investor are modelled by some ϑ where ϑit is the number of shares of asset i that the investor has decided to hold at time t. Trading must be done in a self-financing way so that price fluctuations in S do not generate or require additional cashflows to keep ϑ going. This translates mathematically into the fact that if the investor starts with initial capital x at time 0, his wealth at time t is given by Xtx,ϑ from (1.1). (For a loose explanation, note that ϑu dSu is the infinitesimal gain or loss at time u when the investor holds the portfolio ϑu and prices move by dSu . By the self-financing condition, changes in wealth occur only in this way; so the wealth at time t is obtained by adding, via the integral, to the initial wealth x all infinitesimal changes.) Finally, the lower bound in (1.1) is a budget constraint designed to eliminate economically unreasonable strategies (e.g., doubling). For a textbook account, we refer to Delbaen/Schachermayer (2006) or Föllmer/Schied (2004). Optimal portfolio choice is the problem of finding, for a given time horizon T < ∞, a strategy with “best” final wealth XTx,ϑ . To quantify this, let U be a utility random field; this is a mapping from [0, ∞) × IR × Ω to IR where we think of U (T, y, ω) as the investor’s subjective quantitative assessment of having the amount y at time T in state ω. Naturally, y 7→ U (T, y, ω) is thus assumed to be increasing and concave, and (ω, T ) 7→ U (T, y, ω) should be a decent stochastic process. It is classical that the optimisation problem (1.2) maximise the expected utility E[U (T, XTx,ϑ )] of final wealth over all admissible strategies up to time T 2 has for each time horizon T a unique solution, under suitable (mild) assumptions on U (and S); see for instance Kramkov/Schachermayer (1999) or Karatzas/Žitković (2003). But how does all that depend on the time horizon? In mathematical terms, we have for bT , the solution of (1.2), which is a stochastic integral of S. More each T a random variable X precisely, the integrand is the optimal strategy for the problem (1.2); so if we denote this by ¡ ¢ ϑ̂(T, u) 0≤u≤T , we have (1.3) bT = X x,ϑ̂(T,·) XT =x+ ZT ϑ̂(T, u) dSu . 0 bT )T ≥0 of optimal wealths — and this apparently Our goal is to study the stochastic process (X bT on T in (1.3). In innocent question is quite difficult because of the double dependence of X bT (ω) is not obvious at all and turns out to fact, even getting some path regularity for T 7→ X involve a lot of work. For the purposes of this section’s overview, we are deliberately vague about the precise assumptions on U (and S); full details are given in later sections when they are needed and used. The only exception to such precision is that we shall simply assume that (1.2) has a unique solution for each T , without further specifics and without concrete sufficient conditions on U . That last point will be addressed in forthcoming work. Our three main results clarify in the general semimartingale setting for S the mathematical structure of horizon-dependence in optimal portfolio choice. We prove: b of optimal wealths has a version X e which is a semimartingale. This 1) The process X bT )T ≥0 can be aggregated into a nice means that the collection of random variables (X stochastic process. ¡ ¢ 2) The two-parameter process ϑ̂(T, u) T ≥0, 0≤u≤T of optimal strategies for (1.2) has a version ϑe which is of finite variation with respect to T . This means that the optimal strategy field depends nicely on the time horizon. e of optimal wealths admits a unique decomposition as the sum of a 3) The semimartingale X stochastic integral of S and a predictable process B̄ of finite variation: P -almost surely, (1.4) eT = x + X ZT ϑ̄u dSu + B̄T 0 for each T ≥ 0, where ϑ̄ = (ϑ̄u )u≥0 is an IRd -valued predictable S-integrable process which does not depend on T . The structure and above all existence of the decomposition (1.4) is quite remarkable. eT is a stochastic integral with respect to S, with an For each T , we know from (1.3) that X 3 e as a stochastic integral of one single integrand depending on T . However, (1.4) gives us X integrand ϑ̄, plus a correction term B̄. This kind of decomposition is conceptually reminiscent of, but technically quite different from, the optional decomposition theorem due to Kramkov e and the more detailed study of the horizon-dependence (1996). For the further analysis of X in (1.2), the representation (1.4) will play a key role. The idea for proving 3) is at least conceptually quite simple. Since a good version ϑe of T 7→ ϑ̂(T, ·) is by 2) of finite variation, we can write (1.5) eT = x + X ZT 0 e u) dSu = x + ϑ(T, ZT 0 e u) dSu + ϑ(u, ZT ZT 0 u e ϑ(dt, u) dSu . e ·). From the last term in (1.5), we expect that Not surprisingly, ϑ̄ is then the diagonal of ϑ(·, “ B̄T = ZT Zt 0 0 e dSu ϑ(dt, u) ” by formally applying Fubini. To prove that this is possible and that the last term in (1.5) indeed defines a predictable process of finite variation, we establish in Section 4 a new stochastic Fubini theorem for (measure-valued) stochastic integrals, with respect to a martingale, of measure-valued integrands. This generalises earlier work by several authors and constitutes another new result which is of independent mathematical interest. The proofs for 1) and 2) are considerably more complicated, and the more difficult result is 1). From classical theory, there are two ways of showing that a given process Y is a semimartingale; see the textbook Protter (2005) for more background. One can try to exhibit a decomposition of Y into a local martingale and a process of finite variation if one knows the dynamics of Y sufficiently well. Alternatively, one can appeal to the deep Bichteler–Dellacherie characterisation result for semimartingales if one can show that Y is a good integrator. The latter means that as a functional on simple (i.e. piecewise constant) predictable processes ψ, the stochastic integral ψ . Y (which is for such ψ defined in an elementary way) has a suitable continuity property. Our idea is to use a variant of the second approach which has recently come up in a line of work going back to S. Ankirchner and P. Imkeller: If one has an unbounded increasing function U (to be thought of as a utility function) such that E[U (ψ . YT )] remains bounded over all simple predictable ψ, then Y is a semimartingale (on [0, T ]); see Ankirchner (2005) and Ankirchner/Imkeller (2005). However, there are two major obstacles. The first is that such a result is only available if the utility U (T, y, ω) depends only on y and, more importantly, if Y is locally bounded. We have no such b and we also do not want to restrict ourselves to unbounded utilities. information on our X, Thus we go back to Delbaen/Schachermayer (1994) who essentially already showed that a locally bounded process is a semimartingale if the family of final values of all its stochastic 4 integrals over simple bounded admissible integrands is bounded in probability. We exploit b as process of wealths to translate (in Proposition 2.10) simple admissible the structure of X b into (general, admissible) stochastic integrals over S. Then we stochastic integrals over X deal with the big jumps, and finally, an indirect argument establishes the semimartingale b in Theorem 2.14. property of X But there is a second, fundamental obstacle. All the above results start from a process Y which is adapted and RCLL (i.e. has trajectories which are right-continuous and admit left bT )T ≥0 is just a family of random variables for which we have a priori limits). In contrast, (X no more information than the structure (1.3) for each T . So the very first thing we must do is b has a version X e with RCLL trajectories. This crucially exploits (1.3). We to establish that X bT is right-continuous in probability (Theorem 2.8), by adapting a similar first show that T 7→ X x,ϑ̂(T,·) argument from Kramkov/Schachermayer (1999) who established continuity of XT with bT admits right and left limits respect to x. We then prove (Proposition 2.11) that T 7→ X along the rationals, combining to that end (in Lemma 2.1) classical ideas from martingale theory and mathematical finance in a novel way. Putting all these results together finally e of X, b and then we can proceed yields (Proposition 2.13) the existence of an RCLL version X along the lines sketched above. As is well known, stochastic integrals are often easier to handle when the integrator is not a semimartingale, but a martingale. This is because one has for martingales and their stochastic integrals an isometry property and several classical inequalities. To exploit that, we show (in Proposition 2.7) how one can obtain the martingale property (and even very good integrability properties) for S by passing to a well-chosen equivalent probability measure Qm — at least locally, up to an arbitrarily large stopping time Tm . This important auxiliary result is used in several places in the paper, by working under Qm instead of under P . Finally, it remains to establish the result 2) concerning the structure of ϑ̂. To that end, we view ϑ̂(T, ω, u) as an IRd -valued process indexed by T and defined on Ω̄ = Ω × (0, ∞), and we aim at proving that it is a semimartingale. (Since the filtration on Ω̄ is chosen constant, the b we have to start by obtaining some finite variation property is then immediate.) Like for X, bT , we first prove regularity for ϑ̂. Similarly to the right-continuity in probability of T 7→ X that T 7→ ϑ̂(T, ·) is right-continuous in measure (Lemma 3.2) and has a product-measurable ¡ ¢ is a version ϑe on [0, ∞) × Ω̄ (Lemma 3.3). This easily implies that the diagonal ϑ̂(T, T ) T ≥0 predictable S-integrable process (Corollary 3.6). To obtain the semimartingale property of ϑ̂ e ·) is (locally, under an equivalent with respect to T , we prove (Proposition 3.10) that T 7→ ϑ(T, measure Qm ) a quasimartingale on Ω̄, by first establishing (Proposition 3.7) that it has good continuity properties as an integrator for piecewise constant integrands (with respect to T , over Ω̄). The quasimartingale property then gives the existence of a right-continuous version, and the above continuity property in turn yields by the classical Bichteler–Dellacherie result e ·) is a semimartingale. that T 7→ ϑ(T, 5 2. The process of optimal wealths has a semimartingale version In this section, we prove that the process bt = X x,ϑ̂(t,·) Xt = x + ϑ̂(t, ·) . St = x + Zt 0 ϑ̂(t, u) dSu , t≥0 e which is a semimartingale. This will of optimal wealths over time from (1.3) has a version X be done in Section 2.2, while Section 2.1 prepares the ground with some results of general interest. The main idea and overall structure of the argument are explained in Section 1, but the results are presented here in a different order for reasons of proof efficiency. Since T is b often used as a fixed time horizon in this section, we have changed the running index for X from T to t. 2.1. Some general results We always work on a filtered probability space (Ω, F , IF, P ) with IF = (Ft )t≥0 satisfying W the usual conditions of right-continuity and P -completeness, and set F∞ = t≥0 Ft . Given a ≥ 0 and an arbitrary IRd -valued process Y = (Yt )t≥0 , a simple (a-)admissible integrand I P hi I]]τi−1 ,τi ]] with I ∈ IN , each hi IRd -valued and for Y is a process of the form H = i=1 Fτi−1 -measurable and stopping times 0 ≤ τ0 ≤ τ1 ≤ · · · ≤ τI < ∞ such that the real-valued I R P τi − Y τi−1 ) is uniformly bounded from stochastic integral process H . Y = H dY = htr i (Y i=1 below (by −a, respectively). If all the hi are constant vectors ci ∈ IRd , we call H very simple. Like Delbaen/Schachermayer (1994), we say that Y satisfies (NFLVR) for (very) simple admissible integrands if H n . YT → 0 in L0 for every T ∈ [0, ∞], whenever H n is for each n a (very) simple εn -admissible integrand for Y and (εn )n∈IN decreases to 0. Moreover, in analogy to Karatzas/Kardaras (2007), we say that Y satisfies (NUPBR) for (very) simple a-admissible integrands if the family {H . YT | H (very) simple a-admissible integrand for Y } of random variables is bounded in L0 for every T ∈ [0, ∞]. Note that the value of a > 0 is not important for (NUPBR); all that matters is that we impose the same negative uniform lower bound on H . Y for all H. There is in Delbaen/Schachermayer (1994) also a notion of (NFLVR) with general integrands and we shall use that later. If S = (St )t≥0 is an IRd -valued semimartingale, we say that S satisfies (NFLVR) if ϑn . ST → 0 in L0 for every T ∈ [0, ∞], whenever ϑn is for each n an IRd -valued predictable S-integrable process which is εn -admissible for S (in the usual sense that ϑn . S ≥ −εn ) and (εn )n∈IN decreases to 0. Note that we allow here general (as opposed to simple) integrands, and so we have to assume a priori that S is a semimartingale. 6 Lemma 2.1. Suppose Y = (Yt )t≥0 is an adapted IRd -valued process and uniformly bounded from below in each coordinate. If Y satisfies (NFLVR) for very simple admissible integrands or (NUPBR) for very simple 1-admissible integrands, then Y admits with probability 1 right and left limits along the rationals, i.e. there exists a set C of probability 1 such that for all I + increases or decreases to some ω ∈ C, lim Xrn (ω) exists in IR whenever (rn )n∈IN ⊆ Q n→∞ r ∈ [0, ∞). If Y is progressively measurable, then Y even admits with probability 1 right and left limits (i.e. along monotonic sequences (rn )n∈IN ⊆ IR+ ). Proof. By working with each coordinate, we can and do assume without loss of generality that Y is real-valued. We combine classical ideas from Dellacherie/Meyer (1982) and Delbaen/ Schachermayer (1994), and in particular follow in parts closely the proof of Theorem VI.48 in Dellacherie/Meyer (1982). Assume first only that Y is adapted. Define the processes Ut := lim inf Yr , r&t, r∈I Q Vt := lim sup Yr , r&t, r∈I Q t≥0 and note that U and V are progressively measurable by Theorem IV.17 in Dellacherie/Meyer (1978). To prove the P -a.s. existence of right limits along the rationals, we want to show that the set {(ω, t) | Ut (ω) < Vt (ω)} is evanescent, or equivalently that for all rationals a < b, the set E := {(ω, t) | Ut (ω) < a, Vt (ω) > b} is evanescent. This set is progressive, so its debut D is a stopping time, and we want to reach a contradiction from assuming that P [D < ∞] 6= 0. As in the proof of Theorem VI.48 in Dellacherie/Meyer (1982), let A0 := {(ω, t) | D(ω) < t < D(ω) + 1, Yt (ω) < a}. Then the measure of the projection of A0 on Ω is P [D < ∞] because D is the debut of E. Choose an optional cross-section τ0 of A0 such that P [τ0 < ∞] > (1 − 12 )P [D < ∞]. Let £ ¢ ¡ A1 := {(ω, t) | D(ω) < t < D(ω) + 12 ∧ τ0 (ω), Yt (ω) > b} ¤ note the small typo in Dellacherie/Meyer (1982)! , and choose an optional cross-section τ1 of A1 such that P [τ1 < ∞] > (1 − 14 )P [D < ∞]. Let ¡ ¢ A2 := {(ω, t) | D(ω) < t < D(ω) + 14 ∧ τ1 (ω), Yt (ω) < a}, and choose an optional cross-section τ2 of A2 such that P [τ2 < ∞] > (1 − 18 )P [D < ∞], etc. In this way, we construct stopping times τi with P [τi < ∞] > (1 − 2−(i+1) )P [D < ∞], τi ≡ +∞ on {D = ∞}, τi < τi−1 on {τi < ∞}, and τi < D + 2−i on {τi < ∞}. Since then P [τi = ∞, D < ∞] = P [D < ∞] − P [τi < ∞] ≤ P [D < ∞] 2−(i+1) , Borel-Cantelli implies that almost surely on {D < ∞}, we have τi < ∞ for all large i. So there exists with probability 1 on {D < ∞} a random n0 ∈ IN such that (τi )i≥n0 decreases to D and for i ≥ n0 , we have τ2i−1 < ∞ and τ2i < ∞, hence also Yτ2i ≤ a and Yτ2i−1 ≥ b. 7 Now we deviate from Dellacherie/Meyer (1982) and switch from martingale theory to 2n P cn I]]τ2i ,τ2i−1 ]] mathematical finance. Take T = ∞, define very simple integrands H n := i=n with constants cn ≥ 0 still to be chosen, and set Y∞ := 0. Then Yτ2i−1 − Yτ2i implies that H n. Y∞ = 2n X i=n 0 = Y∞ − Yτ2i ≥ −a Y τ2i−1 − Yτ2i ≥ b − a if τ2i = ∞ if τ2i < ∞, but τ2i−1 = ∞ if τ2i−1 < ∞ cn (Yτ2i−1 − Yτ2i ) 2n X ¯ ª ¯ I{τ2i−1 =∞,τ2i <∞} . ≥ cn (b − a)# i ∈ {n, . . . , 2n} τ2i−1 < ∞ + cn (−a) © i=n The last sum is always 0 or 1 so that the second summand above is always at least −cn |a|. For ω ∈ {D < ∞} and i ≥ n0 (ω), we have P -a.s. τ2i−1 (ω) < ∞; hence H n . Y∞ (ω) ≥ cn (b − a)(n + 1) − cn |a| for n ≥ n0 (ω), and this shows that if cn → 0 and (n + 1)cn → ∞, then H n . Y∞ −→ +∞ (2.1) P -a.s. on {D < ∞}. The same argument applies for any t instead of ∞ if we truncate all the τi at t, i.e. replace τi by τi ∧ t in the above computations. For admissibility of H n , the same computations for Yτ2i−1 ∧t − Yτ2i ∧t lead to H n. 2n X ¯ ª ¯ Yt ≥ cn (b − a)# i ∈ {n, . . . , 2n} τ2i−1 ≤ t + cn (Yt − a) I{τ2i−1 >t,τ2i ≤t} . © i=n The first term is ≥ 0 since cn ≥ 0, and the second term is 0 or cn (Yt − a). So if −y with y > 0 is the uniform lower bound for Y , we see from H n . Y ≥ −cn (y + |a|) =: −εn that each H n is √ εn -admissible for Y , and εn & 0 if cn & 0. Choosing for instance cn = 1/ n + 1 therefore produces with the corresponding (H n ) a sequence of very simple integrands that violates via (2.1) both properties (NFLVR) and (NUPBR) if P [D < ∞] > 0; note that (εn ) is bounded since it converges, and so we have a uniform lower bound for H n . Y simultaneously for all n. This proves existence of right limits along the rationals, and left limits along the rationals are obtained analogously, which ends the proof for the adapted case. 8 If Y is even progressive, the same arguments can be made for the processes Ut0 := lim inf Ys , s&t Vt0 := lim sup Ys , s&t t ≥ 0, since these are then progressive by Theorem IV.33 of Dellacherie/Meyer (1978). This yields the second assertion. q.e.d. The next result is very simple but useful. Lemma 2.2. If a process Y admits right limits along the rationals and is right-continuous in probability, then Y has a version Ye with right-continuous trajectories. Proof. For each t ≥ 0, set Yet := lim r&t, r∈I Q Yr ; this is well defined P -a.s. by the assumption that Y admits right limits along the rationals. For any t ≥ 0 and rationals rn & t, we get Yrn → Yt in L0 by right-continuity in probability, and Yrn → Yet P -a.s. by definition; so Yet = Yt P -a.s. and Ye is a version of Y . By construction, Ye is P -a.s. right-continuous. q.e.d. The next result is very useful for establishing the semimartingale property of a process, in general as well as later in Theorem 2.14. Lemma 2.3. Suppose Y = (Yt )t≥0 is a real-valued adapted locally bounded process and fix T ∈ (0, ∞). If Y is not a semimartingale, then one can find a sequence of simple predictable processes |H n | ≤ 1 such that all the H n are c-admissible for Y for some c > 0 and the sequence (H n . YT )n∈IN is not bounded in L0 . Proof. Our argument follows very closely the work done by F. Delbaen and W. Schachermayer in Section 7 of Delbaen/Schachermayer (1994), but the result itself is not explicitly there. By stopping, we can assume without loss of generality that Y is bounded, and we can also normalise so that |Y | ≤ 1. It is clear that if we then construct the H n to be uniformly bounded (in n as well), we can normalise that bound to 1, too. Now we start with the set G := (m−1 X k=0 (Yτk+1 ¯ ) ¯ ¯ − Yτk )2 ¯ m ∈ IN, 0 ≤ τ0 ≤ τ1 ≤ · · · ≤ τm ≤ T stopping times ¯ and, for a given element of G, put H := −2 and H . Yt = m−1 P k=0 m−1 P k=0 Yτk I]]τk ,τk+1 ]] . Then |H| ≤ 2 since |Y | ≤ 1, (Yτk+1 ∧t − Yτk ∧t )2 − Yτ2m ∧t + Yτ20 ∧t ≥ −1 shows that H is 1-admissible for 9 Y . If G is not bounded in L0 , we can thus take a sequence (H n ) of the above form such that (H n . YT )n∈IN ⊆ G is not bounded in L0 , and we are already done. So suppose from now on that G is bounded in L0 . To follow the rest of the proof, a copy of Section 7 of Delbaen/Schachermayer (1994) should be ready to hand. Start there with the proof of Theorem 7.2, assuming that Y (called X there) is not a semimartingale, and produce a first sequence (H n ) as on page 506, lines 3–10. Use then that G is bounded in L0 to produce a new sequence (K n ) as on page 506, line 20. Go on to construct a next sequence e n . YT )n∈IN is then unbounded in L0 . Note also e n ) as on page 507, line 3, and note that (K (K e n | ≤ 1 for all n. that |K Now go to the proof of Lemma 7.3 in Delbaen/Schachermayer (1994) on page 505, and recall that H there is just some family of simple predictable processes bounded by 1, like e n instead of the H n in Delbaen/Schachermayer e n , n ∈ IN , from above. Use these K the K ¡ e n I]]0,T ]] . These (1994) on page 505, line 2; construct Tn0 , Un and then Tn , and set H̄ n := K n H̄ n n correspond to the H I]]0,Tn ]] on page 505, lines 11 and 12, of Delbaen/Schachermayer ¢ (1994). Then property (a) on page 505, line 11, dropping δn , shows that (H̄ n . YT )− , n ∈ IN , is bounded; but even more, the estimate on page 505, line 7, and Tn ≤ Un imply that H̄ n . Y ≥ −K − 2 so that each H̄ n is (K + 2)-admissible for Y . Finally, property (b) on page 505, line 12, with cn → ∞ and dropping δn shows that (H̄ n . YT )n∈IN is not bounded in L0 . So up to normalising, these H̄ n give our sequence. q.e.d. At this point, we have to introduce our assumptions on the basic process S. As explained in part I of the textbook Delbaen/Schachermayer (2006), the standard approach in mathematical finance would be to suppose that S is a semimartingale and satisfies (NFLVR). The latter condition is motivated from economic postulates, and by the fundamental theorem of asset pricing, it is equivalent to the existence of an equivalent σ-martingale measure for S, which is a probability Q equivalent to P such that S becomes a Q-σ-martingale (i.e. a stochastic integral S = ψ . M , with a one-dimensional integrand ψ > 0, of an IRd -valued local Q-martingale M ); see Delbaen/Schachermayer (1998). Also equivalently, there exists ¯ ¯ a strictly positive P -martingale Z up to ∞ (namely, the process of densities Zt = dQ dP Ft , t ∈ [0, ∞], of Q with respect to P along IF ) such that the product ZS is a P -σ-martingale on [0, ∞); this is just rewriting things from Q to P via the Bayes rule. Our conditions on S will be slightly different. We also assume that S is a semimartingale, and we introduce for fixed x > 0 and each stopping time τ the set ¯ © Xτx := Xτx,ϑ = x + ϑ . Sτ ¯ ϑ is IRd -valued, predictable, S-integrable ª and x-admissible for S up to time τ . Then (NFLVR) is equivalent to the combination of the following two conditions: 0 intersects L0+ only in 0. (2.2) X∞ 10 (2.3) For each (deterministic) T ∈ [0, ∞) and some (or equivalently all) x > 0, the set XTx is bounded in L0 . The above equivalence can be found in Section 3 of Delbaen/Schachermayer (1994) or more succinctly in Lemma 2.2 of Kabanov (1997). The property (2.2) is usually called (NA) for no arbitrage, while (2.3) has been blandly called BK in Kabanov (1997) or labelled (NUPBR), for no unbounded profit with bounded risk , in Karatzas/Kardaras (2007). Instead of (2.3), we consider (2.4) For each bounded stopping time τ and some (or equivalently all) x > 0, the set Xτx is bounded in L0 . We also introduce the localised version (2.5) There exists a sequence (σk )k∈IN of stopping times increasing to ∞ and such that for each bounded stopping time τ and some (or equivalently all) x > 0, the set Xτx∧σk is bounded in L0 for each k ∈ IN . Instead of (NFLVR), we shall need a variant of the existence of an equivalent σ-martingale measure Q for S as above. More precisely, a σ-martingale density for S (and P ) is a local (P -)martingale Z = (Zt )t≥0 with Z0 = 1 and Z ≥ 0 and such that ZS is a (P -)σ-martingale. We denote by Da,σ (S, P ) the family of all σ-martingale densities Z for S and P , and by Za,σ (S, P ) the subset of all those Z with the additional property that Z log Z is locally (P -)integrable; so the semimartingale Z log Z is then P -special which means that its running supremum process (Z log Z)∗t = sup |Zs log Zs |, t ≥ 0, is locally P -integrable. Finally we 0≤s≤t introduce De,σ (S, P ) := {Z ∈ Da,σ (S, P ) | Z > 0} and Ze,σ (S, P ) := Za,σ (S, P ) ∩ De,σ (S, P ) = {Z ∈ Za,σ (S, P ) | Z > 0}. Our main condition on S is that Ze,σ (S, P ) is nonempty, and this is simultaneously a bit stronger and a bit weaker than the existence of an equivalent σ-martingale measure Q for S. Indeed, the density process Z Q would have stronger global properties, because it is a true P -martingale up to ∞ while Z ∈ Ze,σ (S, P ) is only a local P -martingale; but the local integrability of Z log Z for Z is more than what we have for Z Q . Let us briefly clarify how the above concepts and conditions are related. To that end, we first recall two facts that will be used several times in the sequel: (2.6) If Y is an IRm -valued σ-martingale and ψ is an IRm -valued predictable and Y -integrable R process, then also ψ . Y = ψ dY is a (real-valued) σ-martingale. (2.7) Any real-valued σ-martingale uniformly bounded from below is a local martingale. If its initial value is in L1 (e.g. if it is null at 0), then it is also a supermartingale. The result in (2.6) can be found in Lemma 3.3 of Kallsen (2003) or Proposition III.6.42 of Jacod/Shiryaev (2003). The first assertion of (2.7) follows directly from Corollaire 3.5 of Ansel/Stricker (1994); the second one is then a well-known consequence of Fatou’s lemma. 11 Lemma 2.4. For an IRd -valued semimartingale S, conditions (2.4) and (2.5) are equivalent, and follow from the condition that De,σ (S, P ) is nonempty. Proof. To get (2.5) from (2.4), it is enough to take σk = k. For the converse, note that (2.8) P [Xτx,ϑ ≥ α] = P [Xτx,ϑ ≥ α, τ ≤ σk ] + P [Xτx,ϑ ≥ α, τ > σk ] ≤ P [Xτx,ϑ ∧σk ≥ α] + P [σk < τ ]. If ϑ is x-admissible for S up to τ , it is also x-admissible for S up to τ ∧ σk . So after we take the supremum over all the former ϑ, the first term on the right-hand side of (2.8) goes to 0 for each fixed k as α → ∞, due to (2.5). Since σk % ∞ and τ is bounded, the second term P [σk < τ ] can be made arbitrarily small for k large. Hence (2.4) follows from (2.5). Finally we show that De,σ (S, P ) 6= ∅ implies (2.5). Take Z ∈ De,σ (S, P ) and a localising sequence (σk )k∈IN such that each Z σk is a uniformly integrable P -martingale. Then setting dQk := Zσk dP defines a probability measure Qk equivalent to P since Z > 0, and Z σk S σk = (ZS)σk is like ZS a P -σ-martingale. By the Bayes rule, this means that S σk is a Qk -σ-martingale; see Proposition 5.1 of Kallsen (2003). For any ϑ which is S-integrable, hence S σk -integrable, ϑ . S σk is thus a Qk -σ-martingale by (2.6); so if ϑ is also x-admissible for S up to σk , then ϑ . S σk is a Qk -supermartingale by (2.7). For such ϑ, therefore, EQk [x + ϑ . Sτ ∧σk ] = EQk [x + ϑ . Sτσk ] ≤ x for any bounded stopping time τ . This shows that Xτx∧σk is bounded in L1 (Qk ), hence also in L0 (Qk ), which equals L0 (P ) because Qk is equivalent to P . q.e.d. Our key assumption on S is now that S is an IRd -valued semimartingale and the set Ze,σ (S, P ) is nonempty. (2.9) This clearly implies De,σ (S, P ) 6= ∅ and hence (2.4) by Lemma 2.4. Our next auxiliary result is similar to Lemma 3.2 of Delbaen/Schachermayer (1994). Recall that X x,ϑ = x + ϑ . S ≥ 0 if ϑ is x-admissible for S. Lemma 2.5. Suppose that the IRd -valued semimartingale S satisfies (2.4). For any fixed T ∈ (0, ∞), the family ¯ ½ ¾ ¯ x,ϑ ¯ WT := sup Xt ¯ ϑ predictable and x-admissible for S, τ ≤ T stopping time 0≤t≤τ is then bounded in L0 . As a consequence, W (T ) := ess sup WT is finite P -a.s. n Proof. Suppose WT is not bounded in L0 . Take a sequence W n = sup Xtx,ϑ in WT and 0≤t≤τn n some δ > 0 such that P [W > n] ≥ δ for all n. With the stopping time ¯ © ª n σn := inf t ≥ 0 ¯ Xtx,ϑ > n ∧ τn ≤ T, 12 we clearly have © ª © ª © ª n n n n > n ⊆ Xσx,ϑ ≥ n or Xτx,ϑ > n ⊆ Xσx,ϑ {W n > n} ⊆ {σn < τn } ∪ Xτx,ϑ ≥n . n ∧τn n n n So if we set ϕn := ϑn I]]0,σn ∧τn ]] , we get x + ϕn . S ≥ 0 since ϑn is x-admissible, and P [x + ϕn . ST ≥ n] ≥ P [W n > n] ≥ δ. But this means that the family {x + ϕ . ST | ϕ is predictable and x-admissible} ⊆ XTx is not bounded in L0 , which contradicts the property (2.4) for S. q.e.d. The next result is an already announced key technical tool in our analysis. It allows us to obtain good martingale and integrability properties by a suitable change of measure combined with stopping. For its formulation, we first need some notations, and we start by recalling the set Za,σ (S, P ) from above. If the local P -martingale Z ≥ 0 with Z0 = 1 is written as Z = E(N ) with a local (P -)martingale N null at 0 and such that 1 + ∆N ≥ 0, we define ¢ P ¡ (1 + ∆Ns ) log(1 + ∆Ns ) − ∆Ns . For Z ∈ Za,σ (S, P ), the process V (N ) := 12 hN c i + 0<s≤ · the process V (N ) is locally (P -)integrable; the entropy-Hellinger process of Z with respect to P is then the dual predictable projection of V (N ) with respect to P , and we denote this predictable RCLL increasing process by hE (Z, P ). Finally, the minimal entropy-Hellinger σ-martingale density for (S, P ) is that element Z ∗ of Za,σ (S, P ) – if it exists – which minimises Z→ 7 hE (Z, P ) over Z ∈ Za,σ (S, P ) in the sense of the strong order, i.e. hE (Z ∗ , P ) − hE (Z, P ) is increasing for all Z ∈ Za,σ (S, P ). If Z ∗ is a P -uniformly integrable P -martingale, the ∗ dP is called the minimal entropycorresponding probability measure Q∗ with dQ∗ = Z∞ Hellinger σ-martingale measure for (S, P ). More details can be found in Choulli/Stricker (2005, 2006). In several of our proofs, we have to change from P to an equivalent probability measure Q which endows S with better properties. While L0 , hence also (2.4), and the semimartingale property of S are invariant under any such change of measure, condition (2.9) that Ze,σ (S, P ) 6= ∅ involves a local integrability condition and is thus more delicate. It is there- fore quite remarkable that we have the following stability result. Theorem 2.6. Let S be an IRd -valued semimartingale and suppose that the probability measure Q is equivalent to P . If Ze,σ (S, P ) 6= ∅ and if the density dQ dP is bounded, then also Ze,σ (S, Q) 6= ∅. The proof of Theorem 2.6 is too long to be given here; it takes a paper on its own, and so we refer to Choulli/Schweizer (2010) for details. However, we point out that all the difficulties come from the fact that the stochastic logarithm N of a strictly positive local P -martingale Z = E(N ) can have jumps whose size comes arbitrarily close to −1, so that 13 the term log(1 + ∆N ), which appears in the definition of V (N ), is hard to control. If the filtration IF does not contain any discontinuous martingales (e.g., if it is generated by some possibly multidimensional Brownian motion), this problem does not occur, and the result in Theorem 2.6 is then obvious. Here is then the announced technical result for good measure changes. Proposition 2.7. Suppose that the IRd -valued semimartingale S satisfies (2.9). Let W be any nonnegative random variable with W < ∞ P -a.s. For any fixed T ∈ (0, ∞), the following hold: 1) There exists a probability measure Q ≈ P with bounded density (2.10) µ 2 2 exp W + sup |∆St | + sup |St | 0<t≤T 0≤t≤T 2 ¶ dQ dP such that ∈ L1 (Q) and the minimal entropy-Hellinger σ-martingale density Ze for (S T , Q) exists. 2) There exist a sequence of stopping times (Tm )m∈IN increasing stationarily to T and a sequence of probability measures Qm ≈ P with Qm = Qm+1 on FTm such that for each m ∈ IN , S Tm is a Qm -square-integrable Qm -martingale and W ∈ L2 (Qm ). (As a matter of fact, dQm = ZeTm dQ, as can be seen in the proof.) Proof. Since the arguments for Proposition 2.7 are not used in the sequel, this proof is relegated to the Appendix. q.e.d. b 2.2. The semimartingale property of X We now fix x > 0 and study in more detail the process bt = X x,ϑ̂(t,·) Xt =x+ Zt ϑ̂(t, u) dSu , 0 t≥0 of optimal wealths in order to prove that it admits a version which is a semimartingale. A very rough outline of the argument is as follows: b is right-continuous in probability. 1) We first prove that X b satisfies (NUPBR) for simple integrands which are x-admissible 2) We then show that X b and deduce from Lemmas 2.1 and 2.2 that X b has an RCLL version X. e for X, e the big jumps to get a locally bounded process X̄. Assuming that 3) We subtract from X X̄ is not a semimartingale, we use Lemma 2.3 to construct a sequence of strategies that will be shown to contradict the property (2.4) of S. Steps 2) and 3) also require an auxiliary result, where in particular Proposition 2.7 turns out to be very useful. 14 Before embarking on that route, we make precise our assumptions on the utility field U . Recalling that (t, y, ω) 7→ U (t, y, ω) is a mapping from [0, ∞) × IR+ × Ω to IR, we assume that for fixed (ω, t), y 7→ U (t, y, ω) is a standard utility function on IR+ , i.e. strictly concave, strictly increasing, in C 1 , and satisfying the Inada conditions Uy (t, 0, ω) = +∞, Uy (t, ∞, ω) = 0. In addition, we shall need: (2.11) The processes (ω, t) 7→ U (t, 0, ω)I{U (t,0,ω)>−∞} and (ω, t) 7→ U (t, b, ω), for any constant b > 0, are adapted, right-continuous and of locally integrable variation. This condition is easy to verify if there is for example no dependence on ω. We also need: (2.12) For any uniformly bounded sequence of stopping times (τn )n∈IN decreasing to some stopping time τ and any a, b with 0 < a ≤ b < ∞, we have lim sup |U (τn , y, ω) − U (τ, y, ω)| = 0 n→∞ a≤y≤b P -a.s. This essentially says that U is right-continuous with respect to t, locally uniformly in y. Finally, as a joint condition on U and S, we need the earlier mentioned basic assumption that the utility maximisation problem is well posed in the following sense: (2.13a) For each bounded stopping time τ , the problem of maximising the expected utility E[U (τ, g)] over all g ∈ Xτx has a finite value and a unique solution ĝ. We then take an optimal strategy denoted by ϑ̂(τ, ·) so that ¢¤ £ ¡ sup E[U (τ, g)] = E[U (τ, ĝ)] = E U τ, Xτx,ϑ̂(τ,·) < ∞, (2.13b) g∈Xτx x,ϑ̂(τ,·) bτ := ĝ = Xτ . The combination of (2.13a) and (2.13b) is denoted by (2.13), and we set X and this assumption deserves a few comments. First of all, it is reasonable in view of our problem — if we can achieve infinite expected utility for some time horizon, something is wrong with our model at a fundamental level. Secondly, there are known sufficient condi- tions on U and S in a number of cases; see for instance Kramkov/Schachermayer (1999) or Karatzas/Žitković (2003). Finally, note that while the optimal wealth ĝ is naturally unique by the strict concavity of U with respect to y, there can be several strategies leading to ĝ. We take here for ϑ̂(τ, ·) one of them, and we shall come back to the issues of nonuniqueness and of a good choice later in Section 3. b is right-continuous in probability. One inspiration why such We now first prove that X a result could hold comes from Kramkov/Sı̂rbu (2006) who showed that with respect to x, x,ϑ̂(t,·) Xt x,ϑ̂(t,·) Xt is even differentiable (if U is independent of (ω, t) and C 2 with respect to y). Since Rt = x + ϑ̂(t, u) dSu depends on t in a much more complicated way than on x, our 0 result is conceptually quite different from that of Kramkov/Sı̂rbu (2006), but we can still 15 explain the intuition behind it. Increasing t (or τ ) a little has the effect of adding a stochastic integral, which is a right-continuous operation since the stochastic integral is, and this rightcontinuity is then preserved thanks to the assumptions (2.11) and (2.12) on U . Formally, our proof is modified from Lemma 3.6 in Kramkov/Schachermayer (1999), which establishes a similar continuity result (but also with respect to x) for the solution of the dual problem. Theorem 2.8. Suppose that the IRd -valued semimartingale S satisfies (2.9), and U satisfies (2.11)–(2.13). Let (τn )n∈IN be any sequence of stopping times bounded by some T ∈ (0, ∞). bτ in L0 . bτ )n∈IN converges to X If (τn ) decreases to a stopping time τ , then (X n Proof. 1) Take Z ∈ De,σ (S, P ) which is nonempty due to (2.9). As in the proof of Lemma 2.4, let (σk )k∈IN be a localising sequence for Z, define Qk by dQk := Zσk dP and recall that S σk is then a Qk -σ-martingale. For any ε > 0, ¤ ¤ £ ¤ £ £ bτ − X bτ − X b τ | ≥ ε = P |X bτ | ≥ ε, τn ≤ σk + P |X bτ | ≥ ε, τn > σk bτ − X P |X n n n £ ¤ bτ ∧σ − X bτ ∧σ | ≥ ε + P [σk < τ1 ] ≤ P |X n k k because (τn ) is decreasing, and since σk % ∞ and τ1 is bounded, the last summand can be bτ in L0 follows if we bτ → X made arbitrarily small by making k large. So the convergence X n bτ ∧σ in L0 as n → ∞. But in view of the definition bτ ∧σ → X show for each fixed k that X n k k b and because = x + ϑ . S%∧σk = x + ϑ . S%σk for each stopping time %, this means of X that we can in effect argue for each fixed k as if we were working with S σk instead of S. 2) To simplify notations in the rest of the proof, we now fix k and write %0 := % ∧ σk for each stopping time %. As seen in step 1), the process S 0 := S σk is a Qk -σ-martingale, or Qk x,ϑ X%∧σ k is an equivalent σ-martingale measure for S 0 . By the fundamental theorem of asset pricing, this means that S 0 satisfies (NFLVR); see Delbaen/Schachermayer (1998). 3) For each bounded stopping time %, define the set C(%0 , x) := X%x0 − L0+ (F%0 ) := {x + ϑ . S%0 − Y | ϑ is x-admissible for S on [[0, %0 ]], Y ∈ L0+ (F%0 )} = {x + ϑ . S%0 0 − Y | ϑ is x-admissible for S 0 on [[0, %0 ]], Y ∈ L0+ (F%0 )}. As in Proposition 3.1 of Kramkov/Schachermayer (1999), each C(%0 , x) is convex and closed T C(%0n , x) = C(%0 , x) if (%n ) decreases to %; in L0 because S 0 satisfies (NFLVR). Moreover, n∈IN T indeed, each g from the intersection is measurable with respect to F%0n = F%0 , and for n∈IN any equivalent σ-martingale measure Q for S and any ϑ admissible for S 0 , the process ϑ . S 0 is by (2.7) a Q-supermartingale so that 0 g = EQ [g | F%0 ] = EQ [x + ϑn . S%0 0n − Y n | F%0 ] ≤ x + ϑn . S%0 0 − EQ [Y n | F%0 ]. 16 0 x,ϑ̂(% ,·) b%0 because U is Finally, the maximiser of E[U (%0 , g)] over g ∈ C(%0 , x) must be X%0 =X strictly increasing with respect to y. bτ 0 ) does not converge to X bτ 0 in L0 so that for some δ > 0, 4) Suppose now that (X n ¤ £ bτ 0 | ≥ δ ≥ δ for infinitely many n. As in the proof of Lemma 3.6 of bτ 0 − X we have P |X n bτ 0 ) so that ḡn ∈ C(τn0 , x), and use bτ 0 + X Kramkov/Schachermayer (1999), define ḡn := 12 (X n bτ 0 ) + 1 U (τn0 , X bτ 0 ) for all strict concavity of U with respect to y to get U (τn0 , ḡn ) ≥ 1 U (τn0 , X 2 n, as well as n 2 ¤ £ bτ 0 ) + 1 U (τn0 , X bτ 0 ) + η ≥ η P U (τn0 , ḡn ) ≥ 12 U (τn0 , X 2 n for some η > 0 and for infinitely many n. By passing to a subsequence (gn ) that we again index by n for ease of notation, we thus obtain bτ 0 )] + 1 E[U (τn0 , X bτ 0 )] + η 2 ≥ E[U (τn0 , X bτ 0 )] + η 2 E[U (τn0 , gn )] ≥ 12 E[U (τn0 , X 2 n bτ 0 is optimal in C(τ 0 , x) ⊇ C(τ 0 , x). But as we shall argue below in step for all n, because X n n bτ 0 )] = E[U (τ 0 , X bτ 0 )], and so 5), assumption (2.12) implies that lim E[U (τn0 , X n→∞ bτ 0 )] + η 2 . lim inf E[U (τn0 , gn )] ≥ E[U (τ 0 , X (2.14) n→∞ Again by step 5) below, lim E[U (τn0 , gn ) − U (τ 0 , gn )] = 0, and because we always have n→∞ lim inf (an + bn ) ≥ lim inf an + lim inf bn , we obtain from (2.14) that n→∞ n→∞ n→∞ bτ 0 )] + η 2 . lim inf E[U (τ 0 , gn )] ≥ E[U (τ 0 , X (2.15) n→∞ Now since all the gn are nonnegative, the standard Komlós trick from Lemma A1.1 of Delbaen/Schachermayer (1994) gives for all n ∈ IN some g̃n ∈ conv(gn , gn+1 , . . .) with (g̃n )n∈IN converging P -a.s. to some g̃∞ . By the closedness in L0 of each C(%0 , x), this g̃∞ lies T C(τn0 , x) = C(τ 0 , x) since like the gn , gn+1 , . . ., each g̃n lies in the convex set C(τn0 , x). in n∈IN The concavity of U with respect to y yields E[U (τ 0 , g̃n )] ≥ inf E[U (τ 0 , gm )], and since the m≥n last term is monotonic in n, letting n → ∞ and using (2.15) yields bτ 0 )] + η 2 . lim inf E[U (τ 0 , g̃n )] ≥ lim inf E[U (τ 0 , gn )] ≥ E[U (τ 0 , X n→∞ n→∞ Again using (2.12), we get from step 5) below that lim E[U (τ 0 , g̃n )] = E[U (τ 0 , g̃∞ )]. Hence n→∞ 0 we find that g̃∞ ∈ C(τ , x) satisfies bτ 0 )] + η 2 , E[U (τ 0 , g̃∞ )] = lim E[U (τ 0 , g̃n )] ≥ lim inf E[U (τ 0 , gn )] ≥ E[U (τ 0 , X n→∞ n→∞ bτ 0 )n∈IN converges to X bτ 0 in L0 . bτ 0 . Thus (X which clearly contradicts the optimality of X n 17 5) It remains to argue the convergence assertions left open in step 4). For brevity, 0 . Suppose (Yn )n∈IN is any sequence of nonnegative random variables with write τ 0 =: τ∞ U (τn0 , Yn ) ∈ L1 (P ) and U (τ 0 , Yn ) ∈ L1 (P ) for each n ∈ IN . Then we claim that £ ¤ (2.16) lim E |U (τn0 , Yn ) − U (τ 0 , Yn )| = 0. n→∞ Similarly, if U (τ 0 , Yn ) ∈ L1 (P ) for each n ∈ IN ∪ {∞} and Yn → Y∞ P -a.s., we claim that £ ¤ (2.17) lim E |U (τ 0 , Yn ) − U (τ 0 , Y∞ )| = 0. n→∞ bτ 0 and Yn := gn for (2.16), and Yn := g̃n for (2.17); they Examples used in step 4) are Yn := X can all be assumed without loss of generality to satisfy the required integrability. Indeed, we ¢− ¡ ∈ L1 (P ) because we are maximising expected utility, and then may suppose U (τn0 , Yn ) ¢+ ¡ ∈ L1 (P ) since we assume in (2.13) finiteness of the maximal expected also U (τn0 , Yn ) utility. The same arguments apply to U (τ 0 , Yn ). Let us first argue (2.16). For ease of notation, we drop in the rest of this step all the 0 since the presence or absence of the minimum with σk does not change the argument in any ¢ ¡ way. We first fix m and note that U (τn , Yn )I{ m1 ≤Yn ≤m} = U τn , Yn I{ m1 ≤Yn ≤m} I{ m1 ≤Yn ≤m} (with the usual convention that ±∞ · 0 = 0). If |U | denotes the variation of U with respect ¯ ¯ 1 ) and therefore, using the to t, we thus have ¯U (τn , Yn )I{ m1 ≤Yn ≤m} ¯ ≤ |U |(τn , m) + |U |(τn , m assumption (2.11), ¯ ¯ sup ¯U (τn , Yn )I{ m1 ≤Yn ≤m} I{τn <%m } ¯ ∈ L1 (P ) (2.18) n∈IN for a localising sequence (%m ), constructed via a diagonal argument, which makes each 1 ) have integrable variation. Moreover, U (τn , Yn ) ∈ L1 (P ) implies that |U |(·, m) and |U |(·, m U (τn , 0)I{Yn =0} ∈ L1 (P ) so that (2.19) U (τn , Yn )I{Yn =0} = U (τn , 0)I{U (τn ,0)>−∞} I{Yn =0} P -a.s. ¢ ¡ Hence we also have from (2.11) possibly after further modifying (%m ) that ¯ ¯ (2.20) sup ¯U (τn , 0)I{U (τn ,0)>−∞} I{Yn =0} I{τn <%m } ¯ ∈ L1 (P ). n∈IN Clearly, (2.18) and (2.20) also hold with τn replaced by τ everywhere. Now introduce for m, n ∈ IN the double sequence of nonnegative random variables ¡ ¢ Zm,n := |U (τn , Yn ) − U (τ, Yn )| I{ m1 ≤Yn ≤m} + I{Yn =0} I{τn <%m } . Then for each m ∈ IN , (2.12) and (2.19) combined with (2.11) imply that lim Zm,n = 0 n→∞ 1 P -a.s., and since sup Zm,n ∈ L (P ) by (2.18) and (2.20), we get n∈IN for each m ∈ IN lim E[Zm,n ] = 0 n→∞ 18 by dominated convergence. On the other hand, for each n ∈ IN , the sequence (Zm,n )m∈IN increases to Z∞,n := |U (τn , Yn ) − U (τ, Yn )| P -a.s., and so monotone integration yields lim E[Zm,n ] = E[Z∞,n ] = E[|U (τn , Yn ) − U (τ, Yn )|] for each m ∈ IN . m→∞ But now we use the general fact that lim sup lim inf amn ≤ lim inf lim sup amn for any double m→∞ n→∞ m→∞ n→∞ sequence (amn ) in IR to obtain lim sup E[|U (τn , Yn ) − U (τ, Yn )|] = lim sup lim E[Zm,n ] ≤ lim inf lim E[Zm,n ] = 0. n→∞ m→∞ n→∞ m→∞ n→∞ This proves (2.16). Finally, (2.17) is proved completely analogously by working with the ¢ ¡ q.e.d. random variables Z̄m,n := |U (τ, Yn ) − U (τ, Y∞ )| I{ m1 ≤Yn ≤m} + I{Yn =0} I{τ <%m } . Remark 2.9. As can be seen from the above proof, the conclusion of Theorem 2.8 still holds if we replace (2.9) by the slightly weaker assumption that De,σ (S, P ) 6= ∅. In particular, by the fundamental theorem of asset pricing, the latter holds if S satisfies (NFLVR). ¦ b with nice trajectories. In order to apply The next logical step is to obtain a version of X b satisfies (NUPBR) for simple admissible integrands for Lemma 2.1, we want to argue that X b To that end, we first prove that simple admissible integrals of X b can (modulo suitable X. stopping) be realised via (general) admissible integrals of S. More precisely: Proposition 2.10. Suppose that the IRd -valued semimartingale S satisfies (2.9), and U satisfies (2.13). For any fixed T ∈ (0, ∞), there exists a sequence (Tm )m∈IN of stopping times increasing stationarily to T and having the following property: Let H be a simple real-valued I P hi I]]τi−1 ,τi ]] , where 0 ≤ τ0 ≤ τ1 ≤ · · · ≤ τI ≤ T are stopping integrand of the form H = i=1 times and hi is a bounded Fτi−1 -measurable random variable for i = 1, . . . , I. Fix x > 0 and b ≥ 0, then there exist x-admissible integrands take any stopping time % ≤ T . If x + H . X ϑ(m,%) , m ∈ IN , for S Tm such that for each m, (m,%) b%∧T = x + ϑ(m,%) . S%∧T = X x,ϑ x + H .X m m %∧Tm P -a.s. Proof. Take W := W (T ) from Lemma 2.5 and apply Proposition 2.7 to obtain stopping times (Tm )m∈IN increasing stationarily to T and probability measures Qm , m ∈ IN , equivalent to P such that for each m, the random variable W is in L2 (Qm ) and S Tm is in M2 (Qm ). 1) Suppose first that each hi is not a random variable, but a constant ci ∈ IR. Then we have for any stopping time τ ≤ T that bτ = H .X I X i=1 bτ ∧τ − X bτ ∧τ ) = ci ( X i i−1 I X i=1 ci Zτ 0 ¡ 19 ¢ ϑ̂(τi ∧ τ, u) − ϑ̂(τi−1 ∧ τ, u) dSu = ψ (τ ) . Sτ (τ ) with the predictable process ψu := I P i=1 ¡ ¢ ci ϑ̂(τi ∧τ, u)− ϑ̂(τi−1 ∧τ, u) . Since S Tm ∈ M2 (Qm ) := ψ (%∧Tm ) yields that the process ϑ(m,%) . S Tm = ϑ(m,%) . S and W ∈ L (Qm ), choosing ϑ is a martingale in M2 (Qm ), and by construction, 2 (m,%) b%∧T ≥ 0 x + ϑ(m,%) . S%∧Tm = x + H . X m P -a.s. Thus also x + ϑ(m,%) . S ≥ 0 by the martingale property, and so ϑ(m,%) is x-admissible for S Tm . 2) In general, if hi is an Fτi−1 -measurable random variable, we can no longer simply say ¢ ¡ that hi ϑ̂(τi ∧ τ, ·) − ϑ̂(τi−1 ∧ τ, ·) is predictable. Therefore we define (N ) hi (N ) Thus hi := −N I{hi <−N } + N I{hi ≥N } + NX 2N −1 k=−N 2N k2−N I{k2−N ≤hi <(k+1)2−N } . (N ) (N ) is a finite sum of terms of the form ck IA(N ) with sets Ak k (N ) (i) ∈ Fτi−1 and constants (i) ck . So if we define stopping times %k ≤ τi by %k := τi−1 IA(N ) + τi I(A(N ) )c , we obtain k (N ) (N ) ck IA(N ) I]]τi−1 ,τi ]] = ck I]]%(i) ,τi ]] , and so H (N ) := k k I P i=1 (N ) hi k I]]τi−1 ,τi ]] has the same form as in step 1) (although it now involves a double sum over i and k). Hence we get as in step 1) predictable integrands ϑ(N,m,%) for S Tm with b%∧T , x + ϑ(N,m,%) . S%∧Tm = x + H (N ) . X m (N ) m and each ϑ(N,m,%) . S T is in M2 (Qm ). But clearly each hi N → ∞, and therefore also b%∧T = H (N ) . X m I X i=1 (N ) ¡ hi converges to hi pointwise as ¢ bτ ∧%∧T − X bτ ∧%∧T −→ H . X b%∧T X i m i−1 m m P -a.s. as N → ∞. ¢ ¡ b%∧T Moreover, the sequence H (N ) . X m N ∈IN has the majorant ¯ ¯ ¡ ¢ bτ ∧%∧T + X b%∧T ¯ ≤ IkHk∞ max X bτ ∧%∧T ≤ 2IkHk∞ W (T ) ∈ L2 (Qm ), sup ¯H (N ) . X m i m i−1 m N ∈IN i=1,...,I b%∧T also converges in L2 (Qm ) as N → ∞. But the and thus ϑ(N,m,%) . S%∧Tm = H (N ) . X m construction of stochastic integrals with respect to martingales now implies that there exists an S Tm -integrable process ϑ(m,%) such that ϑ(m,%) . S Tm is in M2 (Qm ) and has final value b%∧T ≥ −x. The martingale property of ϑ(m,%) . S Tm therefore implies ϑ(m,%) . S%∧Tm = H . X m that ϑ(m,%) is x-admissible for S Tm , and so the proof is complete. 20 q.e.d. Proposition 2.11. Suppose that the IRd -valued semimartingale S satisfies (2.9), and U b satisfies (NUPBR) for simple x-admissible integrands. As a consesatisfies (2.13). Then X b admits P -a.s. right and left limits along the rationals. quence, X b fails (NUPBR) for simple Proof. We fix T ∈ [0, ∞] and argue by contradiction. If X b x-admissible integrands, we can find a sequence of simple predictable integrands H n for X b ≥ −x and (H n . X bT )n∈IN is not bounded in L0 . By Proposition 2.10, we such that H n . X bT = ϑ(m,n) . ST with ϑ(m,n) . S ≥ −x on [[0, Tm ]] since H n . X b ≥ −x. If we can write H n . X m m extend ϑ(m,n) to [[0, T ]] by setting it to 0 after Tm , we thus get ϑ(m,n) . S ≥ −x on [[0, T ]] so that (ϑ(m,n) )n∈IN is for each m a sequence of x-admissible integrands for S. But S satisfies (2.4) by Lemma 2.4, and so for each m, (2.21) Now ¢ ¡ the sequence ϑ(m,n) . ST n∈IN is bounded in L0 . P [ϑ(m,n) . ST ≥ η] ≥ P [ϑ(m,n) . ST ≥ η, T = Tm ] bT ≥ η, T = Tm ] = P [H n . X bT ≥ η] − P [Tm < T ], ≥ P [H n . X and since Tm % T , we have for any δ > 0 that P [Tm < T ] ≤ 2δ for m large enough. So if bT )n∈IN is not bounded in L0 , we have lim sup P [H n . X bT ≥ η] ≥ δ for some δ > 0 and (H n . X n→∞ all η > 0, and hence for m fixed large enough also lim sup P [ϑ(m,n) . ST ≥ η] ≥ n→∞ δ 2 for all η > 0, b indeed satisfies (NUPBR) for simple x-admissible which contradicts (2.21). Therefore X integrands, and the existence of right and left limits along the rationals then follows from Lemma 2.1. q.e.d. Remark 2.12. 1) By slightly modifying the proof of Proposition 2.11, we could also show b satisfies (NFLVR) that if S satisfies (NFLVR) and (2.9), and U still satisfies (2.13), then X for simple admissible integrands. This gives again existence of right and left limits along the rationals, by Lemma 2.1. 2) For the case where U does not depend on (ω, t), an alternative way to obtain the conclusion of Proposition 2.11 would be to invoke Proposition 7.2.2 of Ankirchner (2005). That result shows that an adapted RCLL process Y satisfies (NFLVR) for simple strategies if its expected utility remains bounded over all final wealths from simple strategies, and a closer look at the proof reveals that the assumption of RCLL trajectories is not needed. Since we b as general stochastic integrals of S via Proposition can rewrite simple stochastic integrals of X 2.10 while preserving admissibility, we could thus invoke (2.13) to apply Ankirchner’s result. However, our result allows more general U and our proof is also more direct. 21 ¦ Proposition 2.13. Suppose that the IRd -valued semimartingale S satisfies (2.9), and U bτ b admits a version X e which is RCLL, and for which X eτ = X satisfies (2.11)–(2.13). Then X e = H .X b for every simple P -a.s. for every bounded stopping time τ . As a consequence, H . X predictable process H. et := Proof. As in Lemma 2.2, set X lim u&t, u∈I Q bu for t ≥ 0; this is well defined by Proposition X bτ P -a.s. for eτ = X 2.11, and Theorem 2.8 implies like in (the proof of) Lemma 2.2 that X e is clearly a version of X. b Moreover, H . X e = H .X b each bounded stopping time τ . Hence X e satisfies for every simple predictable process H, and so Proposition 2.11 implies that also X e But by construction, X e is adapted, like (NUPBR) for simple x-admissible integrands for X. b and right-continuous, hence optional and so progressively measurable, and thus X e also X, admits left limits by the second part of Lemma 2.1. This completes the proof. q.e.d. e from Proposition 2.13 is a semimartingale. The basic Our goal is now to show that X idea for that is similar to the approach taken by S. Ankirchner and P. Imkeller in Ankirchner (2005) and Ankirchner/Imkeller (2005), and goes back to the classical ideas in Delbaen/ Schachermayer (1994). Ankirchner and Imkeller deduce the semimartingale property from the finiteness of the maximal expected utility from final wealth when this is taken over all simple strategies; but all that is really needed is Lemma 2.3 which says in essence that a locally bounded process Y is a semimartingale if the family of final wealths from bounded simple admissible strategies for Y is bounded in L0 . We then argue indirectly to show that S e admits a sequence of simple admissible strategies with final wealths unbounded fails (2.4) if X in L0 ; this can be achieved because Proposition 2.10 allows us to translate simple admissible e into admissible integrals over S. integrals over X e and to be able to use the general results in In order to deal with the jumps of X e Section 2.1 on locally bounded processes, we introduce the process V of large jumps of X P es I e ∆X by Vt := es |>1} for t ≥ 0. This is well defined, since X is RCLL, and clearly {|∆X 0<s≤t e is a semimartingale adapted, RCLL and of finite variation, hence a semimartingale. So X e − V is a semimartingale, and clearly X̄ is locally bounded since its if and only if X̄ := X jumps are bounded by 1, by construction. To prove that X̄ is a semimartingale, it is enough to prove that X̄ T is a semimartingale for each T ∈ (0, ∞). Now fix x > 0 and T ∈ (0, ∞) and construct the stopping times (Tm )m∈IN as in Proposition 2.7 with W = W (T ) taken from Lemma 2.5. Set and define now (2.22) ¯ © σm := inf t ≥ 0 ¯ |V |t ≥ m 2 et > or X %m := Tm ∧ σm ∧ T. 22 m 2 ª So (%m )m∈IN increases to T stationarily, and if each X̄ %m is a semimartingale, so is X̄ T . We are now ready to prove the main result of this section. Theorem 2.14. Suppose that the IRd -valued semimartingale S satisfies (2.9), and U satb of optimal wealths has a version X e which is a isfies (2.11)–(2.13). Then the process X semimartingale (and RCLL). e of X b in Proposition 2.13, and Proof. 1) We have already constructed an RCLL version X e is a semimartingale if and only if the locally bounded process we have seen above that X et − X̄t := X X 0<s≤t es I ∆X e {|∆X s |>1} e t − Vt , =X t≥0 is a semimartingale, or equivalently if and only if X̄ %m is a semimartingale for each m, where %m is constructed in (2.22). So fix m and suppose that X̄ %m is not a semimartingale. We want to show that this leads to a contradiction to (2.9). For brevity, write X := X̄ %m and note that all stochastic integrals of X end at time T at the latest. Since X is adapted and locally bounded but not a semimartingale, Lemma 2.3 gives a sequence of simple predictable processes |H n | ≤ 1 with H n . X ≥ −c for all n, for some c > 0, and (2.23) P [H n . X%m ≥ αn ] = P [H n . XT ≥ αn ] ≥ δ > 0 for all n with a sequence αn → ∞ and some δ > 0, because (H n . XT )n∈IN is not bounded in L0 . So we have a sequence of stochastic integrals of X whose final values are not bounded in L0 ; but e whose final values what we really want is a sequence of stochastic integrals of the process X are not bounded in L0 . This will be constructed next. e ≥ −x with X e0 = x and note from the construction of %m ≤ σm that 2) Recall that X m n e |V |t ≤ m 2 and Xt ≤ 2 for t < %m . Because H ≥ −1, we thus have for t < %m that et − X e0 ≥ −c−|V |t −x− X e0 ≥ −c− m −2x ≥ −c− 3m −4x. et = H n . Xt +H n . Vt + X (H n +1) . X 2 2 Using H n + 1 ≤ 2, we obtain at time %m that e% = (H n + 1) . X e% − + (H n + 1)X e% − (H%n + 1)X e% − (H n + 1) . X %m m m m m m ≥ (−c − m 2 − 2x) − 2x − 2 m 2 = −c − so that (2.24) e %m ≥ −c − (H n + 1) . X 23 3m 2 − 4x. 3m 2 − 4x e we note that To exploit (2.23) in terms of X, e −X e + 2x + |V | e +X e0 − H n . V ≤ (H n + 1) . X H n . X = (H n + 1) . X e − x ≥ −2x. Therefore e −X e0 = X because X e% ≥ αn − k] + P [2x + |V |% ≥ k], P [H n . X%m ≥ αn ] ≤ P [(H n + 1) . X m m for any fixed k, and if we choose k large enough, we can ensure that P [2x + |V |%m ≥ k] ≤ where δ > 0 comes from (2.23). Thus we get, for fixed m and k large but fixed, e% ≥ αn − k] ≥ P [(H n + 1) . X m (2.25) in view of (2.23). e n := Now keep m and k fixed and define H simple predictable processes. Then (2.24) gives e %m = e n .X H (2.26) e n is so that each H x 2 -admissible, 2(c + + 4x) for all n ∈ IN x n 2c+3m+8x (H (αn −k)x 2c+3m+8x , + 1) for n ∈ IN , which are e %m ≥ − x (H n + 1) . X 2 e %m , and hence x-admissible for X £ e% ≥ z] = P (H n + 1) . X e n .X e% ≥ P [H m m shows that for βn := (2.27) x 3m 2 δ 2 δ 2 z(2c+3m+8x) ¤ x we have e % ≥ βn ] ≥ e n .X P [H m δ 2 for all n ∈ IN from (2.25). Moreover, αn → ∞ implies that βn → ∞ since x, c, m, k are all fixed. In e n )n∈IN of simple predictable processes which summary, then, we have found a sequence (H e have final values which e %m and whose integrals with respect to X are all x-admissible for X are not bounded in L0 because they satisfy (2.27) with some δ > 0 and a sequence βn → ∞. bτ for every bounded stopping time τ eτ = H . X 3) Recall from Proposition 2.13 that H . X and every simple predictable process H. For every n ∈ IN , we thus obtain from Proposition 2.10 an x-admissible integrand ϑ(n,m,%m ) for S Tm such that e n .X e% = x + H b% = x + ϑ(n,m,%m ) . S% e n .X x+H m m m P -a.s.; ¢ ¡ note that %m ≤ Tm by (2.22). Therefore we have a sequence x + ϑ(n,m,%m ) . S%m n∈IN in X%xm which is not bounded. But this contradicts (2.4) and hence also (2.9) by Lemma 2.4, and so X̄ % m must be a semimartingale for each m. Thus the proof is complete. 24 q.e.d. Remark 2.15. There is a whole strand of literature with results of the form that, loosely speaking, a process Y is a semimartingale if the maximal expected “utility” from its stochastic integrals over simple integrands remains finite; see Chapter 7 of Ankirchner (2005), Ankirchner/Imkeller (2005), Larsen/Žitković (2008), or Biagini/Øksendal (2005) in a slightly different setting. However, these results typically assume that Y is locally bounded and that the increasing “utility” function U : [0, ∞) → IR satisfies lim U (y) = +∞. Since our utility y→∞ e is not locally bounded, U might be bounded and can depend on (ω, t) as well, and since X a translation to our setting is not straightforward. The reason why we can still obtain the e is that because X e comes from optimal investment, its simple semimartingale property for X integrals can be related (via Proposition 2.10) to integrals of S. In that sense, we exploit e has more structure than some abstract process Y . that X ¦ Our goal in Section 5 is to provide more details on the structure of the semimartingale e In preparation, we study in the next section the structure of the optimal strategy field ϑ̂ X. in more detail. 3. Properties of the optimal strategy field ¡ ¢ In this section, we study the properties of the two-parameter process ϑ̂(t, u) t≥u≥0 in more detail. Recall that ϑ̂(t, ·) is an optimal strategy for the utility maximisation problem with time horizon t. The properties of (a good version ϑe of) ϑ̂ will be crucial in the sequel when e of optimal wealths from we want to establish the precise structure for the semimartingale X Theorem 2.14. In a nutshell, our goal in this section is to prove that ϑ̂ admits a version ϑe such that e u) is of finite variation, and – as a function of the time horizon, t 7→ ϑ(t, e t), t ≥ 0, is predictable and S-integrable. – the diagonal ϑ̄t = ϑ(t, For technical reasons, however, a precise statement is a bit more involved. To the best of our knowledge, there exist no results in the literature that are even remotely close to the above precise structural description of horizon-dependence in optimal portfolio choice. In general terms, financial economists are of course well aware that decision making under uncertainty is strongly affected by the planning horizon. But the considerable literature on that topic is almost exclusively empirical and aims predominantly at a qualitative understanding of which factors have which kind of influence, etc.; see Siebenmorgen/Weber (2004) for a (randomly chosen) recent sample of that literature. The only work we have found giving any theoretical result is an unpublished preprint by A. Lazrak which presents a continuous-time version of an idea from (an early preprint version of) Gollier/Zeckhauser 25 (2002). In the very specific setting of Lazrak (1998), S is a multivariate geometric Brownian motion with a unique equivalent martingale measure, and y 7→ U (y) is a standard utility function on (0, ∞). The optimal strategy can then be computed in almost closed form, and the (from our perspective) main result in Lazrak (1998) more or less shows that this strategy is increasing with respect to the time horizon if and only if the risk tolerance function y 7→ −U 0 (y)/U 00 (y) of U is convex. But it is clear from the arguments that this result and its proof depend very strongly on the specific model in Lazrak (1998). On the quantitative or mathematical side, there is also a substantial literature on the asymptotic behaviour of optimal strategies as the time horizon T tends to ∞. This can be found under headings like risk-sensitive control/portfolio management or portfolio turn- pike theorems. In a different direction, several authors have tried to find conditions or approaches that eliminate horizon-dependence by a modified problem formulation; see for instance Henderson/Hobson (2007), Choulli/Stricker/Li (2007), Zariphopoulou/Žitković (2008) or the numerous papers by M. Musiela and T. Zariphopoulou on forward preference criteria. But neither the results nor the techniques in all these areas seem to help in obtaining more precise information on the structure of horizon-dependence for all time horizons T ∈ (0, ∞). We do believe that our study is the first to address and answer such questions in a general framework in a rigorous way. After these general remarks, let us now get started. Throughout this section, we assume that the IRd -valued semimartingale S satisfies (2.9), and U satisfies (2.11)– (2.13), so that we can use all the results from Section 2. In order to have good properties for several quantities appearing below, we need the following preliminary result. Recall that the set WT is defined in Lemma 2.5. Lemma 3.1. Assume that the IRd -valued semimartingale S satisfies (2.9). For every fixed T ∈ (0, ∞), there exist stopping times (Tm )m∈IN increasing stationarily to T and probability measures (Qm )m∈IN equivalent to P with the following properties for each m: (3.1) W (T ) = ess sup WT ∈ L2 (Qm ). e ∗ = sup X et ∈ L2 (Qm ). (3.2) X T 0≤t≤T (3.3) J (3.4) J (1) (2) ¯ o ¯ ∗ . . := ess sup j ST = sup |j St | ¯ j predictable with kjk∞ ≤ 1 ∈ L2 (Qm ). n := ess sup (Πn )n∈IN ½ 0≤t≤T sup P n∈IN τi ∈Πn eτ − X eτ ) (X i+1 i 2 ¾ ∈ L2 (Qm ), where the essential supremum is taken over all sequences (Πn )n∈IN of random partitions (by stopping times τi ) of [0, T ] whose mesh size |Πn | := sup{τi − τi−1 | τi−1 , τi ∈ Πn } tends to 0. (3.5) S Tm ∈ M2 (Qm ). e Tm = X e0 + M̃ m + Ãm with M̃ m ∈ M2 (Qm ) and Ãm predictable with uniformly (3.6) X 0 26 bounded variation |Ãm |. (3.7) [M̃ m , Ãm ] is a Qm -martingale. R e− dM̃ m is a Qm -martingale. (3.8) X R ∗ e− d|Ãm | + hM̃ m i + [Ãm ] is uniformly bounded. (3.9) B̃ m := 4 X Proof. By Lemma 2.5, the random variable W (T ) is finite P -a.s. So is J (1) because S e is a semimartingale by Theorem 2.14 so that for any is a semimartingale. Moreover, X sequence (Πn )n∈IN of random partitions of [0, T ] whose mesh size |Πn | tends to 0, the sequence ´ ³ P e τ )2 eτ − X e T in L0 ; see Theorem II.22 in Protter (2005). (X converges to [X] i+1 i τi ∈Πn n∈IN Thus also J (2) is finite P -a.s. as the essential supremum of a family which is bounded in L0 . We now apply Proposition 2.7 with the random variable W := W (T ) + J (1) + J (2) to ob0 tain (Tm ) and (Qm ) which already satisfy (3.1)–(3.4); note that (3.2) follows by observing that 0 0 e Tm e ∗ ≤ W (T ) . Moreover, S Tm ∈ M2 (Qm ) and X is a semimartingale which is Qm -special X T 0 e0 + M̃ + à under e Tm since its supremum is in L2 (Qm ). For its canonical decomposition X =X Qm , we can then obtain (3.6)–(3.9) as well as (3.5) by stopping; note that [M̃ , Ã] is a local Qm -martingale by Yoeurp’s lemma (see Dellacherie/Meyer (1982), Theorem VII.36), and |Ã| R ∗ e− d|Ã| + hM̃ i + [Ã] are predictable and RCLL, hence locally bounded. We and B̃ := 4 X 0 and the stopping times required for the above localisations, choose as Tm the minimum of Tm and the proof is complete. q.e.d. Now we introduce a suitably “localised version” of ϑ̂ for which we can subsequently establish precise results. Choose first an arbitrary T ∈ (0, ∞) and construct Tm and Qm as in Lemma 3.1. Then we fix m ∈ IN and define the process ϑ(m) by ϑ(m) (t, u) := ϑ̂(t ∧ Tm , u ∧ Tm ) for 0 ≤ t ≤ T , 0 ≤ u ≤ T . Note that ϑ(m) (t, u) = 0 for u ∧ Tm > t ∧ Tm since ϑ̂(t ∧ Tm , ·), being a strategy on [[0, t ∧ Tm ]], is only defined on that interval. Recall that Ω̄ := Ω × (0, ∞) and set F̄ := F ⊗ B(0, ∞). Since S is IRd -valued, we need a bit of notation in connection with our stochastic integrals. The main issue is that the bracket process ([S] or hSi) of S is matrix-valued. So we choose a bounded (real-valued) predictable increasing process B m null at 0 such that ® T i (S m ) , (S Tm )j ¿ B m for all i, j = 1, . . . , d (where these brackets are taken under Qm ); one ³P d ®´ example would be B m = tanh (S Tm )i . We introduce the matrix-valued predictable i=1 ®± T i m m m process σ by σij = d (S ) , (S Tm )j dB m and the finite measure µm := Qm ⊗ B m on F̄. Then we have the isometry (3.10) kϑ . S Tm kM2 (Qm ) = kϑ . STTm kL2 (Qm ) 27 ° 1° = °(ϑtr σ m ϑ) 2 °L2 (µ m) 1 2 = kϑtr σ m ϑkL1 (µm ) =: kϑkL2 (Qm ,S Tm ) . Moreover, we also have (3.11) ϑ.S Tm , ψ .S Tm ® = Z ϑtr σψ dB m . Finally, we note that by its definition, each σ m (u) is symmetric and nonnegative definite; hence σ m = (cm )tr cm for some matrix-valued process cm , and we can and do choose cm to be predictable as well. The properties (3.10) and (3.11) are precisely what we need to derive properties for ϑ(m) e Tm . We remark that for the case d = 1 where S is real-valued, we can directly from those of X take B m = hS Tm i so that then µm = Qm ⊗ hS Tm i and σ m ≡ 1. Lemma 3.2. Viewed as a process defined on Ω̄, the mapping t 7→ ϑ(m) (t, ·) is right-continuous in L2 (Qm , S Tm ). Equivalently, the mapping t 7→ cm ϑ(m) (t, ·) is right-continuous in L2 (µm ) and hence in (P ⊗ B m )-measure. Proof. Take t0 > t and write (3.12) et∧T = X bt0 ∧T − X bt∧T et0 ∧T − X X m m m m = t0Z ∧Tm ϑ̂(t0 ∧ Tm , u) dSu − 0 = ZT 0 = ZT 0 t∧T Z m 0 ϑ̂(t ∧ Tm , u) dSu ¡ ¢ ϑ̂(t0 ∧ Tm , u) − ϑ̂(t ∧ Tm , u) dSuTm ¡ ¢ ϑ̂(t0 ∧ Tm , u ∧ Tm ) − ϑ̂(t ∧ Tm , u ∧ Tm ) dSuTm by using that ϑ̂ vanishes whenever its second argument exceeds the first one. Since all the ϑ̂(τ, ·) are x-admissible, all stochastic integrals are local Qm -martingales due to (2.7), and they are even in M2 (Qm ) because W (T ) is a majorant for their suprema which is in L2 (Qm ) by (3.1) in Lemma 3.1. Hence we obtain (3.13) £ et∧T ) et0 ∧T − X EQ m ( X m m ¤ 2 = EQ m "µ ZT 0 ¡ ¢ ϑ(m) (t0 , u) − ϑ(m) (t, u) dSuTm °2 ° = °ϑ(m) (t0 , ·) − ϑ(m) (t, ·)° 2 L (Qm ,S Tm ) 28 . ¶2 # et∧T converges to 0 P -a.s. as t0 & t by right-continuity of X et0 ∧T − X e (or in L0 by But X m m ¡ (T ) ¢2 , so that Theorem 2.8), and a Qm -integrable majorant for the squares is given by W ¤ £ et∧T )2 = 0 by dominated convergence. This gives the assertion thanks et0 ∧T − X lim EQ (X t0 &t m m m to the isometry (3.10). q.e.d. By pushing the same idea a little further, we next establish a measurability result for ϑ̂, or rather ϑ(m) . Recall that T ∈ (0, ∞) is fixed but arbitrary. Lemma 3.3. Viewed as a mapping on [0, T ] × Ω̄, ϑ(m) has a version which is (B[0, T ] ⊗ P)¯ measurable, where “version” is with respect to the measure λ¯[0,T ] ⊗ µm on [0, T ] × Ω̄. Proof. Denote by Dk the k-th dyadic partition of [0, T ] and by d0 := d + T 2−k the successor P in Dk of d ∈ Dk . Define then g k (t, ·) := I(d,d0 ] (t)ϑ(m) (d0 , ·) and note that each g k is d∈Dk clearly (B[0, T ] ⊗ P)-measurable because each ϑ(m) (d0 , ·) is a predictable process. Moreover, for each fixed t, we have cm g k (t, ·) → cm ϑ(m) (t, ·) in L2 (µm ) as k → ∞ by Lemma 3.2; but we can do even better. Indeed, write first ¯ ¡ ¢¯2 (3.14) ¯cm (u) g k (t, u) − ϑ(m) (t, u) ¯ ¡ ¢tr ¢ ¡ = g k (t, u) − ϑ(m) (t, u) σ m (u) g k (t, u) − ϑ(m) (t, u) µ X ¶ ¡ (m) 0 ¢ tr m (m) I(d,d0 ] (t) ϑ (d , u) − ϑ (t, u) = σ (u) d∈Dk × = µ X I(d,d0 ] (t) ϑ d∈Dk X ¡ I(d,d0 ] (t) d∈Dk ³¡ ϑ (m) (m) 0 (d , u) − ϑ 0 (d , u) − ϑ (m) (m) (t, u) (t, u) ¢ ¢tr ¶ ¡ m σ (u) ϑ (m) 0 (d , u) − ϑ (m) (t, u) ¢´ ; then integrate with respect to B m in u, take expectations under Qm in ω, and finally integrate with respect to λ in t to obtain, by using (3.10) and (3.13), ZT ° ¢¯ ° ° ¯¯ m ¡ k °2 (m) ° c g (t, ·) − ϑ (t, ·) ¯ ° 2 L (µm ) dt 0 = ZT 0 = ° k ° °g (t, ·) − ϑ(m) (t, ·)°2 2 ZT X 0 d∈Dk L (Qm ,S Tm ) dt ° °2 I(d,d0 ] (t)°ϑ(m) (d0 , ·) − ϑ(m) (t, ·)°L2 (Qm ,S Tm ) dt 29 = ZT X 0 d∈Dk = ZT 0 = ZT 0 if we set Ytk := P d∈Dk EQ m £ ¤ ed0 ∧T − X et∧T )2 dt I(d,d0 ] (t)EQm (X m m "µ X d∈Dk ed0 ∧T − X et∧T ) I(d,d0 ] (t)(X m m ¶2 # dt £ ¤ etTm )2 dt, EQm (Ytk − X e T0m for 0 ≤ t ≤ T . For every fixed t, Ytk converges to X etTm I(d,d0 ] (t)X d e (or in L0 by Theorem 2.8), and P -a.s. as k → ∞ by the right-continuity of X etTm | ≤ W (T ) ∈ L2 (Qm ). sup sup |Ytk | ≤ sup |X k∈IN 0≤t≤T 0≤t≤T £ ¤ etTm )2 tends to 0 as Hence dominated convergence first yields that fk (t) := EQm (Ytk − X £ ¤ k → ∞ for each t; but since we also have fk (t) ≤ 4E (W (T ) )2 for all t and k, again using dominated convergence gives lim RT k→∞ 0 fk (t) dt = 0. Translated back, this means that |cm g k − cm ϑ(m) | −→ 0 ¡ ¯ ¢ in L2 λ¯[0,T ] ⊗ µm as k → ∞, so that cm ϑ(m) agrees almost everywhere on [0, T ] × Ω̄ with lim sup cm g k =: hm , say, which k→∞ m k is (B[0, T ] ⊗ P)-measurable since c and each g are. Finally, we write hm = cm ψIC + 0IC c for a set C ∈ B[0, T ] ⊗ P and a (B[0, T ] ⊗ P)-measurable process ψ, by distinguishing in a ¢ ¢ ¡ ¡ (B[0, T ]⊗P)-measurable way whether hm (t, u) is in Range cm (u) or in Ker cm (u) . Setting then ϑ(m) := ψ gives the desired version. q.e.d. Remark 3.4. Although, as pointed out in Section 2 after (2.13), we have no uniqueness for the optimal strategy ϑ̂(τ, ·) for a given time horizon τ , Lemma 3.3 nevertheless allows us to choose a good version for ϑ(m) . Once we have done that, the nonuniqueness issue is laid to ¦ rest from here on. In the sequel, we choose and fix a version of ϑ(m) as in Lemma 3.3, without changing the notation. For future use, we note that this has no effect on integrals formed of simple integrands in the t-variable with respect to ϑ(m) , since the resulting integrals are then also versions of each other. ¡ ¢ Lemma 3.5. Suppose that the process Y (t, u) t≥0, u≥0 is (B[0, ∞) ⊗ P)-measurable when viewed as a mapping on [0, ∞) × Ω̄. Then the diagonal process Ȳt := Y (t, t), t ≥ 0, is predictable. (An analogous result holds true if P is replaced by the optional σ-field O.) 30 Proof. If Y (t, u) = f (t)gu where g is a predictable process (on Ω) and f is a Borel-measurable ¡ ¢ function on [0, ∞) , then Ȳt = f (t)gt is predictable as the product of two predictable processes, since any nonrandom measurable function is trivially predictable. For general Y , the assertion follows via the monotone class theorem. q.e.d. Now we are ready to establish the first (and simpler) main result of this section. Apart from the localisation issue, this says that taking the diagonal of the optimal strategy field yields a predictable S-integrable process. Corollary 3.6. For the version of ϑ(m) obtained from Lemma 3.3, the diagonal process (m) ϑ̄u := ϑ(m) (u, u) = ϑ̂(u ∧ Tm , u ∧ Tm ), 0 ≤ u ≤ T , is predictable and S Tm -integrable. Proof. Predictability is immediate from Lemma 3.5 and Lemma 3.3. Now we go back to (3.12) in the proof of Lemma 3.2 and write this as et ∧T = et ∧T − X X i+1 m i m Zti 0 ¡ ¢ ϑ(m) (ti+1 , u) − ϑ(m) (ti , u) dSuTm + tZi+1 ϑ(m) (ti+1 , u) dSuTm . ti As already pointed out in the proof of Lemma 3.2, all the stochastic integral processes are in M2 (Qm ), and so the two summands on the right-hand side above are orthogonal in L2 (Qm ). P I(d,d0 ] (u)ϑ(m) (d0 , u), where Dk is again the k-th dyadic partition With the notation ψuk := d∈Dk of [0, T ], we thus obtain from (3.10) similarly as in (3.14) (3.15) X d∈Dk £ ed0 ∧T − X ed∧T ) EQ m ( X m m ¤ 2 ≥ X EQ m d∈Dk = EQm "µ Zd0 ϑ(m) (d0 , u) dSuTm d "µ 0 d X Z ¡ ϑ(m) (d0 , u) dSuTm d∈Dk d ¤ = EQm (ψ k . STTm )2 ° ° = °(ψ k )tr σ m ψ k °L1 (µ £ e is a semimartingale by Theorem 2.14, But since X X ed∧T )2 −→ [X ed0 ∧T − X e Tm ] T (X m m d∈Dk ¢ ¶2 # m) ¶2 # . in L0 as k → ∞, and due to (3.4) in Lemma 3.1, dominated convergence therefore implies that the left-hand £ T ¤ e m ]T < ∞. But for the right-hand side, side of (3.15) converges to EQm [X cm (u)ψuk −→ cm (u)ϑ(m) (u, u) = cm (u)ϑ̄(m) u 31 in µm -measure as k → ∞ by Lemma 3.2, and since σ m = (cm )tr cm is nonnegative definite, Fatou’s lemma yields ° (m) tr m (m) ° °(ϑ̄ ) σ ϑ̄ ° L1 (µm ) ° ° ≤ lim inf °(ψ k )tr σ m ψ k °L1 (µ ) m k→∞ ¸ · X 2 ed∧T ) ed0 ∧T − X ≤ lim EQm (X m m k→∞ d∈Dk ¤ e Tm ]T < ∞. = EQ m [ X £ So ϑ̄(m) is even in L2 (Qm , S Tm ) and hence S Tm -integrable. q.e.d. Having finished the first piece of business in this section, we now turn to showing that ϑ̂ – or rather the chosen good version of ϑ(m) – is of finite variation with respect to its first argument t. The overall idea is to view ϑ(m) as a process indexed by t and defined on Ω̄ = Ω × (0, ∞), and to prove that it is a semimartingale over a constant filtration. To make this more precise, recall that F̄ = F ⊗ B(0, ∞) and define on (Ω̄, F̄ ) the filtration Pt := P, t ≥ 0, where P is the predictable σ-field on Ω̄. (Even more precisely, we take the augmentation with respect to µm = Qm ⊗ B m to have the usual conditions satisfied.) Then ϑ(m) is (Pt )-adapted since each ϑ(m) (t, ·) is a predictable process. To use the vector process ϑ(m) as a stochastic integrator with respect to the t-variable, we start with very simple predictable matrix-valued processes f¯ on Ω̄. Such an f¯ has the I P Cui (ω)I(ti ,ti+1 ] (t) with each C i being a Pti -measurable d × d-matrix on form f¯(t, ω, u) = i=1 i Ω̄ so that each C is a predictable matrix-valued process on Ω. The stochastic integral of f¯ with respect to ϑ(m) is then the vector process defined by (m) f¯. ϑT := ZT 0 f¯t ϑ(m) (dt, ·) := I X i ¡ C ϑ i=1 (m) (ti+1 , ·) − ϑ (m) ¢ (ti , ·) =: I X C i ∆i+1 ϑ(m) , i=1 and this is an IRd -valued predictable process (indexed by u) on Ω. We want to use the classical Bichteler–Dellacherie characterisation of semimartingales as good stochastic integrators, and (m) so we first study the continuity properties of the mapping f¯ 7→ f¯. ϑ . T Proposition 3.7. Take the version of ϑ(m) from Lemma 3.3. Then: (m),k 1) For each coordinate k ∈ {1, . . . , d}, the mapping f 7→ f . ϑT is continuous in ¡ £ T k ¤¢ 1 on the set of bounded very simple real-valued predictable processes f on L Qm ⊗ (S m ) ¯ ¤¢ ¡ £ ⊗ Qm ⊗ (S Tm )k . Ω̄, equipped with the L∞ -norm with respect to the measure λ¯ [0,T ] 2) For each simple (i.e. piecewise constant) real-valued measurable function f on [0, T ], (m) the IRd -valued process f . ϑT is in L2 (Qm , S Tm ), and we have (3.16) ° ° °f . ϑ(m) ° T L2 (Qm ,S Tm ) 32 ≤ Cm kf kL∞ b The notation f . ϑ(m) here should be read with a constant Cm depending only on m and X. T (m) (m) componentwise; so f . ϑT := (f Id×d ) . ϑT . Note that although part 2) of Proposition 3.7 will only be used in Section 4, we provide this result for efficiency already here, since the arguments for 1) and 2) are largely parallel. Before we can prove Proposition 3.7, we need an auxiliary result. Recall from (the proof e0 + M̃ m + Ãm under Qm . e Tm = X of) Lemma 3.1 the canonical decomposition X Lemma 3.8. With the notation ∆i+1 ϑ(m) := ϑ(m) (ti+1 , ·) − ϑ(m) (ti , ·), we have (3.17) EQ m " ZT 0 (3.18) ¯ # ¯ ¯ £ Tm ¤ (m) Tm ¯ etTm ¯ Ft ∧T et − X ∆i+1 ϑ (u) dSu ¯ Fti ∧Tm = EQm X i m i+1 i ¯ ¯ £ ¤ m¯ = EQm Ãm ti+1 − Ãti Fti ∧Tm and (3.19) EQ m "µ ZT 0 (3.20) # ¶2 ¯¯ ¯ £ Tm ¤ ¯ etTm )2 ¯ Ft ∧T et − X ∆i+1 ϑ(m) (u) dSuTm ¯ Fti ∧Tm = EQm (X i m i+1 i ¯ ¯ £ ¤ ≤ EQm B̃tmi+1 − B̃tmi ¯ Fti ∧Tm for an increasing predictable uniformly bounded process B̃ m . Proof. Purely for ease of notation, we drop in this proof all indices m and argue for Tm = T . Moreover, we also shortly write S instead of S Tm . Define the process Y i+1 := ϑ(ti+1 , ·) . S so that Y i+1 ∈ M2 (Q) by Proposition 2.7, and note that et − x =X YTi+1 = Yti+1 i+1 i+1 because ϑ(ti+1 , ·) vanishes after ti+1 . So ZT 0 et − X et , ∆i+1 ϑ(u) dSu = YTi+1 − YTi = X i+1 i and this yields (3.17) and (3.19). Due to (3.6) in Lemma 3.1, (3.18) then follows as well. e Tm = x + M̃ m + Ãm to obtain To prove (3.20), we first use Itô’s formula on X e2 − X et − X et − X e t (X e t )2 = X et ) = 2 e 2 − 2X (X ti+1 ti i+1 i i i+1 i 33 tZi+1 ti e t ) dX e t − [X] es + [X] es− − X et. (X i i+1 i e = [M̃ ] + [Ã] + 2[M̃ , Ã], and by (3.6)–(3.8) and (3.2) in Lemma 3.1, the processes Now [X] R et ) dM̃ on [ti , ti+1 ] are all Q-martingales. With the usual e− − X [M̃ ] − hM̃ i, [M̃ , Ã], and (X i ∗ es |, we thus get e := sup |X supremum notation X t 0≤s≤t ¯ ¤ £ e t )2 ¯ Ft et − X EQ ( X i+1 i i ¯ # " tZi+1 ¯ es− − X et ) dÃs + hM̃ it − hM̃ it + [Ã]t − [Ã]t ¯¯ Ft (X = EQ 2 i i+1 i i+1 i ¯ i ti ≤ EQ [B̃ti+1 − B̃ti | Fti ] R ∗ e d|Ã| + hM̃ i + [Ã]. By (3.9) in Lemma 3.1, B̃ has the claimed with the process B̃ := 4 X − properties. q.e.d. Proof of Proposition 3.7. Like in the proof of Lemma 3.8, we drop all indices m and work with Tm = T for ease of notation. P i a) Suppose f¯ = C I(t ,t ] is a very simple bounded IRd×d -valued predictable process i i i+1 P i C ∆i+1 ϑ. In addition, let the matrices C i all be diagonal so that on Ω̄ so that f¯. ϑT = i i i σC = C σ. Take any IRd -valued predictable process h in L2 (Q, S) and write ZT (3.21) 0 htr u σu ¡ XZ ti ¢ f¯. ϑT (u) dBu = i i htr u Cu σu ∆i+1 ϑ(u) dBu 0 ti+1 X Z i + htr u Cu σu ∆i+1 ϑ(u) dBu , i ti using that ∆i+1 ϑ vanishes after time ti+1 . We first look at the second sum where we fix i. There we use first Cauchy–Schwarz (for the scalar product in IRd with matrix σ) pointwise for the dB-integrand, then Cauchy– Schwarz for the dB-integrals, then conditional Cauchy–Schwarz in L2 (Q) and then the (conditional) isometry property (3.10) of the stochastic integral to obtain ¯ # " tZi+1 ¯ ¯ tr i ¯ ¯ ¯hu Cu σu ∆i+1 ϑ(u)¯ dBu ¯ Ft EQ ¯ i ti ≤ EQ ≤ à " tZi+1 ¡ ti i i tr htr u Cu σu (Cu ) hu ¢ 21 ¡ ∆i+1 ϑtr (u)σu ∆i+1 ϑ(u) ¢ 12 ¯ # ¯ ¯ dBu ¯ Fti ¯ ¯ #! 21 à "µ ZT #! 12 " tZi+1 ¶2 ¯¯ ¯ ¯ ¯ ∆i+1 ϑ(u) dSu ¯ Fti (Cui hu )tr σu (Cui hu ) dBu ¯ Fti EQ . EQ ¯ ¯ ti 0 34 Note that (C i )tr = C i since C i is diagonal and that we have generously estimated the integral ti+1 R ∆i+1 ϑtr (u)σu ∆i+1 ϑ(u) dBu by integrating the same quantity from 0 to T , before using ti (3.10). Next we sum over i, use Cauchy–Schwarz for the sum, take expectations under Q and apply Cauchy–Schwarz for the expectation. Then we use |C i | ≤ kf¯k∞ and the isometry (3.10) for the resulting first factor, and (3.20) from Lemma 3.8 for the second factor, to get (3.22) EQ " # ti+1 X Z ¯ ¯ i ¯htr ¯ u Cu σu ∆i+1 ϑ(u) dBu i ti à ≤ kf¯k∞ EQ " #! 21 #! 12 à " ti+1 X Z X EQ htr (B̃ti+1 − B̃ti ) u σu hu dBu i i ti 1 = kf¯k∞ khkL2 (Q,S) kB̃T kL2 1 (Q) . Now we look at the first sum in (3.21), or more precisely at P Rti ¯¯ i 0 which majorises that first sum. For each i, the process ¯ i ¯ dBu , htr C σ ∆ ϑ(u) u i+1 u u ¢ ¡ h̃i := h sign htr C i σ∆i+1 ϑ (3.23) is IRd -valued, predictable, in L2 (Q, S) like h, with even |h̃i | = |h|, and it satisfies by con¯ ¯ struction ¯htr C i σ∆i+1 ϑ¯ = (h̃i )tr C i σ∆i+1 ϑ. Then we set hi := h̃i I[[0,ti ]] so that clearly (C i hi ) . ST = (C i h̃i ) . Sti is Fti -measurable. Moreover, (C i hi ) . S ∈ M2 (Q) since hi ∈ L2 (Q, S) and C i is bounded, and so by (3.11), EQ " Zti 0 # " ZT # ¯ tr i ¯ ¯hu Cu σu ∆i+1 ϑ(u)¯ dBu = EQ (hiu )tr Cui σu ∆i+1 ϑ(u) dBu 0 ® ¤ = EQ (C i hi ) . S, (∆i+1 ϑ) . S T £¡ ¢¡ ¢¤ = EQ (C i hi ) . ST (∆i+1 ϑ) . ST . £ But the first factor in the last expectation is Fti -measurable; so we can replace the second one by its Fti -conditional expectation, which is given by (3.18) in Lemma 3.8, and we find (3.24) EQ " # # " ti XZ ¯ X¡ ¯ ¢ i ¯htr ¯ (C i hi ) . Sti (Ãti+1 − Ãti ) u Cu σu ∆i+1 ϑ(u) dBu = EQ i i 0 ≤ EQ 35 h³ ¡ max (C h ) . ST∗ i i i ¢´ i |Ã|T . In summary so far, adding (3.24) and (3.22) yields in view of (3.21) that (3.25) EQ " ZT 0 # 1 ¯ tr ¡ ¢¯ ¯hu σu f¯. ϑT (u) ¯ dBu ≤ kf¯k∞ khkL2 (Q,S) kB̃T k 2 1 L (Q) ° ° 12 ° ¡ ¢° . + ° max (C i hi ) . ST∗ °L2 (Q) °|Ã|T °∞ i If we simply use h̃i ≡ ±h instead of (3.23), then we still get with the same arguments ¯ " ZT #¯ ¯ ¯ 1 ¡ ¢ ¯ ¯ 2 ¯ ¯ . f ϑ htr σ (u) dB ¯ EQ T u ¯ ≤ kf k∞ khkL2 (Q,S) kB̃T kL1 (Q) u u ¯ ¯ (3.26) 0 ° ° 12 ° ¡ ¢° . + ° max (C i hi ) . ST∗ °L2 (Q) °|Ã|T °∞ i At this point, we distinguish two cases: a1) If all the C i are deterministic constant matrices ci , we can pull them out of the stochastic integrals and estimate them above by kf¯k∞ . If in addition all the h̃i coincide with ±h, then ¡ ¢ max (C i hi ) . ST∗ ≤ kf¯k∞ (h . ST∗ ) (3.27) i because hi . S = h̃i . S ti = ±h . S ti . a2) If all the |h̃i | are bounded by 1, then |C i hi | ≤ kf¯k∞ and so ¡ ¢ (3.28) max (C i hi ) . ST∗ ≤ kf¯k∞ ess sup{j . ST∗ | j predictable and kjk∞ ≤ 1} = kf¯k∞ J (1) . i In both cases, we subsequently exploit that both |Ã|T and B̃T are bounded, due to (3.6) in Lemma 3.1 and Lemma 3.8, respectively. In case a2), we know from (3.3) in Lemma 3.1 that J (1) is in L2 (Q). In case a1), Doob’s inequality and the isometry property of the stochastic integral yield kh . ST∗ kL2 (Q) ≤ 2kh . ST kL2 (Q) = 2khkL2 (Q,S) . (3.29) b) For the proof of part 1), take now an index k ∈ {1, . . . , d} and a very simple bounded real-valued predictable process f on Ω̄. Define f¯j` := f δjk δ`k so that the only nonzero entry of the matrix f¯ is f¯kk = f . Then f¯. ϑT = (f . ϑk )ek , where ek is the k-th basis vector in T IR , and hence h := sign(f¯. ϑT ) = sign(f . ϑkT )ek , where the first sign function is applied d coordinatewise. With the same h̃i as in step a), we are then in case a2) from above, and so combining (3.25) with (3.28) yields, for the above choice of h and f¯, (3.30) ° ° °f . ϑkT ° L1 (Q⊗[S k ]) = EQ " ZT 0 ¯ ¯ ¯f . ϑkT (u)¯ dhS k iu 36 # = EQ " ZT 0 = EQ " ZT ¯ ¯ ¯f . ϑkT (u)¯σukk dBu ¯. htr u σu f ϑT (u) dBu 0 ≤ kf k∞ ³ ¢ ¡ # 1 2 kSTk kL2 (Q) kB̃T k∞ # + kJ (1) ° ° 12 ´ ° kL2 (Q) |Ã|T °∞ , using that kf¯k∞ = kf k∞ and |h| = 1. This proves assertion 1). P ci I(ti ,ti+1 ] with constants ci is a (bounded) c) To prove part 2), suppose that f = i real-valued piecewise constant function on [0, T ]. Take f¯ := f Id×d and any predictable (IRd -valued) h ∈ L2 (Q, S). Then we get from the definition of f¯, (3.26), (3.27) and (3.29) ¯ " ZT #¯ ¯ ¯ ¯¯ ¯ ¢ ¡ ¯. ϑT (u) dBu ¯¯ ¯(h, f . ϑT )L2 (Q,S) ¯ = ¯EQ htr σ f u u ¯ ¯ 0 ³ 1 ° 12 ´ ° 2 . ≤ kf k∞ khkL2 (Q,S) kB̃T k∞ + 2°|Ã|T °∞ Thus h 7→ (h, f . ϑT )L2 (Q,S) is a continuous linear functional on L2 (Q, S), and this implies by Riesz’ representation theorem that f . ϑT ∈ L2 (Q, S) with ³ 1 ° 21 ´ ° 2 ° . kf ϑT kL2 (Q,S) ≤ kf k∞ kB̃T k∞ + 2 |Ã|T °∞ . This proves assertion 2) and completes the proof. q.e.d. Remark 3.9. 1) Once we establish that each coordinate ϑ(m),k of ϑ(m) is of finite variation (m) with respect to t, we can of course define (the vector process) f . ϑT for every bounded (real-valued) measurable function f on [0, T ], and the estimate (3.16) in Proposition 3.7 then extends to these f as well. 2) In the above proof, we have worked with the sharp bracket process hS Tm i of S Tm , which exists under Qm since S Tm ∈ M2 (Qm ). The reason for not working with [S Tm ] (which always exists, even under P ) is that in the construction of h̃i in step a) of the proof, we have to ensure that h̃i is predictable. For the one-dimensional case d = 1, one could define h̃i directly from h; but for d > 1, the bracket process of S Tm also comes up via the matrix σ m , and so we have to work with hS Tm i because [S Tm ], as well as its corresponding matrix, is only optional in general. A similar issue comes up in the construction of the version ϑ(m) at the end of the proof of Lemma 3.3. ¦ Proposition 3.10. Take the version of ϑ(m) from Lemma 3.3. Viewed as a process on Ω̄, each coordinate t 7→ ϑ(m),k (t, ·) is then a (real-valued) quasimartingale with respect to (Pt ) 37 ¤ ¡ ¤¢ £ £ and Qm ⊗ (S Tm )k . Thus it has a version which is Qm ⊗ (S Tm )k -almost everywhere right-continuous in t. Proof. To simplify notation, we once again drop all indices m and work with Tm = T , writing also for brevity S instead of S Tm . Fix an index k ∈ {1, . . . , d} throughout. 1) To show that ϑk is a quasimartingale, we need to prove that it is adapted, integrable, and has the quasimartingale property (see below). Adaptedness is clear since Pt ≡ P and each ϑk (t, ·) is by definition a predictable process on Ω, and integrability will be argued presently. The main point is to show that ·X ¸ ¯ ¯ k ¯EQ⊗[S k ] [∆i+1 ϑ | Pt ]¯ < ∞, sup EQ⊗[S k ] i (3.31) Π ti ∈Π where ∆i+1 ϑk = ϑk (ti+1 , ·) − ϑk (ti , ·) as before and the supremum is taken over all partitions Π of [0, T ]. But Pti ≡ P; therefore the (inner) conditional expectation vanishes, and only the (outer) expectation remains. Moreover, for any given partition Π, setting P sign(∆i+1 ϑk )I(ti ,ti+1 ] gives a very simple real-valued predictable process on Ω̄ which f := ti ∈Π is bounded by 1. Because f . ϑkT = P sign(∆i+1 ϑk )∆i+1 ϑk = ti ∈Π ti ∈Π (3.30) from (the proof of) Proposition 3.7 gives ·X ¸ ¯ ¯ ° ° k¯ ¯ ∆i+1 ϑ = °f . ϑkT ° EQ⊗[S k ] ti ∈Π ≤ kf k∞ ³ P |∆i+1 ϑk | ≥ 0, the estimate L1 (Q⊗[S k ]) ° °1 2 kSTk kL2 (Q) °B̃T °∞ + kJ (1) ° ° 21 ´ ° kL2 (Q) |Ã|T °∞ . But kf kL∞ = 1, and so the right-hand side above does not depend on Π, which proves (3.31) and etablishes the quasimartingale property of ϑk in t. Finally, integrability follows ¢ ¡ in the same way, by choosing f := sign ϑk (t, ·) I(0,t] to obtain that f . ϑkT = |ϑk (t, ·)| is in L1 (Q ⊗ [S k ]). 2) Exactly as in the proof of Rao’s theorem (see Dellacherie/Meyer (1982), Theorem VI.40), the quasimartingale ϑk can be decomposed as the difference of two supermartingales ψ ± . Both of these have (Q⊗[S k ])-a.e. right limits (in the argument t) along the rationals, and because Pt ≡ P is obviously right-continuous, the processes ϕ± obtained as the right limits along the rationals are again supermartingales (see Dellacherie/Meyer (1982), Theorem VI.2) and of course right-continuous in t by construction. But now the argument goes as in Section 2 for Proposition 2.13: The process ϑk = ψ + − ψ − admits right limits along the rationals since ψ ± do, and the process obtained as those limits equals ϕ+ − ϕ− and is right-continuous. But we also know from Lemma 3.3 that ϑk is right-continuous in measure (with respect to µm , hence with respect to Q ⊗ [S k ] ¿ µm as well), and so ϕ+ − ϕ− is also a version of ϑk by (the proof of) Lemma 2.2. This completes the argument. q.e.d. 38 Putting everything together now readily yields the main result of this section. Theorem 3.11. As in Theorem 2.14, suppose that the IRd -valued semimartingale S satisfies (2.9), and U satisfies (2.11)–(2.13). Viewed as a process on Ω̄ = Ω × (0, ∞), each coordinate ¤¢ £ ¡ t 7→ ϑ(m),k (t, ·) has a version which is Qm ⊗ (S Tm )k -almost everywhere right-continuous and of finite variation in t. Proof. Take the right-continuous version from Proposition 3.10. This does not change the integrals, with respect to the vector-valued ϑ(m) , of very simple matrix-valued predictable processes on Ω̄, so that the conclusions of Proposition 3.7 hold for this version as well. Thus the Bichteler–Dellacherie characterisation of semimartingales (see Dellacherie/Meyer (1982), Theorem VIII.80) implies for each k that this version t 7→ ϑ(m),k (t, ·) is on Ω̄ a semimartingale; note that this uses right-continuity of ϑ(m),k in t, which is why we had to establish Proposition 3.10 beforehand. But the filtration Pt ≡ P on Ω̄ is constant; hence ϑ(m),k has no nontrivial martingale part and is therefore of finite variation (in the t-argument). q.e.d. Remark 3.12. Because ϑ(m) (t, u) = ϑ̂(t ∧ Tm , u ∧ Tm ) is not simply “the process ϑ̂ stopped at Tm ”, it is not clear whether we can easily (or at all) formulate an analogous result for the process ϑ̂ itself. The problem is that while Tm acts on the u-argument in the usual “stopping way”, its effect on the t-argument is less clear — recall that this first argument is the time ¦ horizon of a utility maximisation problem. If we forget the financial origin of the problem studied in this paper, we can view Theorems 2.14 and 3.11 as providing an answer to the following general question. Suppose ¡ ¢ Y = (Yt )t≥0 is a semimartingale and ψ(t, u) t≥0, 0≤u≤t is a two-parameter process such that ¡ ¢ for each fixed t, the process ψ(t, u) 0≤u≤t is predictable and Y -integrable on [0, t]. For each t ≥ 0, the random variable Zt = ψ(t, ·) . Yt = Zt ψ(t, u) dYu 0 is hence well defined. In that abstract setting, when is Z = (Zt )t≥0 a semimartingale? More precisely, there are two questions: 1) Which conditions on ψ and Y imply that Z is a semimartingale? 2) Conversely, if Z is a semimartingale, what does this imply about ψ? Even more generally, one can ask about the properties of a process Z that arises in such a way from a pair (ψ, Y ). At this level of generality, the question is too vague. If Y is Brownian motion and ψ is a nonrandom function, then Z is a Volterra-type integral process and there are many Gaussian-type results; see for instance Chapter VI in Hida/Hitsuda (1993). If in 39 R addition ψ(t, u) = f (t − u) for some deterministic function f , then Z = f (· − u) dWu is a convolution-type integral or moving average of Brownian motion W . In that specific case, questions 1) and 2) have very precise answers; we refer to the recent work of Basse/Pedersen (2009) which generalises these classical results from Brownian motion to the case where Y is a Lévy process, and which also contains a very good overview of earlier and recent literature. But for a general semimartingale Y and a general (stochastic) ψ, results seem to be much more sparse. For the easier question 1), there is an affirmative result in Protter (1985) under a rather restrictive assumption on ψ; see the end of Section 4 for a more detailed discussion. Concerning question 2) at the general semimartingale level, we are not aware of any results in the literature so far. If one looks more carefully at the arguments that lead to Theorem 3.11, one can see that our approach actually provides a partial answer to question 2) in the following way. All we really need for our arguments is that locally, after an equivalent change of measure, Y is a square-integrable martingale, all stochastic integral processes ψ(t, ·) . Y can be dominated in L2 , and the semimartingale Z is in S 2 . We could formulate this as an abstract theorem and it seems very likely that these ideas can be further extended, but we leave this issue for future work. 4. A new stochastic Fubini theorem In this section, we establish a new type of weak stochastic Fubini theorem. This is needed in e in more explicit form. However, the order to obtain in the next section the structure of X result in this section is also of independent interest in itself. The key idea for the argument in the next section is to view the process ϑ̂ as a process on Ω, indexed by u and taking values in the space of signed measures on [0, ∞); this is justified by Theorem 3.11 since (a good version of) each ϑ(m) (·, u) is of finite variation with respect to t (at least after suitable “stopping” with Tm ). Moreover, we integrate in the next section with respect to S Tm which is a Qm -square-integrable Qm -martingale. The abstract formulation we need thus involves integrals of the form ϕ . M with a martingale M and a measure-valued process ϕ. Our construction of ϕ . M is strongly inspired by Björk/Di Masi/ Kabanov/Runggaldier (1997), but there are important differences that we explain below. (∞) the space of all signed (possibly So fix a time horizon T ∈ (0, ∞), denote by IMT infinite) measures m on [0, T ] with m({0}) = 0 and by CT = C([0, T ]; IR) the space of all continuous functions f : [0, T ] → IR with the sup-norm kf k∞ . Denote the integral of f with respect to m by m(f ) := f . mT = ZT f (t) m(dt) 0 40 (∞) for f ∈ CT , m ∈ IMT (∞) and write kmkv := sup {m(f ) | f ∈ CT with kf k∞ ≤ 1} for the variation norm of m ∈ IMT . (∞) IMT the σ-field MT generated by the weak∗ topology, i.e. by all mappings © ª (∞) ¯ m 7→ m(f ) for f ∈ CT . Finally, we set IMT := m ∈ IMT ¯ kmkv < ∞ . For d ∈ IN and m = (mi )i=1,...,d ∈ IMTd , we use the notation kmkv := (kmi kv )i=1,...,d ∈ IRd . We take on A process ϕ = (ϕu )0≤u≤T defined on (Ω, F, P ) and with values in IMTd is called weakly ¢ ¡ predictable if the IRd -valued process ϕ(f ) = ϕu (f ) 0≤u≤T is predictable for each test function f ∈ CT . In view of the definition of MT , this is the same as saying that as a mapping from Ω̄ = Ω × (0, ∞) (or, more precisely, Ω̄T = Ω × (0, T ]) to IMTd , ϕ is P-MdT -measurable. We remark for later use that the IRd -valued process kϕkv = (kϕu kv )0≤u≤T is then predictable. Indeed, because CT is separable for the sup-norm, we can take a countable dense subset (fk )k∈IN of CT and obtain sup G(f ) = sup G(fk ) for any continuous function G : CT → IR. i f ∈CT k∈IN Taking G(f ) = m (f ) for i = 1, . . . , d thus yields for each (ω, u) ∈ Ω̄ that kϕi kv (ω, u) = kϕiu kv (ω) = sup f ∈CT ¢ ¢ 1 ¡ i 1 ¡ i ϕu (f ) (ω) = sup ϕu (fk ) (ω). kf k∞ k∈IN kfk k∞ Since each ϕ(fk ) is predictable, so is then kϕkv as a countable supremum. Now let M = (Mu )0≤u≤T be an IRd -valued martingale in M20 (P ) and denote by E b the family of all IMTd -valued processes ϕ of the form ϕ= L−1 X `=0 m` ID` ×(u` ,u`+1 ] with L ∈ IN , 0 ≤ u0 < u1 < · · · < uL ≤ T , m` ∈ IMTd and D` ∈ Fu` . For each f ∈ CT , L−1 P m` (f )ID` ×(u` ,u`+1 ] is then a very simple IRd -valued predictable process (on Ω), ϕ(f ) = `=0 which justifies calling ϕ ∈ E b a very simple IMTd -valued weakly predictable process (on Ω̄). Similarly, we call an IMT -valued process N = (Nu )0≤u≤T a weak martingale in M20 (P ) if ¡ ¢ the real-valued process N (f ) = Nu (f ) 0≤u≤T is in M20 (P ) for each f ∈ CT . To define an IMT -valued stochastic integral ϕ . M for general IMTd -valued ϕ, we start with ϕ ∈ E b and then extend to a larger class of integrands. Note that while both ϕ and M are d-dimensional (like b ϑ̂ and S, respectively), the resulting integral ϕ . M will be one-dimensional (like X). Lemma 4.1. Fix an IRd -valued martingale M in M20 (P ). For each ϕ ∈ E b , there exists a process ϕ . M = (ϕ . Mu )0≤u≤T with values in IMT which is a weak martingale in M20 (P ) and which has the weak Fubini property (4.1) (ϕ . Mu )(f ) = ϕ(f ) . Mu , 0 ≤ u ≤ T , for each f ∈ CT . 41 Written out in terms of integrals, (4.1) takes the form ZT f (t) 0 µ Zu ¶ ϕs dMs (dt) = 0 Zu µ ZT 0 ¶ 0 ≤ u ≤ T, f (t) ϕs (dt) dMs , 0 which explains the terminology. In particular, we have the isometry property (4.2) k(ϕ . M )(f )kM2 (P ) = k(ϕ . MT )(f )kL2 (P ) = kϕ(f ) . MT kL2 (P ) = kϕ(f )kL2 (P,M ) for f ∈ E b . Proof. Writing ϕ = L−1 P `=0 m` ID` ×(u` ,u`+1 ] , it is clear that we shall set ϕ . Mu := L−1 X `=0 ID` (Mu`+1 ∧u − Mu` ∧u )tr m` , 0≤u≤T which clearly has values in IMT . Then we get for each f ∈ CT that (4.3) (ϕ . Mu )(f ) = L−1 X `=0 ID` m` (f )tr (Mu`+1 ∧u − Mu` ∧u ) = Zu hs dMs 0 with the process (4.4) hs (ω) := L−1 X m` (f )ID` (ω)I(u` ,u`+1 ] (s) `=0 = T L−1 XZ `=0 0 = µ ZT ID` ×(u` ,u`+1 ] (ω, s)f (t) m` (dt) ¶ f (t) ϕs (dt) (ω) 0 = ϕs (f )(ω). Moreover, the predictable process h = ϕ(f ) is bounded uniformly in ω, u, by (4.5) ¯ ª © C := max kmi` kv kf k∞ ¯ i = 1, . . . , d, ` = 0, . . . , L − 1 , and hence it is in L2 (P, M ). Now (4.3) immediately shows that (ϕ . M )(f ) is a martingale and in M20 (P ), and combining (4.3) with (4.4) also directly gives (4.1). Finally, (4.2) follows from (4.1). q.e.d. 42 Remark 4.2. The above construction and setup as well as the subsequent extension to general ϕ are strongly inspired by the work of Björk/Di Masi/Kabanov/Runggaldier (1997). However, there is an important difference. In Björk/Di Masi/Kabanov/Runggaldier (1997), the goal is to develop stochastic integration with respect to a process (M̄ , say) with values in CT ; the natural integrands ϕ are then IMT -valued and the resulting integral ϕ . M̄ has values in IR. In contrast, our integrator M is IRd - or real-valued, and so using the same IMTd - or IMT -valued integrands ϕ leads to an integral process ϕ . M which is again measure-valued . ¦ To extend both ϕ . M and the weak Fubini property (4.1) from ϕ ∈ E b to more general integrands, we need appropriate seminorms on ϕ and on weak martingales N like ϕ . M . Recall that (fk )k∈IN is a countable dense subset of CT . Definition 4.3. For an IMT -valued weak martingale N = (Nu )0≤u≤T in M20 (P ), we define ¶ 21 µX ∞ 2 γk kN (fk )kM2 (P ) , rγ (N ) := k=1 where the sequence γ = (γk )k∈IN of weights in (0, ∞) will be chosen later. We then denote by Nγ2 the space of all IMT -valued weak martingales N in M20 (P ) with rγ (N ) < ∞. Definition 4.4. For an IMTd -valued weakly predictable process ϕ = (ϕu )0≤u≤T , we define ¶ 12 µX ∞ 2 γk kϕ(fk )kL2 (P,M ) , qγ (ϕ) := k=1 where the sequence γ = (γk )k∈IN of weights in (0, ∞) will be chosen later. We then denote by L2BDKR the space of all IMTd -valued weakly predictable processes ϕ with qγ (ϕ) < ∞. Because M is IRd -valued and in M20 (P ), we can represent its (matrix-valued) bracket R process similarly as in Section 3 in the form hM i , M j i = σuij dBu for i, j = 1, . . . , d, with an increasing predictable integrable RCLL process B null at 0 and a predictable process σ taking values in the symmetric nonnegative definite d × d-matrices. (Of course B and σ here are not the same as in Section 3, but this should not lead to any confusion.) For any IRd -valued predictable process ψ, we then have " ZT # (4.6) kψk2L2 (P,M ) = E[hψ . M iT ] = E ψutr σu ψu dBu = kψ tr σψkL1 (P ⊗B) . Moreover, |σ ij | ≤ √ 0 σ ii σ jj by the Kunita–Watanabe inequality, and combining this with the Cauchy–Schwarz inequality yields tr |ψ σψ| ≤ d X i,j=1 i ij j |ψ σ ψ | ≤ 43 d kψk2∞ d X i=1 σ ii for any bounded IRd -valued predictable process ψ. Integrating with respect to B and taking expectations gives (4.7) tr kψ σψkL1 (P ⊗B) ≤ d kψk2∞ d X i=1 ° °2 E[hM i iT ] = d kψk2∞ °|MT |°L2 (P ) < ∞ because M is in M20 (P ). This will be useful later on. Lemma 4.5. Suppose γk , k ∈ IN , are in (0, ∞) and satisfy ∞ P k=1 γk < ∞ and ∞ P k=1 γk kfk k2∞ ≤ 1. (Clearly, such sequences γ exist.) Then: 1) rγ and qγ are seminorms. 2) For each b > 0 and with 1 = (1, . . . , 1) ∈ IRd , the functional qγ is continuous with respect to the weak∗ topology on the subspace of L2BDKR consisting of those ϕ taking values ¯ © ª in Ubd := m ∈ IMTd ¯ kmkv ≤ b1 : If (ϕn )n∈IN is a sequence in L2BDKR with values in Ubd and converging (P ⊗ B)-almost everywhere to some Ubd -valued ϕ for the weak∗ topology, then lim qγ (ϕn ) = qγ (ϕ). n→∞ 3) For ϕ ∈ E b , we have the isometry rγ (ϕ . M ) = qγ (ϕ). (4.8) Proof. 1) It is clear that rγ and qγ are both positively homogeneous. For the triangle inequality, we start by noting that any m ∈ IMT (or IMTd ) can be viewed as a continuous (IRd -valued) linear functional on CT via f 7→ m(f ) and is hence determined by the sequence ¢ ¡ m(fk ) k∈IN of its values on the dense subset (fk )k∈IN of CT . Hence N ∈ Nγ2 can be identified ¢ ¡ with the sequence of random variables NT (fk ) k∈IN , since each N (fk ) is a martingale, and rγ2 (N ) = ∞ X k=1 γk kNT (fk )k2L2 (P ) = ∞ X γk x2k k=1 with xk := kNT (fk )kL2 (P ) , and analogously for N̄ with x̄. This shows that rγ (N ) is simply the `2 -norm of the sequence x for the measure having weights γk , k ∈ IN , on IN . But yk := kNT (fk ) + N̄T (fk )kL2 (P ) ≤ xk + x̄k by the usual triangle inequality in L2 (P ), and this gives rγ (N + N̄ ) = kyk`2 (γ) µX ¶ 12 ∞ 2 ≤ γk (xk + x̄k ) k=1 = kx + x̄k`2 (γ) ≤ kxk`2 (γ) + kx̄k`2 (γ) = rγ (N ) + rγ (N̄ ). 44 The argument for qγ is completely analogous. 2) Take a sequence (ϕn )n∈IN of weakly predictable processes with kϕn kv ≤ b1 for all n and suppose that (ϕn ) converges to ϕ for the weak∗ topology (P ⊗ B)-a.e. This implies that lim ϕnu (fk ) = ϕu (fk ) for all k ∈ IN (P ⊗ B)-a.e. and hence also (P ⊗ B)-a.e. n→∞ lim ϕnu (fk )tr σu ϕnu (fk ) = ϕu (fk )tr σu ϕu (fk ) n→∞ for all k ∈ IN . √ ¯ ¯ But |ϕnu (fk )| ≤ ¯kϕnu kv ¯kfk k∞ ≤ d bkfk k∞ (P ⊗ B)-a.e. and so we can apply dominated ¢ ¡ convergence to obtain that ϕnu (fk )tr σu ϕnu (fk ) n∈IN tends to ϕu (fk )tr σu ϕu (fk ) in L1 (P ⊗ B) for all k ∈ IN , which means by (4.6) that lim ϕn (fk ) = ϕ(fk ) in L2 (P, M ) for all k ∈ IN . n→∞ Moreover, Fatou’s lemma yields for all k ∈ IN that kϕ(fk )k2L2 (P,M ) = kϕ(fk )tr σϕ(fk )kL1 (P ⊗B) ≤ lim inf kϕn (fk )tr σϕn (fk )kL1 (P ⊗B) n→∞ ≤ sup kϕn (fk )tr σϕn (fk )kL1 (P ⊗B) n∈IN ° °2 ≤ d2 b2 kfk k2∞ °|MT |°L2 (P ) thanks to (4.7) since each ϕn has values in Ubd . So if we first take K large enough to obtain ∞ ° °2 P γk kfk k2∞ ≤ ε2 and then n large enough to get 4d2 b2 °|MT |°L2 (P ) k=K+1 γk kϕn (fk ) − ϕ(fk )k2L2 (P,M ) ≤ ε2 /K for k = 1, . . . , K, we obtain from the definition of qγ and the above estimates that qγ2 (ϕn − ϕ) = ∞ X k=1 γk kϕn (fk ) − ϕ(fk )k2L2 (P,M ) ≤ 2ε2 . This proves the assertion. 3) The isometry (4.2) in Lemma 4.1 yields kϕ(fk )k2L2 (P,M ) = k(ϕ . M )(fk )k2M2 (P ) for ϕ ∈ E b and all k ∈ IN . Hence (4.8) follows directly from the definitions of rγ and qγ . q.e.d. Lemma 4.6. If we choose the γk as in Lemma 4.5, then E b is dense in L2BDKR with respect to the seminorm qγ . Proof. This is very similar to Lemma 2.3 of Björk/Di Masi/Kabanov/Runggaldier (1997), but for completeness we give the proof in detail. First of all, we have E b ⊆ L2BDKR ; indeed, L−1 P m` ID` ×(u` ,u`+1 ] is in E b , then |ϕ(f )| ≤ Ckf k∞ with a constant C like in (4.5) if ϕ = `=0 45 only depending on ϕ, and so we obtain qγ (ϕ) < ∞ by using (4.6), (4.7), the definition of ∞ P γk kfk k2∞ < ∞. To prove denseness, note that by Banach–Alaoglu, each ball qγ and k=1 ¯ © ª Ubd := m ∈ IMTd ¯ kmkv ≤ b1 is compact in the weak∗ topology on IMTd and metrisable. Thus any P-MdT -measurable mapping ϕ : Ω̄ → Ubd can be approximated in the sense of weak∗ J P convergence by P-measurable step functions of the form ϕn = mj IAj with mj ∈ IMTd and ¡ Aj ∈ P. Hence ϕn (f ) ¢ j=1 n∈IN converges to ϕ(f ) (P ⊗ B)-a.e. for every f ∈ CT , and part 2) of Lemma 4.5 then yields that (ϕn ) tends to ϕ in the seminorm qγ since all the ϕn are chosen to have values in Ubd like ϕ. Now the predictable σ-field P is generated by the sets of the form D` × (u` , u`+1 ] with L−1 P D` ∈ Fu` , so that the real-valued predictable processes of the form g = c` ID` ×(u` ,u`+1 ] are `=0 2 dense in L (P, P ⊗ B). By passing to a subsequence, we can find for any A ∈ P a sequence (g n )n∈IN of the above form such that g n → IA (P ⊗ B)-a.e. For m ∈ Ubd and f ∈ CT , we then get first m(f )g n → m(f )IA in IRd (P ⊗ B)-a.e. and then also in L2 (P, M ) by using (4.6), (4.7) and dominated convergence. Because the same applies to finite linear combinations, we see that E b is dense, for the L2 (P ⊗ B)-norm, in the set of all P-measurable step functions ϕn as used above. But in view of the definition of qγ , this is enough to conclude that E b is dense, with respect to qγ , in the set of all weakly predictable Ubd -valued processes, and hence also in L2BDKR . q.e.d. Lemma 4.7. If we choose the γk as in Lemma 4.5, then Nγ2 is complete with respect to the seminorm rγ . Proof. This is fairly straightforward. If (N n )n∈IN is Cauchy in Nγ2 with respect to rγ , the ¢ ¡ definition of rγ implies that N n (fk ) n∈IN is Cauchy in M2 (P ) for each k ∈ IN and thus convergent to a limit process in M20 (P ) that we call N ∞,k . Since each IMT -valued weak martingale L in M2 (P ) is determined by the values of LT (fk ), k ∈ IN , we can define such a weak martingale N ∞ by setting NT∞ (fk ) := NT∞,k . For each finite K, we then get K X k=1 n γk kN (fk ) − N ∞ (fk )k2M2 (P ) = lim m→∞ K X k=1 ≤ lim sup m→∞ γk kN n (fk ) − N m (fk )k2M2 (P ) ∞ X k=1 γk kN n (fk ) − N m (fk )k2M2 (P ) ≤ ε2 for large n by the Cauchy property in Nγ2 , and this shows by letting K → ∞ that we have rγ (N n − N ∞ ) ≤ ε for n large enough. So N ∞ lies in Nγ2 and (N n ) converges to N ∞ with respect to rγ . This proves the assertion. q.e.d. 46 Putting everything together, we can now establish the main result of this section. Theorem 4.8. Fix an IRd -valued martingale M in M20 (P ), choose a countable dense subset ∞ P (fk )k∈IN of CT and let γ = (γk )k∈IN be a sequence in (0, ∞) satisfying γk < ∞ and k=1 ∞ P k=1 γk kfk k2∞ ≤ 1. For every ϕ ∈ L2BDKR , the process ϕ . M is then well defined, in Nγ2 , and we have the weak Fubini property (4.9) (ϕ . Mu )(f ) = ϕ(f ) . Mu , 0 ≤ u ≤ T , for each f ∈ CT . In particular, ϕ . M is an IMT -valued weak martingale in M20 (P ). Proof. The mapping ϕ 7→ ϕ . M from E b to Nγ2 is by construction linear, and due to the isometry in part 3) of Lemma 4.5 continuous with respect to qγ and rγ . Moreover, E b is dense in L2BDKR with respect to qγ by Lemma 4.6, and Nγ2 is complete with respect to rγ by Lemma 4.7. Therefore we can extend the above mapping in a unique way to a linear mapping from L2BDKR to Nγ2 , and the isometry (4.8) still holds for that extension by continuity so that (4.10) rγ (ϕ . M ) = qγ (ϕ) for all ϕ ∈ L2BDKR . To show that the weak Fubini property (4.1) extends from ϕ ∈ E b to ϕ ∈ L2BDKR as well, ¡ we first note that for each f ∈ CT , both sides of (4.1) and (4.9) are martingales even in ¢ M20 (P ) so that it is enough to argue for u = T . By the separability of CT , it is also enough to take only fk , k ∈ IN , instead of all f ∈ CT , so that we need to show for all k ∈ IN and for ϕ ∈ L2BDKR that (4.11) (ϕ . MT )(fk ) = ϕ(fk ) . MT . So take (ϕn )n∈IN in E b with qγ (ϕn − ϕ) → 0 as n → ∞. Then ϕn (fk ) → ϕ(fk ) in L2 (P, M ) for each k by the definition of qγ , and so ϕn (fk ) . MT → ϕ(fk ) . MT in L2 (P ) for each k by the usual isometry property of stochastic integrals. But due to (4.10) (or from the very construction), we also have rγ (ϕn . M − ϕ . M ) → 0 as n → ∞, which implies by the definition of rγ that (ϕn . MT )(fk ) → (ϕ . MT )(fk ) in L2 (P ) for each k. Because (4.11) holds for each ϕn by Lemma 4.1, we thus conclude that it holds for ϕ as well, and this ends the proof. q.e.d. For our subsequent application, we need to extend the weak Fubini property (4.9) from f ∈ CT to f being an indicator function of an interval. This can be done at the cost of slightly stronger assumptions on ϕ; but these will be satisfied in our application, thanks to part 2) of Proposition 3.7. 47 Theorem 4.9. Under the same assumptions as in Theorem 4.8, let ϕ be an IMTd -valued weakly predictable process satisfying kϕ(f )kL2 (P,M ) ≤ Ckf k∞ (4.12) for all f ∈ CT with a constant C not depending on f . Then ϕ is in L2BDKR , and for any bounded measurable function f on [0, T ], (ϕ . M )(f ) is a martingale in M20 (P ) and the weak Fubini property (4.9) holds. In particular, for f = I[0,t] , we get Zt µ Zu 0 ¶ ϕs dMs (dr) = 0 Zu µ Zt 0 ¶ 0 ≤ t ≤ T, 0 ≤ u ≤ T. ϕs (dr) dMs , 0 Proof. Using the definition of qγ and (4.12) gives qγ2 (ϕ) = ∞ X k=1 γk kϕ(fk )k2L2 (P,M ) ≤C 2 ∞ X k=1 γk kfk k2∞ < ∞ by the choice of γ so that ϕ is indeed in L2BDKR . Moreover, ϕu (f ) = RT f (t) ϕu (dt) is well 0 defined for (P ⊗B)-almost all (ω, u) for every bounded measurable function f on [0, T ] because ¯ ¯ ¯kϕu kv ¯ < ∞ (P ⊗ B)-a.e. since ϕ has values in IM d . T Now take any bounded measurable function f on [0, T ] and choose a sequence (f n )n∈IN in CT with f n → f pointwise and kf n k∞ ≤ kf k∞ . Then ϕ(f n ) → ϕ(f ) in IRd by dominated convergence, and so (ϕ(f n )tr σϕ(f n ) → ϕ(f )tr σϕ(f ) (P ⊗ B)-a.e. As a consequence, (4.12) extends from f ∈ CT to any bounded real-valued measurable function f on [0, T ]; indeed, combining Fatou’s lemma with (4.6) and (4.12) for f n yields kϕ(f )k2L2 (P,M ) = kϕ(f )tr σϕ(f )kL1 (P ⊗B) ≤ lim inf kϕ(f n )k2L2 (P,M ) n→∞ ≤ lim inf C 2 kf n k2∞ n→∞ ≤ C 2 kf k2∞ . Applying this to f n − f yields ϕ(f n ) → ϕ(f ) in L2 (P, M ) since kf n − f k∞ → 0, and so the usual isometry property of stochastic integrals gives ϕ(f n ) . M → ϕ(f ) . M in M2 (P ) and also ϕ(f n ) . MT −→ ϕ(f ) . MT in L2 (P ). In particular, ϕ(f ) . M is again a martingale in M20 (P ). But on the other hand, ϕ . MT is by construction in IMT . Hence f n → f implies as above by dominated convergence that (ϕ . MT )(f n ) −→ (ϕ . MT )(f ) 48 P -a.s., and we have (ϕ . MT )(f n ) = ϕ(f n ) . MT for each f n by (4.9) since f n ∈ CT . Hence uniqueness of the limit gives (ϕ . MT )(f ) = ϕ(f ) . MT . Replacing f by f I[0,u] gives in the same way that (ϕ . Mu )(f ) = ϕ(f ) . Mu , and so (ϕ . M )(f ) is a martingale in M20 (P ) because ϕ(f ) . M is. This completes the proof. q.e.d. At this point, it seems appropriate to comment on how our stochastic Fubini theorem and in particular Theorem 4.9 relate to the existing literature. The typical formulation is that one has a semimartingale Y , a parametric family ψ x of processes with x from some space X , and a measure η on X . For each x, the process ψ x lies in L(Y ), and so does (by R assumption) ψ̄ η := ψ x η(dx). The typical stochastic Fubini theorem then has the form Z ³Z ´ x ψ η(dx) dY = Z η ψ̄ dY = Z ³Z x ´ ψ dY η(dx); see for instance Théorème 5.44 of Jacod (1979), Lebedev (1995) [which corrects an error in Jacod (1979)], Theorem IV.65 of Protter (2005), or Bichteler/Lin (1995) [which extends the result in Protter (2005) from an L2 - to an L1 -setting]. However, all these results share the restrictive assumption that one has a fixed measure η which only depends on the parameter x and not on the randomness ω. For comparison, our setting in Section 3 has X = b [0, T ] and b ϑ(t∧Tm , ω, u∧Tm ) with x = b t. However, while each ϑ(·∧Tm , ω, u∧Tm ) is by Theorem ϕxu (ω) = 3.11 of finite variation and can hence be used as an integrator, there need not exist any measure on [0, T ] with respect to which all the ϑ(· ∧ Tm , ω, u ∧ Tm ) are absolutely continuous. Hence already our basic framework is considerably more complicated, and establishing a stochastic Fubini theorem therefore needs quite different, new techniques. The paper that is probably closest to our basic question is Protter (1985). The problem studied there is to establish the semimartingale property and decomposition for a stochastic Rt Volterra integral process of the form t 7→ Zt := ψ(t, u) dYu for a semimartingale Y . This 0 ¡ looks indeed very similar to (1.5), and the idea that we borrowed in (1.5) from Protter ¢ (1985) is to decompose Zt as Zt 0 ψ(t, u) dYu = Zt ψ(u, u) dYu + 0 Zt 0 ¡ ¢ ψ(t, u) − ψ(u, u) dYu and to use a stochastic Fubini theorem. But the crucial assumption in Protter (1985) is that t 7→ ψ(t, u) is in C 1 (with a locally Lipschitz derivative) so that one has ψ(t, u) − ψ(u, u) = 49 Zt u ∂ψ (r, u) dr. ∂t This means that with respect to the first parameter t, all the ψ(t, u) are absolutely continuous with respect to one fixed (namely Lebesgue) measure, and so things can be reduced to a “standard” stochastic Fubini theorem. This is not possible in our situation, so that our generalisation is really needed. Remark 4.10. While our stochastic Fubini theorem generalises earlier work with regard to the allowed integrands, we should in fairness also mention that we are in contrast a bit more restrictive about the integrator, M , which is a martingale. We are confident that we can extend our approach to a semimartingale; but the martingale case is sufficient for our needs here, and also (as far as we can see) already contains almost all the important new ideas. Therefore (and also for reasons of space) we have decided to postpone the more general ¦ semimartingale case to future work. 5. The structure of the process of optimal wealths b of optimal final wealths has a version X e which We have seen in Section 2 that the process X b the process X e has the form is a semimartingale. From the definition of X, et = x + X Zt ϑ̂(t, u) dSu , 0 t≥0 which is a Volterra-type of stochastic integral. With the stochastic Fubini theorem from Section 4, we now have everything we need to find the structural decomposition of the semie martingale X. Throughout this section, we again assume that the IRd -valued semimartingale S satisfies (2.9), and U satisfies (2.11)–(2.13), so that we can use all the results from Sections 2 and 3. We start with an arbitrary but fixed T ∈ (0, ∞) and construct the stopping times (Tm )m∈IN and probability measures (Qm )m∈IN via Lemma 3.1. Then we write for fixed m and 0 ≤ t ≤ T (5.1) etTm = X btTm = x + X =x+ t∧T Z m 0 Zt 0 + Zt 0 ϑ̂(t ∧ Tm , u) dSu ¡ ¢ ϑ̂(t ∧ Tm , u ∧ Tm ) − ϑ̂(u ∧ Tm , u ∧ Tm ) dSuTm ϑ̂(u ∧ Tm , u ∧ Tm ) dSuTm . 50 We recall that ϑ̂(t ∧ Tm , u ∧ Tm ) = ϑ(m) (t, u) and take for ϑ(m) the version from Proposition 3.10. Then Corollary 3.6 tells us that ϑ̂(u ∧ Tm , u ∧ Tm ) = ϑ(m) (u, u) = ϑ̄(m) u , 0≤u≤T is predictable and S Tm -integrable so that the last term in (5.1) is well defined and equals Zt (5.2) ϑ̄u(m) dSuTm = 0 Zt 0 ϑ̂(u ∧ Tm , u ∧ Tm )I{u≤Tm } dSu = t∧T Z m ϑ̂(u, u) dSu . 0 For the first term, we recall from Theorem 3.11 that each coordinate of ϑ(m) is, as a function of its first argument t, right-continuous and of finite variation. So the “random signed IRd -valued distribution functions” b 7→ ϕ̄u(m) (b) := ϑ(m) (b, u) − ϑ(m) (u, u), (5.3) 0≤u≤T on [0, ∞) can be associated with an IMTd -valued process ϕ(m) , which is clearly weakly predict¢ (m) ¡ able since ϕu (0, t] = ϑ(m) (t, u) − ϑ(m) (0, u) = ϑ(m) (t, u) is predictable as a process in u. (m) Moreover, ϕu ({0}) = ϑ(m) (0, u) = 0 since any ϑ(m) (t, u) vanishes at u = 0 and we can only take u ≤ t (= 0 here); and part 2) of Proposition 3.7 tells us that kϕ(m) (f )kL2 (Qm ,S Tm ) ° ZT ° ° ° ° ° (m) = ° f (t) ϑ (dt, ·)° ° ° 0 L2 (Qm ,S Tm ) ° (m) ° = °f . ϑT °L2 (Qm ,S Tm ) ≤ Cm kf k∞ for each bounded measurable real-valued function f on [0, T ]. More precisely, we get this inequality from Proposition 3.7 for piecewise constant measurable f ; but as pointed out in Remark 3.9, it then extends directly to all bounded measurable f on [0, T ] because we know now that ϑ(m) is of finite variation in t. Therefore the condition (4.12) in Theorem 4.9 is also satisfied, and this allows us to handle the more complicated first term in (5.1) via our stochastic Fubini theorem. Lemma 5.1. For each m ∈ IN , the real-valued process (m) Bt := Zt 0 ¡ ¢ ϑ(m) (t, u) − ϑ(m) (u, u) dSuTm , 0 ≤ t ≤ T, is RCLL, of finite variation and predictable. Proof. Because ϑ(m) (t, u) vanishes for u > t, we can write (m) Bt = ZT 0 ¡ ϑ (m) (t, u) − ϑ (m) (u, u) ¢ dSuTm = ZT 0 51 ¡ ¢ ϕu(m) (I(0,t] ) dSuTm = ϕ(m) (I(0,t] ) . STTm . (m) But the definition of ϕ(m) via (5.3) immediately shows that ϕu (IA ) = 0 for any Borel set (m) (m) A ⊆ [0, u]; hence we have ϕu (I(0,t] ) = ϕu (I(0,t] )I{u<t} and therefore (5.4) (m) Bt = ZT ϕu(m) (I(0,t] )I{u<t} dSuTm 0 Zt− ¡ (m) ¢ Tm Tm = ϕ(m) (I(0,t] ) . St− . u (I(0,t] ) dSu = ϕ 0 The discussion before Lemma 5.1 now shows that we may apply to ϕ(m) the result in Theorem 4.9, since all its assumptions are satisfied, and therefore we get (5.5) giving that ¡ ¢ ¡ ¢ ϕ(m) (I(0,t] ) . SuTm = ϕ(m) . SuTm (I(0,t] ) (m) Bt for all u ∈ [0, T ], ¡ m¢ = ϕ(m) . STT (I(0,t] ). Because ϕ(m) . S Tm has by construction values in IMT , the last expression shows that the (m) process B (m) = (Bt )0≤t≤T is RCLL and of finite variation with respect to t. To see that it ¢ ¡ Tm (I(0,s] ) on [0, ∞) × Ω̄ is is also predictable, note first that the process V (s, t) := ϕ(m) . St− ¡ (m) Tm ¢ well defined, since t 7→ ϕ . St (I(0,s] ) is a martingale by Theorem 4.9 and hence RCLL with respect to t. Moreover, V (s, t) is (B[0, T ] ⊗ P)-measurable when restricted to s ∈ [0, T ], since it is right-continuous in s, and left-continuous in t and adapted to IF . Finally, (5.5) also yields that (m) Bt ¡ ¢ Tm ¡ (m) T m ¢ = ϕ(m) (I(0,t] ) . St− = ϕ . St− (I(0,t] ) = V (t, t); so B (m) is the diagonal of V , and we conclude as in Lemma 3.5 that B (m) is indeed predictable. This ends the proof. q.e.d. The main result of this section is now the following explicit decomposition of the process of optimal wealths. Theorem 5.2. Suppose that the IRd -valued semimartingale S satisfies (2.9), and U satisfies e from Theorem 2.14 of the process X b of (2.11)–(2.13). Then the semimartingale version X optimal wealths admits a decomposition as (5.6) e =x+ X Z ϑ̄ dS + B̄, where ϑ̄ is an IRd -valued predictable S-integrable process and B̄ is a predictable RCLL process of finite variation and null at 0. More explicitly, ϑ̄ is the diagonal ϑ̄u = ϑ̂(u, u) 52 for u ≥ 0 of ϑ̂, and B̄ is given by B̄t = (5.7) Zt− ¡ 0 ¢ ϑ̂(t, u) − ϑ̂(u, u) dSu for t ≥ 0. The decomposition (5.6) is unique. Proof. 1) With the familiar stopping times Tm from Lemma 3.1, we have from (5.1), (5.2) and Lemma 5.1 that etTm X with (m) Bt = Zt 0 = Zt 0 = (m) ϑ̂(u, u) dSu + Bt 0 ¡ ¢ ϑ(m) (t, u) − ϑ(m) (u, u) dSuTm ¡ ¢ ϑ̂(t ∧ Tm , u ∧ Tm ) − ϑ̂(u ∧ Tm , u ∧ Tm ) I{u≤Tm } dSu t∧T Z m ¡ 0 = =x+ t∧T Z m ¢ ϑ̂(t, u) − ϑ̂(u, u) dSu (t−)∧T Z m 0 ¡ ¢ ϑ̂(t, u) − ϑ̂(u, u) dSu , where the last equality follows from (5.4) in the proof of Lemma 5.1. This gives (5.6) as well as (5.7). 2) In general, if Z > 0 is a local P -martingale with Z0 = 1, then also Z− > 0 and Z = E(N ) for some local P -martingale N null at 0. For any semimartingale Y , the product rule and Z = 1 + Z− . N yield ZY = Y− . Z + Z− . Y + [Z, Y ] = Y− . Z + Z− . (Y + [Y, N ]). Because Z− > 0, (2.6) therefore implies that ZY is a P -σ-martingale if and only if Y + [Y, N ] is a P -σ-martingale. 3) To prove uniqueness of the decomposition (5.6), suppose by taking differences that ϑ . S + C ≡ 0 for some IRd -valued predictable S-integrable ϑ and some predictable RCLL process C of finite variation. Then ϑ . S = −C which is locally bounded since it is predictable and RCLL. Thanks to (2.9), we can take some Z ∈ De,σ (S, P ) and write it as Z = E(N ). Since ZS is a P -σ-martingale, so is S + [S, N ] by step 2), hence also ϑ . (S + [S, N ]) = ϑ . S + [ϑ . S, N ] 53 ¡ ¢ by (2.6) for any ϑ which is S-integrable because such a ϑ is then also (S +[S, N ])-integrable . Again by step 2), this implies that the product Z(ϑ . S) is a P -σ-martingale, and this is also locally P -integrable, because Z is so as a local P -martingale and ϑ . S is locally bounded. Thus Z(ϑ . S) is a special semimartingale and therefore by Corollary 3.1 of Kallsen (2003) a local P -martingale, which means that ZC is a local P -martingale. But ZC = Z− . C+C− . Z+[Z, C] by the product rule, and the last two terms are local P -martingales like Z; for [Z, C], this is ¡ ¢ due to Yoeurp’s lemma see Theorem VII.36 in Dellacherie/Meyer (1982) . Because Z− > 0, we conclude that C is a local P -martingale; but then C ≡ C0 = 0 since C is predictable and of finite variation, and thus also ϑ . S ≡ 0. q.e.d. A. Appendix This section contains a proof which has been omitted from the main body of the paper to facilitate a more streamlined presentation. Proof of Proposition 2.7. a) Since S is a semimartingale and hence RCLL, we have d P [S i ]T < ∞ P -a.s. So setting sup |St |2 < ∞ P -a.s. and sup |∆St |2 ≤ 0≤t≤T (A.1) 0<t≤T i=1 ¶ µ dQ 2 2 2 := const. exp −W − sup |∆St | − sup |St | dP 0<t≤T 0≤t≤T clearly defines a probability measure Q ≈ P with bounded density and (2.10). Moreover, Ze,σ (S T , P ) ⊇ Ze,σ (S, P ) 6= ∅ by (2.9), and hence Ze,σ (S T , Q) 6= ∅ by Theorem 2.6. b) To prove the second half of 1), we need some results and notations from Choulli/ Stricker (2006); see also Sections III.2 and III.3 in Jacod/Shiryaev (2003). Denote by µ the random measure associated to the jumps of the (P - and Q-)semimartingale S and represent its respective compensators under P and Q as ν(ω, dt, dx) = Ft (ω, dx) dAt (ω), ν Q (ω, dt, dx) = FtQ (ω, dx) dAt (ω). £ ¯ ¤ ¯ If we denote by ZtQ := EP dQ dP Ft , 0 ≤ t ≤ T , the density process of Q with respect to P , Theorem III.3.24 of Jacod/Shiryaev (2003) gives us the kernel F Q in terms of F by µ Q¯ ¶ ¯ Q P Z ¯ Ft (ω, dx) = Mµ P̃ (ω, t, x) Ft (ω, dx). Q ¯ Z− Writing c for the constant in the definition (A.1) of Q, we get µ Q¯ ¶ ¯ 2 c P Z ¯ e−|x| P̃ (ω, t, x) ≤ Q (A.2) Mµ Q ¯ Z− Zt− (ω) 54 in view of (A.1), the definition of MµP and the fact that W ≥ 0. By Theorem 3.3 of Choulli/Stricker (2006), a sufficient condition for the existence of the minimal entropy-Hellinger σ-martingale density Ze for (S T , Q) is that for every λ ∈ IRd , Z |x|eλ tr x FtQ (ω, dx) < ∞ P -a.s. for every t ≤ T . {|x|>1} ¡ To be accurate, the results in Choulli/Stricker (2006) are given for local (instead of σ-) martingale densities for S, which means that ZS is required to be a local (instead of a σ-) martingale under P . However, all the theory extends in a straightforward way from local to σ-martingales; in fact, it becomes even simpler since one has fewer integrability conditions ¢ to check. For more details, see Section III.6e in Jacod/Shiryaev (2003), or Kallsen (2003). Using the above expression for F Q and the estimate (A.2), we obtain for every λ ∈ IRd that Z |x|e λtr x FtQ (ω, dx) ≤ {|x|>1} ≤ c Q Zt− (ω) c Q Zt− (ω) Z |x|eλ tr x −|x|2 e {|x|>1} exp µ (1 + |λ|)2 4 ¶ Ft (ω, dx) Z Ft (ω, dx) < ∞ {|x|>1} since Q ≈ P . This completes the proof of part 1) of Proposition 2.7. e Q) for (S T , Q) exists and is locally c) Thanks to 1), the entropy-Hellinger process hE (Z, Q-integrable since it is predictable and RCLL. So take as (Tm ) a localising sequence which also e Q) ∈ L1 (Q) e Then each ZeTm is a Q-uniformly integrable Q-martingale, and hE (Z, reduces Z. Tm implies that ¤ £ EQ ZeTm log ZeTm < ∞ (A.3) by Proposition 3.6 of Choulli/Stricker (2005). (The latter result also holds without the assumptions that IF is quasi-left-continuous and S is bounded and quasi-left-continuous, as can easily be seen from its proof.) Setting dQm := ZeTm dQ therefore defines for each m a probability measure Qm ≈ Q ≈ P such that S Tm is a Qm -σ-martingale. But moreover, combining (A.3) with Young’s inequality xy ≤ ex + y log y − y for any x, y ≥ 0 shows that for any random variable X ≥ 0, EQ [eX ] < ∞ implies that EQm [X] < ∞. So (2.10) gives W ∈ L2 (Qm ) and sup |St |2 ∈ L1 (Qm ), and so S Tm is indeed a Qm -square-integrable 0≤t≤T Qm -martingale. This ends the proof. q.e.d. Acknowledgments. Tahir Choulli thanks Freddy Delbaen and Martin Schweizer for their invitation to spend a sabbatical year at ETH Zurich where a large part of this research was 55 done. Their hospitality and warm welcome to the research group in mathematical finance at the Department of Mathematics in ETH Zurich is greatly appreciated. Financial support by FIM (the Institute for Mathematical Research at ETH Zurich), by Credit Suisse (through the funds of Freddy Delbaen) and by NSERC (the Natural Sciences and Engineering Research Council of Canada, Grant G121210818) is gratefully acknowledged. Martin Schweizer gratefully acknowledges financial support by the National Centre of Competence in Research “Financial Valuation and Risk Management” (NCCR FINRISK), Project D1 (Mathematical Methods in Financial Risk Management). The NCCR FINRISK is a research instrument of the Swiss National Science Foundation. Finally, helpful comments and criticisms by several people at several conferences and workshops have helped us to improve the presentation, and we thank them all for that. References S. Ankirchner (2005), “Information and Semimartingales”, PhD thesis, HU Berlin, http://www.math.hu-berlin.de/∼ankirchn/information.pdf S. Ankirchner and P. Imkeller (2005), “Finite utility on financial markets with asymmetric information and structure properties of the price dynamics”, Annales de l’Institut Henri Poincaré B 41, 479–503 J.-P. Ansel and C. Stricker (1994), “Couverture des actifs contingents et prix maximum”, Annales de l’Institut Henri Poincaré B 30, 303–315 A. Basse and J. Pedersen (2009), “Lévy driven moving averages and semimartingales”, Stochastic Processes and their Applications 119, 2970–2991 F. Biagini and B. Øksendal (2005), “A general stochastic calculus approach to insider trading”, Applied Mathematics and Optimization 52, 167–181 K. Bichteler and S. Lin (1995), “On the stochastic Fubini theorem”, Stochastics 54, 271–279 T. Björk, G. Di Masi, Yu. M. Kabanov and W. J. Runggaldier (1997), “Towards a general theory of bond markets”, Finance and Stochastics 1, 141–174 T. Choulli and M. Schweizer (2010), “Stability of σ-martingale densities in L log L under a bounded change of measure”, working paper [under construction], ETH Zürich T. Choulli and C. Stricker (2005), “Minimal entropy-Hellinger martingale measure in incomplete markets”, Mathematical Finance 15, 465–490 56 T. Choulli and C. Stricker (2006), “More on minimal entropy-Hellinger martingale measure”, Mathematical Finance 16, 1–19 T. Choulli, C. Stricker and J. Li (2007), “Minimal Hellinger martingale measures of order q”, Finance and Stochastics 11, 399–427 F. Delbaen and W. Schachermayer (1994), “A general version of the fundamental theorem of asset pricing”, Mathematische Annalen 300, 463–520 F. Delbaen and W. Schachermayer (1998), “The fundamental theorem of asset pricing for unbounded stochastic processes”, Mathematische Annalen 312, 215–250 F. Delbaen and W. Schachermayer (2006), “The Mathematics of Arbitrage”, Springer C. Dellacherie and P.-A. Meyer (1978), “Probabilities and Potential”, North-Holland, Amsterdam C. Dellacherie and P.-A. Meyer (1982), “Probabilities and Potential B. Theory of Martingales”, North-Holland, Amsterdam H. Föllmer and A. Schied (2004), “Stochastic Finance: An Introduction in Discrete Time”, second revised and extended edition, de Gruyter T. Goll and J. Kallsen (2003), “A complete explicit solution to the log-optimal portfolio problem”, Annals of Applied Probability 13, 774–799 C. Gollier and R. J. Zeckhauser (2002), “Horizon length and portfolio risk”, Journal of Risk and Uncertainty 24, 195–212 N. H. Hakansson (1969), “Optimal investment and consumption strategies under risk, an uncertain lifetime, and insurance”, International Economic Review 10, 443–466 V. Henderson and D. G. Hobson (2007), “Horizon-unbiased utility functions”, Stochastic Processes and their Applications 117, 1621–1641 T. Hida and M. Hitsuda (1993), “Gaussian Processes”, Translations of Mathematical Monographs, Volume 120, American Mathematical Society J. Jacod (1979), “Calcul Stochastique et Problèmes de Martingales”, Lecture Notes in Mathematics 714, Springer J. Jacod and A. N. Shiryaev (2003), “Limit Theorems for Stochastic Processes”, second edition, Springer S. Jaggia and S. Thosar (2000), “Risk aversion and the investment horizon: A new perspective on the time diversification debate”, Journal of Behavioral Finance 1, 211–215 57 Yu. M. Kabanov (1997), “On the FTAP of Kreps–Delbaen–Schachermayer”, in: Yu. M. Kabanov, B. L. Rozovskii and A. N. Shiryaev (eds.), ‘Statistics and Control of Stochastic Processes. The Liptser Festschrift’, World Scientific, 191–203 J. Kallsen (2003), “σ-localization and σ-martingales”, Theory of Probability and its Applications 48, 152–163 I. Karatzas and C. Kardaras (2007), “The numéraire portfolio in semimartingale financial models”, Finance and Stochastics 11, 447–493 I. Karatzas and G. Žitković (2003), “Optimal consumption from investment and random endowment in incomplete semimartingale markets”, Annals of Probability 31, 1821–1858 D. O. Kramkov (1996), “Optional decomposition of supermartingales and hedging contingent claims in incomplete security markets”, Probability Theory and Related Fields 105, 459–479 D. Kramkov and W. Schachermayer (1999), “The asymptotic elasticity of utility functions and optimal investment in incomplete markets”, Annals of Applied Probability 9, 904– 950 D. Kramkov and M. Sı̂rbu (2006), “On the two-times differentiability of the value functions in the problem of optimal investment in incomplete markets”, Annals of Applied Probability 16, 1352–1384 K. Larsen and G. Žitković (2008), “On the semimartingale property via bounded logarithmic utility”, Annals of Finance 4, 255–268 A. Lazrak (1998), “Optimal martingale restrictions and horizon effect on portfolio selection”, Prépublications de l’Equipe d’Analyse et Probabilités, No. 91, Université d’Évry, http://www.maths.univ-evry.fr/prepubli/index.html V. A. Lebedev (1995), “The Fubini theorem for stochastic integrals with respect to L -valued random measures depending on a parameter”, Theory of Probability and its Applications 40, 285–293 0 P. Protter (1985), “Volterra equations driven by semimartingales”, Annals of Probability 13, 519–530 P. Protter (2005), “Stochastic Integration and Differential Equations. A New Approach”, second edition, version 2.1 (corrected 3rd printing), Springer N. Siebenmorgen and M. Weber (2004), “The influence of different investment horizons on risk behavior”, Journal of Behavioral Finance 5/2, 75–90 58 M. E. Yaari (1965), “Uncertain lifetime, life insurance, and the theory of the consumer”, Review of Economic Studies 32, 137–150 T. Zariphopoulou and G. Žitković (2008), “Maturity-independent risk measures”, preprint, University of Texas at Austin, http://arxiv.org/abs/0710.3892v2 59
© Copyright 2026 Paperzz