Weak Dynamic Programming for Generalized State Constraints β Bruno Bouchard Marcel Nutz β October 19, 2012 Abstract We provide a dynamic programming principle for stochastic optimal control problems with expectation constraints. A weak formulation, using test functions and a probabilistic relaxation of the constraint, avoids restrictions related to a measurable selection but still implies the Hamilton-Jacobi-Bellman equation in the viscosity sense. We treat open state constraints as a special case of expectation constraints and prove a comparison theorem to obtain the equation for closed state constraints. Keywords weak dynamic programming, state constraint, expectation constraint, Hamilton-Jacobi-Bellman equation, viscosity solution, comparison theorem AMS 2000 Subject Classications 93E20, 49L20, 49L25, 35K55 1 Introduction We study the problem of stochastic optimal control under state constraints. In the most classical case, this is the problem of maximizing an expected reward, subject to the constraint that the controlled state process has to remain in a given subset πͺ of the state space. There is a rich literature on the associated partial dierential equations (PDEs), going back to [13, 14, 9, 16, 17] in the rst order case and [12, 11, 10] in the second order case. The connection between the control problem and the equation is given by the dynamic programming principle. However, in the stochastic case, it is frequent practice to make this connection only formally, and in fact, we are not aware of a generally applicable, rigorous technique in the literature. Of course, there are specic situations where it is indeed possible to avoid proving the state-constrained dynamic programming principle; in particular, penalization arguments can be useful to reduce to the unconstrained case (e.g., [11]). We refer to [6, 18] for further background. β β [email protected] [email protected]. CEREMADE, Université Paris Dauphine and CREST-ENSAE, Dept. of Mathematics, Columbia University, New York, Research supported by Swiss National Science Foundation Grant PDFM2-120424/1. The authors thank the two referees for their helpful comments. 1 Generally speaking, it is dicult to prove the dynamic programming principle when the regularity of the value function is not known a priori, due to certain measurable selection problems. It was observed in [3] that, in the unconstrained case, these diculties can be avoided by a weak formulation of the dynamic programming principle where the value function is replaced by a test function. This formulation, which is tailored to the derivation of the PDE in the sense of viscosity solutions, avoids the measurable selection and uses only a simple covering argument. It turns out that the latter does not extend directly to the case with state constraints. Essentially, the reason is that if π π₯ β πͺi.e., π₯ remains in πͺthen π may fail β² condition π₯ β πͺ . is some admissible control for the initial condition π the controlled state process ππ‘,π₯ started at to have this property for a nearby initial However, if πͺ is open and mild continuity assumptions are satised, then π ππ‘,π₯ β² will violate the state constraint with at most small probability when π₯β² is close to π₯. This observation leads us to consider optimization problems with constraints in probability, and more generally expectation constraints of the form π (π ))] β€ π for given π β β. πΈ[π(ππ‘,π₯ We shall see that, following the idea of [2], such problems are amenable to dynamic programming if the constraint level π is formulated dynamically via an auxiliary family of mar- tingales. A key insight of the present paper is that relaxing the level π by a small constant allows to prove a weak dynamic programming principle for general expectation constraint problems (Theorem 2.4), while the PDE can be derived despite the relaxation. We shall then obtain the dynamic pro- gramming principle for the classical state constraint problem (Theorem 3.1) by passing to a limit π β 0, with a suitable choice of π and certain regularity assumptions. Of course, expectation constraints are of independent interest and use; e.g., in Mathematical Finance, where one considers the problem of π (π )] under πΈ[π0,π₯ given π, π > 0. maximizing expected terminal wealth π (π ) πΈ[([π0,π₯ β π₯]β )π ] β€ π, for some the loss constraint We exemplify the use of these results in the setting of controlled diusions and show how the PDEs for expectation constraints and state constraints can be derived (Theorems 4.2 and 4.6). For the latter case, we introduce an appropriate continuity condition at the boundary, under which we prove a comparison theorem. While the above concerned an open set apply directly to the closed domain that the value function for πͺ πͺ, πͺ and does not we show via the comparison result coincides with the one for πͺ, under certain conditions. The remainder of the paper is organized as follows. In Section 2 we introduce an abstract setup for dynamic programming under expectation constraints and prove the corresponding relaxed weak dynamic programming principle. In Section 3 we deduce the dynamic programming principle for the state constraint πͺ. We specialize to the case of controlled diusions in Sec- tion 4, where we study the Hamilton-Jacobi-Bellman PDEs for expectation and state constraints. Appendix A provides the comparison theorem. 2 Throughout this paper, (in)equalities between random variables are in the almost sure sense and relations between processes are up to evanescence, unless otherwise stated. 2 Dynamic Programming Principle for Expectation Constraints π β (0, β) and a probability space (Ξ©, β±, π ) equipped π½ = (β±π‘ )π‘β[0,π ] . For each π‘ β [0, π ], we are given a set π°π‘ whose elements are seen as controls at time π‘. Given a separable metric space π , the state space, we denote by S := [0, π ] × π the time-augmented state space. For each (π‘, π₯) β S and π β π°π‘ , we are given a càdlàg adapted process π = {π π (π ), π β [π‘, π ]} with values in π , the controlled state process. ππ‘,π₯ π‘,π₯ Finally, we are given two measurable functions π, π : π β β. We assume We x a time horizon with a ltration that π πΈ[β£π (ππ‘,π₯ (π ))β£] < β so that the π πΈ[β£π(ππ‘,π₯ (π ))β£] < β and for all π β π°π‘ , (2.1) reward and constraint functions π πΉ (π‘, π₯; π) := πΈ[π (ππ‘,π₯ (π ))], π πΊ(π‘, π₯; π) := πΈ[π(ππ‘,π₯ (π ))] are well dened. We also introduce the π (π‘, π₯, π) := sup value function πΉ (π‘, π₯; π), (π‘, π₯, π) β SΜ, (2.2) πβπ° (π‘,π₯,π) where SΜ := S × β β‘ [0, π ] × π × β and { } π°(π‘, π₯, π) := π β π°π‘ : πΊ(π‘, π₯; π) β€ π is the set of controls admissible at constraint level π. Here (2.3) sup β := ββ. The following observation is the heart of our approach: Given a control π admissible at level π at the point π₯, relaxing the level π will make π adπ₯. This will be crucial for the covering missible in an entire neighborhood of arguments used below. We use the acronym u.s.c. (l.s.c.) to indicate upper (lower) semicontinuity. Lemma 2.1. Let (π‘, π₯, π) β SΜ, π β π°(π‘, π₯, π) and assume that the function π₯β² β 7 πΊ(π‘, π₯β² ; π) is u.s.c. at π₯. For each πΏ > 0 there exists a neighborhood π΅ of π₯ β π such that π β π°(π‘, π₯β² , πβ² + πΏ) for all π₯β² β π΅ and all πβ² β₯ π. Proof. π°(π‘, π₯, π). In view of the upper semicontinuity, there exists a neighborhood π΅ of π₯ such that πΊ(π‘, π₯β² ; π) β€ π + πΏ β€ πβ² + πΏ for π₯β² β π΅ ; that is, π β π°(π‘, π₯β² , πβ² + πΏ). We have πΊ(π‘, π₯; π) β€ π by the denition (2.3) of 3 A control problem with an expectation constraint of the form (2.3) is not amenable to dynamic programming if we just consider a xed level π. Extending an idea from [2, 1], we shall see that this changes if the constraint is formulated dynamically by using auxiliary martingales. end, we consider for each π = {π (π ), π β [π‘, π ]} π‘ β [0, π ] β³π‘,0 of π (π‘) = 0. We a family with initial value β³π‘,π := {π + π : π β β³π‘,0 }, We assume that, for all there exists (π‘, π₯) β S ππ‘π [π₯] β β³π‘,π where, necessarily, and also introduce π β β. π β π°π‘ , such that π (π ))]. π = πΈ[π(ππ‘,π₯ To this càdlàg martingales π (π )), ππ‘π [π₯](π ) = π(ππ‘,π₯ In particular, given π β π°π‘ , (2.4) the set π β³+ π‘,π,π₯ (π) := {π β β³π‘,π : π (π ) β₯ π(ππ‘,π₯ (π ))} is nonempty for any π (π ))]. π β₯ πΈ[π(ππ‘,π₯ More precisely, we have the follow- ing characterization. Lemma 2.2. Proof. Let (π‘, π₯) β S and π β β. Then { } π°(π‘, π₯, π) = π β π°π‘ : β³+ π‘,π,π₯ (π) β= β . π β π°π‘ . If there exists some π β β³+ π‘,π,π₯ (π), then taking expecπ tations yields πΈ[π(ππ‘,π₯ (π ))] β€ πΈ[π (π )] = π and hence π β π°(π‘, π₯, π). β² π Conversely, let π β π°(π‘, π₯, π); i.e., we have π := πΈ[π(ππ‘,π₯ (π ))] β€ π. + π π β² With ππ‘ [π₯] as in (2.4), π := ππ‘ [π₯]+πβπ is an element of β³π‘,π,π₯ (π). Let π‘ β [0, π ] an auxiliary π and π are π½π‘ -adapted for π½π‘ = (β±π π‘ )π β[0,π ] of π½ such that ππ‘,π₯ π‘ the set of all π₯ β π , π β π°π‘ and π β β³π‘,0 . Moreover, we denote by π― π‘ π½ -stopping times with values in [π‘, π ]. It will be useful in the following to consider for each subltration The following assumption corresponds to one direction of the dynamic programming principle; cf. Theorem 2.4(i) below. Assumption A. π β π―π‘ and π -a.e. (π‘, π₯, π) β SΜ, π β π°(π‘, π₯, π), π β β³+ π‘,π,π₯ (π), π π β Ξ©, there exists ππ β π°(π (π), ππ‘,π₯ (π )(π), π (π )(π)) For all such that ] ( ) [ π π (π )(π); ππ . πΈ π (ππ‘,π₯ (π ))β±π (π) β€ πΉ π (π), ππ‘,π₯ (2.5) Next, we state two variants of the assumptions for the converse direction of the dynamic programming principle; we shall comment on the dierences in Remark 2.5 below. In the rst variant, the intermediate time is deter- ministic; this will be enough to cover stopping times with countably many values in Theorem 2.4(ii) below. We recall the notation 4 ππ‘π [π₯] from (2.4). Assumption B. Let (π‘, π₯) β S, π β π°π‘ , π β [π‘, π ], π¯ β π°π and Ξ β β±π π‘ . (B1) There exists a control π Λ β π°π‘ such that πΛ π ππ‘,π₯ (β ) = ππ‘,π₯ (β ) on [π‘, π ] × (Ξ© β Ξ); (2.6) πΛ ππ‘,π₯ (β ) on [π , π ] × Ξ; (2.7) on Ξ. (2.8) π¯ (β ) ππ ,π π π‘,π₯ (π ) = [ ] πΛ π πΈ π (ππ‘,π₯ (π ))β±π β₯ πΉ (π , ππ‘,π₯ (π ); π¯) πΛ is (π , Ξ). The control B2) ( and π¯ Let π β β³π‘,0 . on denoted by π β(π ,Ξ) π¯ and called a There exists a process concatenation ¯ = {π ¯ (π), π β [π , π ]} π of π such that ( ) ¯ (β )(π) = π π¯ [π π (π )(π)](β ) (π) π π π‘,π₯ on [π , π ] π -a.s. and ( [ ] ) ¯ βπ ¯ (π ) + π (π ) 1Ξ β β³π‘,0 . π 1[π‘,π ) + 1[π ,π ] π 1Ξ©βΞ + π B3) ( π β β and π β β³+ π‘,π,π₯ (π). π (π )(π), π (π )(π)). ππ β π°(π , ππ‘,π₯ Let For π -a.e. π β Ξ©, there exist a control In the second variant, the intermediate time is a stopping time and we have an additional assumption about the structure of the sets π°π . This variant corresponds to Theorem 2.4(ii') below. Assumption B'. Let (π‘, π₯) β S, π β π°π‘ , π β π― π‘ , Ξ β β±ππ‘ and π¯ β π°β₯π β₯πΏβ . β² (B0') π°π β π°π β² for all 0 β€ π β€ π β€ π . (B1') There exists a control π Λ β π°π‘ , denoted by π β(π,Ξ) π¯, such that πΛ π ππ‘,π₯ (β ) = ππ‘,π₯ (β ) on [π‘, π ] × (Ξ© β Ξ); πΛ ππ‘,π₯ (β ) on [π, π ] × Ξ; on Ξ. π¯ ππ,π (β ) π π‘,π₯ (π ) = [ ] πΛ π πΈ π (ππ‘,π₯ (π ))β±π β₯ πΉ (π, ππ‘,π₯ (π ); π¯) B2') ( Let π β β³π‘,0 . There exists a process (2.9) (2.10) ¯ = {π ¯ (π), π β [π, π ]} π such that ( ) ¯ (β )(π) = π π¯ [π π (π )(π)](β ) (π) π π‘,π₯ π (π) on [π, π ] π -a.s. and ( [ ] ) ¯ βπ ¯ (π ) + π (π ) 1Ξ β β³π‘,0 . π 1[π‘,π ) + 1[π,π ] π 1Ξ©βΞ + π B3') ( π β β and π β β³+ π‘,π,π₯ (π). For π -a.e. π β Ξ©, there exist a control π ππ β π°(π (π), ππ‘,π₯ (π )(π), π (π )(π)). Let 5 Remark 2.3. (i) Assumption B' implies Assumption B. Moreover, As- sumption A implies (B3'). D := {(π‘, π₯, π) β SΜ : π°(π‘, π₯, π) β= β } (ii) Let denote the natural domain of our optimization problem. Then (B3') can be stated as follows: for π β π°(π‘, π₯, π) and π β β³+ π‘,π,π₯ (π), ( ) π π, ππ‘,π₯ (π ), π (π ) β D π -a.s. any π Under (B3), the same holds if for all π β π― π‘. (2.11) takes countably many values. We remark that the invariance property (2.11) corresponds to one direction in the geometric dynamic programming of [15]. (iii) We note that (2.9) (and similarly (2.7)) states in particular that ( π¯ (π, π) 7β ππ,π π ) (π) π‘,π₯ (π ) ( (π) := πππ¯(π),π π ) (π) π‘,π₯ (π )(π) (π) [π‘, π ] × Ξ. Of π¯ . (π‘, π₯) 7β ππ‘,π₯ is a well-dened adapted process (up to evanescence) on course, this is an implicit measurability condition on (iv) For an illustration of (B1'), let us assume that the controls are predictable processes. In this case, one can often take ( ) π β(π,Ξ) π¯ := π1[0,π ] + 1(π,π ] π¯1Ξ + π1Ξ©βΞ (2.12) π β(π,Ξ) π¯ β π°π‘ is called stability under concatenation. The idea is that we use the control π up to time π ; after time π , we continue using π in the event Ξ© β Ξ while we switch to the control π¯ in the event Ξ. Of course, one can omit the set Ξ β β±ππ‘ by observing and the condition that that π β(π,Ξ) π¯ = π1[0,π β² ] + π¯1(π β² ,π ] for the π½π‘ -stopping time π β² := π 1Ξ + π 1Ξ©βΞ . We can now state our weak dynamic programming principle for the stochastic control problem (2.2) with expectation constraint. The formu- lation is weak in two ways; namely, the value function is replaced by a test function and the constraint level constant πΏ>0 π is relaxed by an arbitrarily small (cf. (2.15) below). The exibility of choosing the set pearing below, will be used in Section 3. We recall the set D π ap- introduced in Remark 2.3(ii). Theorem 2.4. Let (π‘, π₯, π) β SΜ, π β π°(π‘, π₯, π) and π β β³+ π‘,π,π₯ (π). Let π‘ π π β π― and let π β SΜ be a set such that (π, ππ‘,π₯ (π ), π (π )) β π holds π -a.s. (i) Let Assumption A hold true and let π : SΜ [β [ββ, β] be a measurable ] π (π ), π (π ))β < β function such that π β€ π on π. Then πΈ π(π, ππ‘,π₯ and ] [ π (π ), π (π )) . (2.13) πΉ (π‘, π₯; π) β€ πΈ π(π, ππ‘,π₯ 6 (ii) Let πΏ > 0, let Assumption B hold true and assume that π takes countably many values (π‘π )πβ₯1 . Let π : SΜ β [ββ, β) be a measurable function such that π β₯ π on π. Assume that for any xed π¯ β π°π‘π , β« (π₯β² , πβ² ) 7β π(π‘π , π₯β² , πβ² ) is u.s.c.  β¬ π₯β² 7β πΉ (π‘π , π₯β² ; π¯) is l.s.c. on ππ (2.14)  β β² β² π₯ 7β πΊ(π‘π , π₯ ; π¯) is u.s.c. for all π β₯ 1, where ππ := {(π₯β² , πβ² ) : (π‘π , π₯β² , πβ² ) β π} β π × β. Then [ ] π π (π‘, π₯, π + πΏ) β₯ πΈ π(π, ππ‘,π₯ (π ), π (π )) . (2.15) (ii') Let πΏ > 0 and let Assumption B' hold true. Let π : SΜ β [ββ, β) be a measurable function such that π β₯ π on π. Assume that π β© D β SΜ is π -compact and that for any xed π¯ β π°π‘0 , π‘0 β [π‘, π ], β« (π‘β² , π₯β² , πβ² ) 7β π(π‘β² , π₯β² , πβ² ) is u.s.c.  β¬ (π‘β² , π₯β² ) 7β πΉ (π‘β² , π₯β² ; π¯) is l.s.c. on π β© {π‘β² β€ π‘0 }. (2.16)  (π‘β² , π₯β² ) 7β πΊ(π‘β² , π₯β² ; π¯) is u.s.c. β Then (2.15) holds true. π The following convention is used on the right hand side of (2.15): if is any random variable, we set πΈ[π ] := ββ πΈ[π + ] = πΈ[π β ] = β. if We note that Assumption (B0') ensures that the expressions πΊ(π‘β² , π₯β² ; π¯) in (2.16) are well dened for π‘β² Remark 2.5. (2.17) πΉ (π‘β² , π₯β² ; π¯) and β€ π‘0 . The dierence between parts (ii) and (ii') of the theorem [0, π ] × π × β π × β and hence no stems from the fact that in the proof of (ii') we consider as the state space while for (ii) it suces to consider assumptions on the time variable are necessary. Regarding applications, (ii) is obviously the better choice if stopping times with countably many values (and in particular deterministic times) are sucient. There is a number of cases where the extension to a general stopping time π can be accomplished a posteriori by approximation, in particular if one has a priori estimates for the value function so that one can restrict to test functions with specic growth properties. Assume for illustration that π is bounded from above, then so is functions π hurt to assume that let (ππ ) π and one will be interested only in test which are bounded from above; moreover, it will typically not π is u.s.c. (or even continuous) in all variables. Now be a sequence of stopping time taking nitely many (e.g., dyadic) values such that ππ β π π -a.s. Applying (2.15) to each 7 ππ and using Fatou's π ππ‘,π₯ time π . lemma as well as the right-continuity of the paths of nd that (2.15) also holds for the general stopping and π, we then On the other hand, it is not always possible to pass to the limit as above and then (ii') is necessary to treat general stopping times. The compactness provided that π is π -compact; e.g., π = βπ . Then, the typical way to apply (ii') is to take (π‘, π₯, π) β int D and let π be a open or closed neighborhood of (π‘, π₯, π) such that π β D. assumption should be reasonable Proof of Theorem 2.4. (i) Fix (π‘, π₯, π) β SΜ. With ππ as in (2.5), the denition (2.2) of π and π β€ π on π imply that [ ] ( ) π π πΈ π (ππ‘,π₯ (π ))β£β±π (π) β€ πΉ π (π), ππ‘,π₯ (π )(π); ππ ( ) π β€ π π (π), ππ‘,π₯ (π )(π), π (π )(π) ( ) π β€ π π (π), ππ‘,π₯ (π )(π), π (π )(π) π -a.s. After noting that the left hand side is integrable by (2.1), the result follows by taking expectations. (π‘, π₯, π) β SΜ and let π > 0. 1. Countable Cover. Like π, the set π β© D has the property ( ) π π, ππ‘,π₯ (π ), π (π ) β π β© D π -a.s.; (ii) Fix that (2.18) π takes countably many values. Replacing π π β© D if necessary, we may therefore assume that π β D. Using also that π β₯ π on π , the denition (2.2) of π shows that there exists for each (π , π¦, π) β π some π π ,π¦,π β π°(π , π¦, π) satisfying this follows from (2.11) since by πΉ (π , π¦; π π ,π¦,π ) β₯ π(π , π¦, π) β π. Fix one of the points π‘π (2.19) (π¦, π) β ππ , in time. For each the semicontinuity assumptions (2.14) and Lemma 2.1 imply that there exists a neighborhood (π¦, π) β π΅ π (π¦, π) β πΛ := π × β (of size depending on ⫠ π π‘π ,π¦,π β π°(π‘π , π¦ β² , πβ² + πΏ) β¬ β² β² π(π‘π , π¦ , π ) β€ π(π‘π , π¦, π) + π  πΉ (π‘π , π¦ β² ; π π‘π ,π¦,π ) β₯ πΉ (π‘π , π¦; π π‘π ,π¦,π ) β π β for all π¦, π, π, π, πΏ ) such that (π¦ β² , πβ² ) β π΅ π (π¦, π) β© ππ . (2.20) Here the rst inequality may read ββ β€ β πΛ is metric Λ. topology on π ββ. We note that ππ separable for the subspace topology relative to the product Λ forms an open cover {π΅ π (π¦, π) β© ππ : (π¦, π) β π} π π π Λ of π , there exists a sequence (π¦π , ππ )πβ₯1 in π such that {π΅ (π¦π , ππ ) β© π }πβ₯1 π π π‘ ,π¦ ,π π π is a countable subcover of π . We set ππ := π π π π and π΅π := π΅ (π¦π , ππ ), so Therefore, since the family that ππ β βͺ πβ₯1 8 π΅ππ . (2.21) We can now dene, for π still being xed, a measurable partition (π΄ππ )πβ₯1 of βͺπβ₯1 π΅ππ by π΄π1 := π΅1π , Since π΄ππ β π΅ππ , π π΄ππ+1 := π΅π+1 β (π΅1π βͺ β β β βͺ π΅ππ ), π β₯ 1. the inequalities (2.19) and (2.20) yield that πΉ (π‘π , π¦ β² ; πππ ) β₯ π(π‘π , π¦ β² , πβ² ) β 3π for all (π¦ β² , πβ² ) β π΄ππ β© ππ . (2.22) 2. Concatenation. Fix an integer π β₯ 1; we now focus on (π‘π )1β€πβ€π . We may assume that π‘1 < π‘2 < β β β < π‘π , by eliminating and relabeling some of π‘ the π‘π . We dene the β±π -measurable sets βͺ { } π Ξππ := π = π‘π and (ππ‘,π₯ (π‘π ), π (π‘π )) β π΄ππ β β±π‘π and Ξ(π) := Ξππ . 1β€π,πβ€π β² π‘π are distinct and π΄ππ β© π΄ππ β² = β for π β= π β² , we have Ξππ β© Ξππ β² = β (π, π) β= (πβ² , π β² ). We can then consider the successive concatenation Since the for π(π) := π β(π‘1 ,Ξ11 ) π11 β(π‘1 ,Ξ12 ) π21 β β β β(π‘1 ,Ξ1 ) ππ1 β(π‘2 ,Ξ21 ) π12 β β β β(π‘π ,Ξπ ) πππ , π π π π β π π β π π := (π π β π π ) β π π . It follows from an iterated application of (B1) that π(π) is well dened and in particular π(π) β π°π‘ . (To understand the meaning of π(π), it may be useful which is to be read from the left with to note that in the example considered in (2.12), we would have ) ( β π(π) = π1[0,π ] + 1(π,π ] π1Ξ©βΞ(π) + πππ 1Ξπ ; π 1β€π,πβ€π πππ i.e., at time π, we switch to π(π) ππ‘,π₯ on Ξ© β Ξ(π) β β±π π = ππ‘,π₯ on Ξππ .) We note that (2.6) implies that and hence ] ] [ [ π(π) π πΈ π (ππ‘,π₯ (π ))β±π = πΈ π (ππ‘,π₯ (π ))β±π Moreover, the fact that π = π‘π on on Ξ© β Ξ(π). Ξππ , repeated application of (2.23) (2.6), and (2.8) show that ] [ π(π) π πΈ π (ππ‘,π₯ (π ))β±π β₯ πΉ (π, ππ‘,π₯ (π ); πππ ) on Ξππ , for 1 β€ π, π β€ π, and we deduce via (2.22) that β ] [ π(π) π πΈ π (ππ‘,π₯ (π ))β±π 1Ξ(π) β₯ πΉ (π, ππ‘,π₯ (π ); πππ )1Ξπ π 1β€π,πβ€π β₯ β ( ) π π(π, ππ‘,π₯ (π ), π (π )) β 3π 1Ξπ π 1β€π,πβ€π π (π ), π (π ))1Ξ(π) β 3π. β₯ π(π, ππ‘,π₯ 9 (2.24) 3. Admissibility. Next, we show that there exists, for each pair such that 1 β€ π, π β€ π , π(π) β π°(π‘, π₯, π + πΏ). By (B2) πππ = {πππ (π’), π’ β [π‘π , π ]} a process ( ππ ) π πππ (β )(π) = ππ‘ππ [ππ‘,π₯ (π‘π )(π)](β ) (π) on Ξππ (2.25) and such that π (π) := (π + πΏ)1[π‘,π ) ) ( β [ ] π π + 1[π,π ] (π + πΏ)1Ξ©βΞ(π) + ππ β ππ (π‘π ) + π (π‘π ) + πΏ 1Ξπ π 1β€π,πβ€π β³π‘,π+πΏ . We note that π (π) (π ) β₯ πππ (π ) on Ξππ since πππ (π‘π ) β€ π (π‘π ) + πΏ on Ξππ as a result of the rst condition in (2.20). Hence, is an element of using (2.7) and (2.25), we have ( ππ π(π) π(ππ‘,π₯ (π )) = π ππ‘ππ,π π ) (π ) π‘,π₯ (π‘π ) This holds for all πβ 1 β€ π, π β€ π . β€ πππ (π ) β€ π (π) (π ) Ξππ . on (2.26) On the other hand, using (2.6) and that β³+ π‘,π,π₯ (π), we have that π(π) π π(ππ‘,π₯ (π )) = π(ππ‘,π₯ (π )) β€ π (π ) β€ π (π ) + πΏ = π (π) (π ) on Ξ© β Ξ(π). (2.27) π(π) π(ππ‘,π₯ (π )) β€ π (π) (π ) π(π) β π°(π‘, π₯, π + πΏ). Combining (2.26) and (2.27), we have Lemma 2.2 yields that on Ξ© and so 4. π-Optimality. We may assume that either the positive or the negative π part of π(π, ππ‘,π₯ (π ), π (π )) is integrable, as otherwise our claim (2.15) is trivial by (2.17). Using the denition (2.2) of π as well as (2.23) and (2.24), we have that [ ] π(π) π (π‘, π₯, π + πΏ) β₯ πΈ π (ππ‘,π₯ (π )) ]] [ [ π(π) = πΈ πΈ π (ππ‘,π₯ (π ))β±π [ ] [ ] π π β₯ πΈ π(π, ππ‘,π₯ (π ), π (π ))1Ξ(π) β 3π + πΈ π (ππ‘,π₯ (π ))1Ξ©βΞ(π) for every π β₯ 1. Letting π β β, we have Ξ(π) β Ξ© π -a.s. by (2.18) and (2.21); therefore, [ ] π πΈ π (ππ‘,π₯ (π ))1Ξ©βΞ(π) β 0 by dominated convergence and (2.1). Moreover, monotone convergence yields [ ] [ ] π π πΈ π(π, ππ‘,π₯ (π ), π (π ))1Ξ(π) β πΈ π(π, ππ‘,π₯ (π ), π (π )) ; to see this, consider separately the cases where the positive or the negative part of π (π ), π (π )) π(π, ππ‘,π₯ are integrable. Hence we have shown that π π (π‘, π₯, π + πΏ) β₯ πΈ[π(π, ππ‘,π₯ (π ), π (π ))] β 3π. 10 As π>0 was arbitrary, this completes the proof of (ii). (π‘, π₯, π) β SΜ and let π > 0. In contrast to the proof of (ii), we SΜ rather than π × β. By (2.11), we may again assume that π β© D = π . Since π β₯ π on π , the denition (2.2) of π shows that π ,π¦,π β π°(π , π¦, π) satisfying there exists for each (π , π¦, π) β π some π (ii') Fix shall cover a subset of πΉ (π , π¦; π π ,π¦,π ) β₯ π(π , π¦, π) β π. Given (π , π¦, π) β π, (2.28) the semicontinuity assumptions (2.16) and a variant of Lemma 2.1 (including the time variable) imply that there exists a set π΅(π , π¦, π) β SΜ of the form ( ) π΅(π , π¦, π) = (π β π, π ] β© [0, π ] × π΅π (π¦, π), where π>0 and π΅π (π¦, π) π × β+ , is an open ball in ⫠ β¬ π π ,π¦,π β π°(π β² , π¦ β² , πβ² + πΏ) π(π β² , π¦ β² , πβ² ) β€ π(π , π¦, π) + π  πΉ (π β² , π¦ β² ; π π ,π¦,π ) β₯ πΉ (π , π¦; π π ,π¦,π ) β π β such that (π β² , π¦ β² , πβ² ) β π΅(π , π¦, π) β© π. for all (2.29) Note that we have exploited Assumption (B0'), which forces us to use the half-closed interval for the time variable. As a result, the product topology on and [0, π ] [0, π ]×π ×β if π and π΅(π , π¦, π) is open for β are given the usual topology is given the topology generated by the half-closed intervals (the upper limit topology). Let us denote the latter topological space by [0, π ]β . β β² Like the Sorgenfrey line, [0, π ] is a Lindelöf space. Let π be the canoniβ² cal projection of π β SΜ β‘ [0, π ]×π×β to π×β. Then π is again π -compact. π -compact space π β was π -compact for the original topology on SΜ and the topology of [0, π ] is β β² ner than the original one, we still have that π β [0, π ] × π is a countable union of closed subsets and therefore π is Lindelöf also for the new topology. By the Lindelöf property, the cover {π΅(π , π¦, π) β© π : (π , π¦, π) β SΜ} of π π ,π¦ ,π admits a countable subcover {π΅(π π , π¦π , ππ ) β© π}πβ₯1 . We set ππ := π π π π and π΅π := π΅(π π , π¦π , ππ ), then π β βͺπβ₯1 π΅π and It is easy to see that the product of a Lindelöf space with a is again Lindelöf; in particular, π΄1 := π΅1 , [0, π ]β × πβ² is Lindelöf. Moreover, since π΄π+1 := π΅π+1 β (π΅1 βͺ β β β βͺ π΅π ), denes a measurable partition of βͺπβ₯1 π΅π . Since πβ₯1 π΄π β π΅ π , the inequali- ties (2.28) and (2.29) yield that πΉ (π β² , π¦ β² ; ππ ) β₯ π(π β² , π¦ β² , πβ² ) β 3π Similarly as above, we rst x π β₯ 1, for all (π β² , π¦ β² , πβ² ) β π΄π β© π. dene the { } π Ξπ := (π, ππ‘,π₯ (π ), π (π )) β π΄π and β±ππ‘ -measurable sets βͺ Ξ(π) := Ξπ 1β€πβ€π 11 and set ( (( ) ) ) π(π) := β β β π β(π,Ξ1 ) π1 β(π,Ξ2 ) π2 β β β β(π,Ξπ ) ππ . ] [ π(π) π πΈ π (ππ‘,π₯ (π ))β±π β₯ πΉ (π, ππ‘,π₯ (π ); ππ ) we use that π β€ π π πΛ := π β§ π π satisfying on on Ξπ for To check that 1 β€ π β€ π, Ξπ , so that we can apply (2.10) with the stopping time β₯Λ π β₯πΏβ β€ π π and thus ππ β π°π π β π°β₯Λπ β₯πΏβ ; c.f. (B0'). The rest of the proof is analogous to the above. Remark 2.6. π -compactness in Theorem 2.4(ii') was used only to ensure the Lindelöf property of πβ©D for the topology introduced The assumption on in the proof. Therefore, any other assumption ensuring this will do as well. Let us record a slight generalization of Theorem 2.4(ii),(ii') to the case of controls which are not necessarily admissible at the given point (π‘, π₯, π). The intuition for this result is that the dynamic programming principle holds as before if we use such controls for a suciently short time (as formalized by condition (2.30) below) and then switch to admissible ones. More precisely, the proof also exploits the relaxation which is anyway present in (2.15). We use the notation of Theorem 2.4. Corollary 2.7. Let the assumptions of Theorem 2.4(ii) hold true except that π β π°π‘ and π β β³π‘,π are not necessarily in π°(π‘, π₯, π) and β³+ π‘,π,π₯ (π), respectively. In addition, assume that ( ) π π, ππ‘,π₯ (π ), π (π ) β D π -a.s. (2.30) Then the conclusion (2.15) of Theorem 2.4(ii) still holds true. Moreover, the same generalization holds true for Theorem 2.4(ii'). Proof. Let us inspect the proof of Theorem 2.4(ii). Using directly (2.30) rather than appealing to (2.11), the construction of the covering in Step 1 remains unchanged and the same is true for the concatenation in Step 2. In Step 3, we proceed as above up to and including (2.26). Note that (2.27) no longer holds as it used the assumption that π(π) view of (2.26) and ππ‘,π₯ (π ) = π (π ) on ππ‘,π₯ π β β³+ π‘,π,π₯ (π). Ξ© β Ξ(π), However, in we still have [ ] [ ] [ ] π(π) π πΈ π(ππ‘,π₯ (π )) β€ πΈ π (π) (π )1Ξ(π) + πΈ π(ππ‘,π₯ (π ))1Ξ©βΞ(π) . π (π )) is integrable by (2.1) and Ξ(π) β Ξ©, the latter expectation π(ππ‘,π₯ is bounded by πΏ for large π . Moreover, Ξ(π) β β±π , the martingale property (π) , and the fact that π (π) = π + πΏ on [0, π ] yield that of π [ ] [ ] [ ] πΈ π (π) (π )1Ξ(π) = πΈ π (π) (π )1Ξ(π) = πΈ (π (π ) + πΏ)1Ξ(π) . Since Since πΈ[π (π )] = π, the right hand side is dominated by Together, we conclude that π(π) πΈ[π(ππ‘,π₯ (π ))] β€ π + 3πΏ; 12 π + 2πΏ for large π. i.e., π(π) β π°(π‘, π₯, π + 3πΏ) π . Step 4 of the previous proof then π (π )) is integrable by (2.1)), except that π (ππ‘,π₯ π(π) now results in for all large applies as before (recall that the changed admissibility of π π (π‘, π₯, π + 3πΏ) β₯ πΈ[π(π, ππ‘,π₯ (π ), π (π ))]. However, since πΏ > 0 was arbitrary, this is the same as (2.15). The argument to extend Theorem 2.4(ii') is analogous. While we shall see that the relaxation πΏ>0 in (2.15) is harmless for the derivation of the Hamilton-Jacobi-Bellman equation, it is nevertheless inter- πΏ esting to know when the in π. can be omitted; i.e., when π is right continuous The following sucient condition will be used when we consider state constraints. Lemma 2.8. such that Let (π‘, π₯, π) β D. For each πΏ > 0 there is π πΏ β π°(π‘, π₯, π + πΏ) πΉ (π‘, π₯; π πΏ ) β₯ πΏ β1 β§ π (π‘, π₯, π + πΏ) β πΏ. Let πΏ0 > 0 and assume that for all 0 < πΏ β€ πΏ0 there exists πΛπΏ β π°(π‘, π₯, π) such that { ππΏ } πΛπΏ lim π ππ‘,π₯ (π ) β= ππ‘,π₯ (π ) = 0 πΏβ0 and such that the set {[ ]+ } ππΏ πΛπΏ π (ππ‘,π₯ (π )) β π (ππ‘,π₯ (π )) : 0 < πΏ β€ πΏ0 β πΏ1 (π ) (2.31) is uniformly integrable. Then πβ² 7β π (π‘, π₯, πβ² ) is right continuous at π. Proof. πβ² 7β π°(π‘, π₯, πβ² ) is increasing, we have that (π‘, π₯, π) β D πΏ implies π°(π‘, π₯, π + πΏ) β= β for all πΏ β₯ 0. Hence π exists; of course, the β1 is necessary only if π (π‘, π₯, π + πΏ) = β. Moreover, the truncation at πΏ β² β² monotonicity of π 7β π (π‘, π₯, π ) implies that the right limit π (π‘, π₯, π+) exists and that π (π‘, π₯, π+) β₯ π (π‘, π₯, π); it remains to prove the opposite πΏ ππΏ πΛπΏ inequality. Let 0 < πΏ β€ πΏ0 and set π΄ := {ππ‘,π₯ (π ) β= ππ‘,π₯ (π )}, then Since π (π‘, π₯, π) β₯ πΉ (π‘, π₯, πΛπΏ ) [ ( )] ππΏ πΛπΏ = πΉ (π‘, π₯, π πΏ ) β πΈ 1π΄πΏ π (ππ‘,π₯ (π )) β π (ππ‘,π₯ (π )) [ ( )+ ] ππΏ πΛπΏ β₯ πΏ β1 β§ π (π‘, π₯, π + πΏ) β πΏ β πΈ 1π΄πΏ π (ππ‘,π₯ (π )) β π (ππ‘,π₯ (π )) . Letting πΏ β 0, Remark 2.9. we deduce by (2.31) that π (π‘, π₯, π) β₯ π (π‘, π₯, π+). The integrability assumption (2.31) is clearly satised if π is bounded. In the general case, it may be useful to consider the value function for a truncated function π in a rst step. 13 Remark 2.10. Our results can be generalized to a setting with multiple π β β and π β βπ , let { π°(π‘, π₯, π) := π β π°π‘ : πΊπ (π‘, π₯; π) β€ ππ constraints. Given for } π = 1, . . . , π , π (π ))] for some measurable function π π . In πΊπ (π‘, π₯; π) := πΈ[π π (ππ‘,π₯ this case, β³π‘,0 is dened as the set of càdlàg π -dimensional martingales π = {π (π ), π β [π‘, π ]} with initial value π (π‘) = 0, and where π β βπ . β³π‘,π := {π + π : π β β³π‘,0 }, This generalization, which has been considered in [4] within the framework of stochastic target problems with controlled loss, also allows to impose constraints at nitely many intermediate times 0 β€ π1 β€ π2 β€ β β β β€ π . Indeed, we can increase the dimension of the state process and add the components π (β β§ π ). ππ‘,π₯ π 3 Application to State Constraints πͺ β π := βπ We consider an open set and study the stochastic control problem under the constraint that the state process has to stay in πͺ. Namely, we consider the value function π¯ (π‘, π₯) := (π‘, π₯) β S, sup πΉ (π‘, π₯; π), (3.1) πβπ°¯(π‘,π₯) where { π ¯ π₯) := π β π°π‘ : ππ‘,π₯ π°(π‘, (π ) β πͺ for all } π β [π‘, π ], π -a.s. . In the following discussion we assume that, for all π ππ‘,π₯ (π‘, π₯) 7β (π‘, π₯) β S and π β π°π‘ , has continuous paths; (3.2) π (π) is continuous in probability, uniformly in ππ‘,π₯ ¯ π₯) β= β π°(π‘, for π; (π‘, π₯) β [0, π ] × πͺ. (3.3) (3.4) (π‘π , π₯π ) β (π‘, π₯) implies ( π ) π sup π ππ‘π ,π₯π (π), ππ‘,π₯ (π) β 0 in probability, Explicitly, the condition (3.3) means that πβ[0,π ] where we set π (π) := π₯ ππ‘,π₯ for π < π‘, π(β , β ) π β π°π‘π and it is implicitly assumed that denotes the Euclidean metric, for all π. We shall augment the state process so that the state constraint becomes a special case of an expectation constraint. π(π₯) := inf{π(π₯, π₯β² ) : π₯β² To this end, we introduce the distance function β π β πͺ} for π₯βπ π π ππ‘,π₯,π¦ (π ) := π¦ β§ inf π(ππ‘,π₯ (π)), πβ[π‘,π ] 14 and the auxiliary process π β [π‘, π ], π¦ β [0, β). (3.5) π (π)(π), π β [π‘, π ]} β π is compact; therefore, {ππ‘,π₯ distance to π β πͺ whenever it is contained in πͺ : By (3.2), each trajectory it has strictly positive π {ππ‘,π₯ (π)(π), π β [π‘, π ]} β πͺ if and only if π ππ‘,π₯,1 (π )(π) > 0. We consider the augmented state process ( π ) π π ¯ π‘,π₯,π¦ π (β ) := ππ‘,π₯ (β ), ππ‘,π₯,π¦ (β ) on the state space π × [0, β), then π π ¯ π‘,π₯,π¦ πΈ[π(π (π ))] = π {ππ‘,π₯,π¦ (π ) β€ 0} for (π₯, π¦) β π × [0, β). Now the ¯ π (π ))] β€ 0 and therefore πΈ[π(π π‘,π₯,1 ¯ π₯) = π°(π‘, π₯, 1, 0) π°(π‘, of course, the value by setting π(π₯, π¦) := 1(ββ,0] (π¦) state constraint may be expressed as and π¯ (π‘, π₯) = π (π‘, π₯, 1, 0); 1 may be replaced by any number π¦ > 0. Here and in the sequel, we use the notation from the previous section applied to the controlled ¯ π π × [0, β); i.e., we tacitly replace the variable π₯ by (π₯, π¦) π°(π‘, π₯, π¦, π) of admissible controls and the associated value π (π‘, π₯, π¦, π). state process on to dene the set function One direction of the dynamic programming principle will be a consequence of the following counterpart of Assumption A. Assumption there exists ¯ π₯), π β π― π‘ and π -a.e. π β Ξ©, AΜ. For all (π‘, π₯) β S, π β π°(π‘, π ¯ ππ β π°(π (π), ππ‘,π₯ (π )(π)) such that ] [ ( ) π π πΈ π (ππ‘,π₯ (π ))β±π (π) β€ πΉ π (π), ππ‘,π₯ (π )(π); ππ . The more dicult direction of the dynamic programming principle will be inferred from Theorem 2.4 under a right-continuity condition; we shall exemplify in the subsequent section how to verify this condition. Theorem 3.1. ¯ π₯)} β π― π‘ . Consider (π‘, π₯) β S and a family {π π , π β π°(π‘, (i) Let Assumption π΄¯ hold true and let π[: S β [ββ, β] be ] a measurable π (π π ))β < β for all function such that π¯ β€ π. Then πΈ π(π π , ππ‘,π₯ ¯ π₯) and π β π°(π‘, [ ] π π¯ (π‘, π₯) β€ sup πΈ π(π π , ππ‘,π₯ (π π )) . πβπ°¯(π‘,π₯) ¯ on π × [0, β) and (ii) Let Assumption B' hold true for the state process π let (3.2)(3.4) hold true. Moreover, assume that π (π‘, π₯, 1, 0) = π (π‘, π₯, 1, 0+) 15 (3.6) and that (π β² , π₯β² ) 7β πΉ (π β² , π₯β² ; π) is l.s.c. on [0, π‘0 ] × πͺ for all π‘0 β [π‘, π ] and π β π°π‘0 . Then [ ] π π¯ (π‘, π₯) β₯ sup πΈ π(π π , ππ‘,π₯ (π π )) (3.7) πβπ°¯(π‘,π₯) for any u.s.c. function π : S β [ββ, β) such that π¯ β₯ π. Proof. ¯ π₯) β= β as otherwise π¯ (π‘, π₯) = ββ. As π°(π‘, π in the proof of (2.13), we obtain that πΉ (π‘, π₯; π) β€ πΈ[π(π, ππ‘,π₯ (π ))] for all ¯ π₯). The claim follows by taking supremum over π . π β π°(π‘, ¯ π₯) β= β as otherwise the right hand (ii) Again, we may assume that π°(π‘, side of (3.7) equals ββ. We set (i) We may assume that π := [π‘, π ] × πͺ × (0, β) × {0}. ¯ π₯), then π(π ¯ π (π )) = 0 and hence the constant martingale π β π°(π‘, π‘,π₯,1 π π := 0 is contained in β³+ π‘,0,(π₯,1) (π). Moreover, ππ‘,π₯,1 (π ) > 0 and hence ¯ π (π ), π (π )) β π for all π β [π‘, π ], π -a.s. Furthermore, if we dene (π , π π‘,π₯,1 { π(π‘β² , π₯β² ), (π‘β² , π₯β² , π¦ β² , πβ² ) β π π(π‘β² , π₯β² , π¦ β² , πβ² ) := ββ, otherwise, If then π is u.s.c. on π. To see that the third semicontinuity condition in (2.16) is also satised, note that for any and (π‘β² , π₯β² , π¦ β² ), (π‘β²β² , π₯β²β² , π¦ β²β² ) β [0, π‘0 ] × π × [0, β) π β π°π‘0 , β£ππ‘πβ² ,π₯β² ,π¦β² (π )βππ‘πβ²β² ,π₯β²β² ,π¦β²β² (π )β£ β€ β£π¦ β² β π¦ β²β² β£ + inf π(ππ‘πβ² ,π₯β² (π)) β inf π(ππ‘πβ²β² ,π₯β²β² (π)) πβ[π‘β² ,π ] πβ[π‘β²β² ,π ] ) ( β€ β£π¦ β² β π¦ β²β² β£ + sup π ππ‘πβ² ,π₯β² (π), ππ‘πβ²β² ,π₯β²β² (π) . πβ[0,π ] Hence (3.3) implies that As (ββ, 0] β β (π‘β² , π₯β² , π¦ β² ) 7β ππ‘πβ² ,π₯β² ,π¦β² (π ) is continuous in probability. is closed, we conclude by the Portmanteau theorem that { } (π‘β² , π₯β² , π¦ β² ) 7β π ππ‘πβ² ,π₯β² ,π¦β² (π ) β (β, 0] β‘ πΊ(π‘β² , π₯β² , π¦ β² ; π) as required. Since any open subset of a Euclidean space is since (3.4) implies π =0 π β© D = π, = π (π‘, π₯, 1, 0+) [ ( )] π π β₯ πΈ π π π , ππ‘,π₯ (π π ), ππ‘,π₯,1 (π π ), 0 π = πΈ[π(π π , ππ‘,π₯ (π π ))]. ¯ π₯) π β π°(π‘, π -compact and we can use (3.6) and Theorem 2.4(ii') with to obtain that π¯ (π‘, π₯) = π (π‘, π₯, 1, 0) As is u.s.c. was arbitrary, the result follows. 16 Remark 3.2. Similarly as in Theorem 2.4(ii), there is also a version of The- orem 3.1(ii) for stopping times taking countably many values. In this case, Assumption B replaces Assumption B', all conditions on the time variable are superuous, and one can consider a general separable metric space π. 4 Application to Controlled Diusions In this section, we show how the weak formulation of the dynamic programming principle applies in the context of controlled Brownian stochastic differential equations and how it allows to derive the Hamilton-Jacobi-Bellman PDEs for the value functions associated to optimal control problems with expectation or state constraints. As the main purpose of this section is to illustrate the use of Theorems 2.4 and 3.1, we shall choose a fairly simple setup allowing to explain the main points without too many distractions. Given the generality of those theorems, extensions such as singular control, mixed control/stopping problems, etc. do not present any particular diculty. 4.1 Setup for Controlled Diusions From now on, we take π = βπ Ξ© = πΆ([0, π ]; βπ ) measure on Ξ©, and π and let be the space π the Wiener the canonical ππ‘ (π) = ππ‘ . Let π½ = (β±π‘ )π‘β[0,π ] be the augmentation of the ltration generated by π ; without loss of generality, β± = β±π . For π‘ β [0, π ], the π‘ π‘ auxiliary ltration π½ = (β±π )π β[0,π ] is chosen to be the augmentation of π(ππ β ππ‘ , π‘ β€ π β€ π ); in particular, π½π‘ is independent of β±π‘ . π Consider a closed set π β β and let π° be the set of all π -valued preβ«π 2 dictable processes π satisfying πΈ[ 0 β£ππ‘ β£ ππ‘] < β. Then we set { } π°π‘ = π β π° : π is π½π‘ -predictable . of continuous paths, process This choice will be convenient to verify Assumption B'. We remark that the π½π‘ -predictable controls entails no loss of generality, in the sense alternative choice π°π‘ = π° would result in the same value function. restriction to that the Indeed, this follows from a well known randomization argument (see, e.g., [3, Remark 5.2]). Let ππ denote the set of π×π matrices. Given two Lipschitz continuous functions π : βπ × π β βπ , and π : βπ × π β ππ π (β ) as the unique strong solution (π‘, π₯, π) β [0, π ] × βπ × π° , we dene ππ‘,π₯ of the stochastic dierential equation (SDE) β« π(π ) = π₯ + π π β« π(π(π), ππ ) ππ + π‘ π(π(π), ππ ) πππ , π‘ 17 π‘ β€ π β€ π, (4.1) where we set π β π°, π (π) = π₯ ππ‘,π₯ for π β€ π‘. As π (π ) ππ‘,π₯ is square integrable for any (2.1) is satised whenever π and π have quadratic growth, (4.2) which we assume from now on. In addition, we also impose that { π π π β has subquadratic growth, + has subquadratic growth, and π is l.s.c. and is u.s.c. (4.3) β : βπ β β is said to have subquadratic growth if β(π₯)/β£π₯β£2 β 0 as β£π₯β£ β β. This will be used to obtain the semicontinuity properties (2.16). Furthermore, we take β³π‘,0 to be the family of all càdlàg martingales π‘ which start at 0 and are adapted to π½ . By the independence of the increments of π , we see that π β β³π‘,0 is then also a martingale in the ltration π (π ) β πΏ2 (β± π‘ , π ) π½π‘ and that ππ = 0 for π β€ π‘. For π β π°π‘ , we have ππ‘,π₯ π π π π‘ and hence (2.4) is satised with ππ‘ [π₯](β ) = πΈ[ππ‘,π₯ (π )β£β±β ]. It will be useful to express the martingales as stochastic integrals. Let ππ‘ denote the set of β«π βπ -valued π½π‘ -predictable processes πΌ such that 0 β£πΌπ‘ β£2 ππ‘ < β π -a.s. and such that β« β πΌ ππ‘,0 (β ) := πΌπ β€ πππ where π‘ β€ is a martingale ( denotes transposition). Then the Brownian representation theorem yields that { πΌ } β³π‘,0 = ππ‘,0 : πΌ β ππ‘ . In the following, we also use the notation πΌ := π + π πΌ . ππ‘,π π‘,0 Lemma 4.1. In the above setup, Assumptions A, π΄¯, B, B' are satised and πΉ and πΊ satisfy (2.16). Proof. Assumption (B0') is immediate from the denition of π°π‘ . We dene the concatenation of controls by (2.12). The validity of Assumptions A, AΜ and (B1') follows from the uniqueness ππ in Assumption A ππ (π β² ) := π(π βπ π β² ), π β² β Ξ©, where the concatenated path and the ow property of (4.1); in particular, the control can be dened by π βπ π β² is given by ( ) (π βπ π β² )π := ππ 1[0,π (π)] (π) + ππβ² β ππβ² (π) + ππ (π) 1(π (π),π ] (π). While we refer to [3, Proposition 5.4] for a detailed proof, we emphasize that the choice of π°π is crucial for the validity of (2.10): in the notation of Assumption B', (2.10) essentially requires that Let π‘, π, π, π¯ π¯ be independent of β±π . be as in Assumption B', we show that (B2') holds. Let be a càdlàg version of πΛ ¯ (π) := πΈ[π(ππ‘,π₯ π (π ))β£β±π ], π β [0, π ], 18 where πΛ := π1[0,π ] + 1(π,π ] π¯. ¯ π By the same argument as in [3, Proposition 5.4], we deduce from the uniqueness and the ow property of (4.1) and the fact that π¯ is independent of β±π that πΛ π¯ πΈ[π(ππ‘,π₯ (π ))β£β±π ] = πΈ[π(ππ,π π π‘,π₯ (π ) π (π ))β£β±π ] = πππ¯ [ππ‘,π₯ (π )](π) on [π, π ]. ¯ = π π¯ [π π (π )] on [π, π ]. The last assertion of (B2') is clear by the π π π‘,π₯ denition of β³π‘,0 . As already mentioned in Remark 2.3, Assumption (B3') Hence follows from Assumption A and Assumption B follows from Assumption B'. πΉ satises (2.16); i.e., that πΉ is l.s.c. For xed π β π° , πΏ2 -continuous. Hence the semicontinuity from (4.3) π + and Fatou's lemma yield that (π‘, π₯) 7β πΈ[π (ππ‘,π₯ (π )) ] is l.s.c. By the subπ β : (π‘, π₯) β π΅} is quadratic growth from (4.3), we have that {π (ππ‘,π₯ (π )) uniformly integrable whenever π΅ β S is bounded, hence the semicontinuity π β of π also yields that (π‘, π₯) 7β πΈ[π (ππ‘,π₯ (π )) ] is u.s.c. As a result, πΉ is l.s.c. The same arguments show that πΊ also satises (2.16). Next, we check that π (π ) (π‘, π₯) 7β ππ‘,π₯ is 4.2 PDE for Expectation Constraints In this section, we show how to deduce the Hamilton-Jacobi-Bellman PDE for the optimal control problem (2.2) with expectation constraint from the weak dynamic programming principle stated in Theorem 2.4. Given a suit- π(π‘, π₯) on [0, π ] × βπ , we shall denote by βπ‘ π 2 its derivative with respect to π‘ and by π·π and π· π the Jacobian and the Hessian matrix with respect to π₯, respectively. ably dierentiable function In the context of the setup introduced in the preceding Section 4.1, the Hamilton-Jacobi-Bellman operator is given by π»(π₯, π, π) := inf ( (π’,π)βπ ×βπ ) β πΏπ’,π (π₯, π, π) , (π₯, π, π) β βπ × βπ+1 × ππ+1 , where 1 β€ πΏπ’,π (π₯, π, π) := ππ,π (π₯, π’)β€ π+ Tr[ππ,π ππ,π (π₯, π’, π)π], 2 (π’, π) β π ×βπ is the Dynkin operator with coecients ( ππ,π (π₯, π’) := Since the set π × βπ π(π₯, π’) 0 ) ( and is unbounded, π» ππ,π (π₯, π’, π) := π(π₯, π’) πβ€ ) . may be discontinuous and viscos- ity solution properties need to be stated in terms of the upper and lower semicontinuous envelopes of π», π» β (π₯, π, π) := lim sup π»(π₯β² , πβ² , πβ² ), (π₯β² ,πβ² ,πβ² )β(π₯,π,π) π»β (π₯, π, π) := lim inf (π₯β² ,πβ² ,πβ² )β(π₯,π,π) 19 π»(π₯β² , πβ² , πβ² ). The value function π dened in (2.2) may also be discontinuous and so we introduce π β (π‘, π₯, π) := π (π‘β² , π₯β² , πβ² ), lim sup β² β² β² (π‘ , π₯ , π ) β (π‘, π₯, π) (π‘β² , π₯β² , πβ² ) β int D πβ (π‘, π₯, π) := lim inf (π‘β² , π₯β² , πβ² ) β (π‘, π₯, π) (π‘β² , π₯β² , πβ² ) β int D π (π‘β² , π₯β² , πβ² ). Here int D denotes the parabolic interior; i.e, the interior of D β {π‘ = π } in SΜ, where {π‘ = π } := {(π‘, π₯, π) β SΜ : π‘ = π }. Moreover, we shall denote by D the closure of D. The main result of this subsection is the following PDE. Theorem 4.2. Assume that π is locally bounded on int D. (i) The function π β is a viscosity subsolution on D β {π‘ = π } of ββπ‘ π + π»β (β , π·π, π·2 π) β€ 0. (ii) The function πβ is a viscosity supersolution on int D of ββπ‘ π + π» β (β , π·π, π·2 π) β₯ 0. We refer to [5] for the various equivalent denitions of viscosity superand subsolutions. We merely mention that subsolution on π΄ means that π΄ which are local maxima of π β βπ on π΄, where π is a test function, and analogously for the supersolution. the subsolution property is satised at points of We shall not discuss in this generality the boundary condition and the validity of a comparison principle. In the subsequent section, these will be studied in some detail for the case of state constraints. We also refer to [1] for the study of the boundary conditions in a similar framework. Remark 4.3. We observe that the domain of the PDE in Theorem 4.2 is not given a priori; it is itself characterized by a control problem: if we dene π π£(π‘, π₯) := inf πΈ[π(ππ‘,π₯ (π ))], πβπ°π‘ (π‘, π₯) β S, (4.4) then { } int D = (π‘, π₯, π) β SΜ : π > π£ β (π‘, π₯), π‘ < π , π£ β is the upper semicontinuous envelope of π£ on [0, π ) × βπ . In particint D β= β since π£ is locally bounded from above. In fact, in the present where ular, setup, we also have { } int D = (π‘, π₯, π) β SΜ : π > π£(π‘, π₯), π‘ < π . (4.5) Indeed, a well known randomization argument (e.g., [3, Remark 5.2]) yields that π (π ))] π£(π‘, π₯) = inf πβπ° πΈ[π(ππ‘,π₯ for all 20 (π‘, π₯) β S; i.e., the set π°π‘ in (4.4) π°. π£ inherits the upper semicontinuity of πΊ π£ = π£ β . Using (4.5), we obtain that { } { } (π‘, π₯, π) β SΜ : π β₯ π£(π‘, π₯) β (π‘, π₯, π) β SΜ : π β₯ π£β (π‘, π₯) = int D, can be replaced by Therefore, (c.f. Lemma 4.1) and we have where π£β is the lower semicontinuous envelope of π£ on [0, π ] × βπ , D β int D. If furthermore (π‘, π₯), π£ π£ π£, (4.6) is continuous and the inmum in (4.4) is attained for all then the converse inclusion is also satised. In applications, it may be desirable to have continuity of by and hence π£ so that int D and D are described directly (rather than a semicontinuous envelope). To analyze the continuity of one can study the comparison principle for the Hamilton-Jacobi-Bellman equation associated with the control problem (4.4). tions 5 and 6] for the explicit computation of π£ We refer to [8, Sec- in an example from Mathe- matical Finance. The rest of this subsection is devoted to the proof of Theorem 4.2. We rst state a version of Theorem 2.4 which is suitable to derive the PDE. Lemma 4.4. (i) Let π΅ be an open neighborhood of a point (π‘, π₯, π) β D such that π (π‘, π₯, π) < β and let π : π΅ β β be a continuous function such that π β€ π on π΅ . For all π > 0 there exist (π, πΌ) β π°π‘ × ππ‘ such that [ ] π πΌ πΌ π π (π‘, π₯, π) β€ πΈ π(π, ππ‘,π₯ (π ), ππ‘,π (π )) + π and ππ‘,π (π ) β₯ π(ππ‘,π₯ (π )), π (π ), π πΌ (π )) where π is the rst exit time of (π , ππ‘,π₯ π β₯π‘ from π΅ . π‘,π (ii) Let π΅ be an open neighborhood of a point (π‘, π₯, π) β int D such that π΅ β D and let (π, πΌ) β π°π‘ × ππ‘ . For any continuous function π : π΅ β β satisfying π β₯ π on π΅ and for any π > 0, [ ] π πΌ π (π‘, π₯, π + π) β₯ πΈ π(π, ππ‘,π₯ (π ), ππ‘,π (π )) , (4.7) π (π ), π πΌ (π )) where π is the rst exit time of (π , ππ‘,π₯ π β₯π‘ from π΅ . π‘,π Proof. In view of Lemmata 2.2 and 4.1, part (i) is immediate from Theo- rem 2.4(i). For part (ii) we use the extension of Theorem 2.4(ii') as stated in Corollary 2.7 with πβ©D=π π := π΅ . Note that is closed and hence π (π ), π πΌ (π )) β π΅ β D; in particular, (π, ππ‘,π₯ π‘,π₯ π -compact. We can now deduce the PDE for π from the dynamic programming principle in the form of Lemma 4.4. Although the arguments are the usual ones, we shall indicate the proof, in particular to show that the relaxation π + π in (4.7) does not aect the PDE. 21 Proof of Theorem 4.2. (i) We rst prove the subsolution property. Let π πΆ 1,2 -function and let (π‘0 , π₯0 , π0 ) β D be such that π‘0 β (0, π ) and (π‘0 , π₯0 , π0 ) is a maximum point of π β β π satisfying be a (π β β π)(π‘0 , π₯0 , π0 ) = 0. (4.8) Assume for contradiction that ( ) β βπ‘ π + π»β (β , π·π, π·2 π) (π‘0 , π₯0 , π0 ) > 0. D = int D by (4.6), (π‘0 , π₯0 , π0 ) such that Since of there exists a bounded open neighborhood ββπ‘ π¯ β πΏπ’,π (β , π·π, ¯ π·2 π) ¯ >0 on π΅ β© int D, for all π΅ β SΜ (π’, π) β π × βπ , (4.9) where ( ) π(π , ¯ π¦, π) := π(π‘0 , π₯0 , π0 ) + β£π β π‘0 β£2 + β£π¦ β π₯0 β£4 + β£π β π0 β£4 . Moreover, we have π := min(π¯ β π) > 0. (4.10) βπ΅ Given π > 0, let (π‘π , π₯π , ππ ) β π΅ β© int D be such that π (π‘π , π₯π , ππ ) β₯ π β (π‘0 , π₯0 , π0 ) β π. (4.11) (π, πΌ) β π°π‘ × ππ‘ such that ππ‘πΌπ ,ππ (π ) β₯ π(ππ‘ππ ,π₯π (π )) and π πΌ let π be the rst exit time of (π , ππ‘ ,π₯ (π ), ππ‘ ,π (π ))π β₯π‘ from π΅ . We recall π π π π π πΌ from Remark 2.3(ii) that (π , ππ‘ ,π₯ (π ), ππ‘ ,π (π ))π β₯π‘ remains in D on [π‘, π ], π π π π and hence also in int D by (4.6). Now, it follows from Itô's formula and (4.9) Consider arbitrary that [ ] π(π‘ ¯ π , π₯π , ππ ) β₯ πΈ π(π, ¯ ππ‘ππ ,π₯π (π ), ππ‘πΌπ ,ππ (π )) . For (π‘π , π₯π , ππ ) close enough to (π‘0 , π₯0 , π0 ), this implies that ] π(π‘ ¯ 0 , π₯0 , π0 ) β₯ πΈ π(π, ¯ ππ‘ππ ,π₯π (π ), ππ‘πΌπ ,ππ (π )) β π(1) [ as π β 0, which, by (4.8), (4.10) and (4.11), leads to [ ] π (π‘π , π₯π , ππ ) β₯ πΈ π(π, ππ‘ππ ,π₯π (π ), ππ‘πΌπ ,ππ (π )) + π β π(1). This contradicts Lemma 4.4(i) for π>0 small enough. (ii) We now prove the supersolution property. Let and let πβ β π (π‘0 , π₯0 , π0 ) β int D be such that (π‘0 , π₯0 , π0 ) π be a πΆ 1,2 -function is a minimum point of satisfying (πβ β π)(π‘0 , π₯0 , π0 ) = 0. Assume for contradiction that ( ) β βπ‘ π + π» β (β , π·π, π·2 π) (π‘0 , π₯0 , π0 ) < 0. 22 (4.12) Then there exist (π‘0 , π₯0 , π0 ) such (Λ π’, π Λ) β π × βπ and that π΅ β int D and a bounded open neighborhood ββπ‘ π¯ β πΏπ’Λ,Λπ (β , π·π, ¯ π·2 π) ¯ <0 on π΅, π΅ of (4.13) where ( ) π(π , ¯ π¦, π) := π(π‘0 , π₯0 , π0 ) β β£π β π‘0 β£2 + β£π¦ β π₯0 β£4 + β£π β π0 β£4 . Note that π := min(π β π) ¯ > 0. (4.14) βπ΅ Given π > 0, let (π‘π , π₯π , ππ ) β π΅ be such that π (π‘π , π₯π , ππ + π) β€ πβ (π‘0 , π₯0 , π0 ) + π. Viewing (Λ π’, π Λ) (4.15) as a constant control, it follows from Itô's formula and (4.13) that [ ] π(π‘ ¯ π , π₯π , ππ ) β€ πΈ π(π, ¯ ππ‘π’Λπ ,π₯π (π ), ππ‘πΛπ ,ππ (π )) , π is the rst exit time of (π , ππ‘π’Λπ ,π₯π (π ), ππ‘πΛπ ,ππ (π ))π β₯π‘ from π΅ . (π‘π , π₯π , ππ ) close enough to (π‘0 , π₯0 , π0 ), this implies that [ ] π(π‘ ¯ 0 , π₯0 , π0 ) β€ πΈ π(π, ¯ ππ‘π’Λπ ,π₯π (π ), ππ‘πΛπ ,ππ (π )) + π(1) as π β 0, where For which, by (4.12), (4.14) and (4.15), leads to [ ] π (π‘π , π₯π , ππ + π) β€ πΈ π(π, ππ‘π’Λπ ,π₯π (π ), ππ‘πΛπ ,ππ (π )) β π + π(1). π>0 For small enough, this yields a contradiction to Lemma 4.4(ii). 4.3 PDE for State Constraints In this section, we discuss the Hamilton-Jacobi-Bellman PDE for the state constraint problem (cf. Section 3) in the case where the state process is given by a controlled SDE as introduced in Section 4.1 and required to stay in an open set πͺ β βπ . Note that in this setup, the continuity conditions (3.2) and (3.3) are satised. We shall derive the PDE via Theorem 3.1. The basic idea to guaran- tee its condition (3.6) about right continuity in the constraint level runs as follows. Consider a control small probability πΏ β₯ 0. π β π°π‘ such that π ππ‘,π₯ state constraint, by switching to some admissible π ππ‘,π₯ exits πͺ. πͺ with at most π πΏ , satisfying the control π Λ shortly before leaves Then we shall construct a control As a result, ππΏ coincides with π on a set of large probability and therefore the reward is similar. Along the lines of Lemma 2.8 we shall then obtain the desired right continuity (cf. Lemma 4.7 below). To make this work, we clearly need to have [0, π ] × πͺ, ¯ π₯) β= β π°(π‘, for all which is anyway necessary for the value function 23 π¯ (π‘, π₯) in from (3.1) to be nite. However, we need a slightly stronger condition; namely, that we can switch to an admissible control in a measurable way. A particularly simple condition ensuring this, is the existence of an admissible feedback control: Assumption C. such that, for all Λ π(π ) = π₯+ β« π π’ Λ:πͺβπ There exists a Lipschitz continuous mapping (π‘, π₯) β [0, π ] × πͺ, the solution ( ) Λ Λ π π(π), π’ Λ(π(π)) ππ + π β« Λ π‘,π₯ π of ( ) Λ Λ π π(π), π’ Λ(π(π)) πππ , π β [π‘, π ] π‘ π‘ (4.16) satises Λ π‘,π₯ (π ) β πͺ π for all π β [π‘, π ], π -a.s. π(β , π’0 ) = 0 and π(β , π’0 ) = 0 for some π’0 β π , then Assumption C is clearly satised for π’ Λ β‘ π’0 . Or, under an additional smoothness condition, Λ π‘,π₯ will stay in πͺ if the Lipschitz function π’ π Λ satises ( ) 1 β€ β€ β£ππβ£(β , π’ Λ) = 0 and π π + Trace[π·π ππ ] (β , π’ Λ) > 0 on βπͺ, 2 If, e.g., where π denotes the inner normal to βπͺ; see also [11, Proposition 3.1] and [7, Lemma III.4.3]. The following is a simple condition guaranteeing the uniform integrability required in (2.31). Assumption D. Either π π(π₯, π’) π’. is bounded or the coecients in the SDE (4.1) have linear growth in π₯, uniformly in This assumption holds in particular if the control domain Remark 4.5. Assumption C implies that and Assumption D implies that π¯ π¯ π and π(π₯, π’) is bounded. is locally bounded from below is locally bounded from above. Next, we introduce the notation for the PDE related to the value function π¯ from (3.1). The associated Hamilton-Jacobi-Bellman operator is given by ¯ π»(π₯, π, π) := inf π’βπ ( ) ¯ π’ (π₯, π, π) , βπΏ (π₯, π, π) β βπ × βπ × ππ , where the Dynkin operator is dened by ¯ π’ (π₯, π, π) := π(π₯, π’)β€ π + 1 Tr[ππ β€ (π₯, π’)π], πΏ 2 π’ β π. Similarly as above, we introduce the semicontinuous envelopes ¯ β (π₯, π, π) := π» lim sup ¯ β² , πβ² , πβ² ), π»(π₯ (π₯β² , πβ² , πβ² ) β (π₯, π, π) (π₯β² , πβ² , πβ² ) β πͺ × βπ × ππ ¯ β (π₯, π, π) := π» lim inf (π₯β² , πβ² , πβ² ) β (π₯, π, π) (π₯β² , πβ² , πβ² ) β πͺ × βπ × ππ 24 ¯ β² , πβ² , πβ² ) π»(π₯ (4.17) and π¯ β (π‘, π₯) := π¯ (π‘β² , π₯β² ), lim sup (π‘β² , π₯β² ) β (π‘, π₯) (π‘β² , π₯β² ) β [0, π ) × πͺ π¯β (π‘, π₯) := lim inf (π‘β² , π₯β² ) β (π‘, π₯) (π‘β² , π₯β² ) β [0, π ) × πͺ π¯ (π‘β² , π₯β² ). We can now state the Hamilton-Jacobi-Bellman PDE; the boundary condition is discussed in Proposition 4.11 below. Theorem 4.6. Assume that π¯ is locally bounded on [0, π ) × πͺ. (i) The function π¯ β is a viscosity subsolution on [0, π ) × πͺ of ¯ β (β , π·π, π·2 π) β€ 0. ββπ‘ π + π» (4.18) (ii) Under Assumptions C and D, the function π¯β is a viscosity supersolution on [0, π ) × πͺ of ¯ β (β , π·π, π·2 π) β₯ 0. ββπ‘ π + π» The proof is given below, after some auxiliary results. the right-continuity condition (3.6) for the value function We rst verify π (π‘, π₯, π¦, π) intro- duced in Section 3. Lemma 4.7. Let Assumptions C and D hold true. Then π (π‘, π₯, 1, 0+) = π (π‘, π₯, 1, 0) for all (π‘, π₯) β [0, π ] × πͺ. Proof. For πΏ > 0, let π πΏ β π°(π‘, π₯, πΏ) be as in Lemma 2.8. Then the process π πΏ dened in (3.5) satises π π πΏ (π ) > 0 outside of a set of measure at ππ‘,π₯,π¦ π‘,π₯,1 most πΏ . It follows that we can nd ππΏ β (0, 1 β§ π(π₯)) such that the set π (π ) β€ π } satises π [π΄πΏ ] β€ 2πΏ . Let π πΏ denote the rst time π΄πΏ := {ππ‘,π₯,1 πΏ ππΏ when ππ‘,π₯,1 reaches the level ππΏ and set Λ πΏ ), Λ(π πΛπΏ := π πΏ 1[π‘,π πΏ ] + 1(π πΏ ,π ] π’ Λ πΏ is the solution of (4.16) on [π πΏ , π ] with initial condition given π π πΏ (π πΏ ). Since the paths of π π πΏ are nonincreasing, we have Λ πΏ (π πΏ ) = ππ‘,π₯ π π‘,π₯,1 where πΏ by πΏ πΛ π lim π [ππ‘,π₯ (π ) β= ππ‘,π₯ (π )] β€ lim π [π πΏ β€ π ] = lim π [π΄πΏ ] = 0. πΏβ0 Next, we check that trivial if π πΏβ0 π (π )), π β π°} {π (ππ‘,π₯ πΏβ0 is uniformly integrable. This is is bounded. Otherwise, Assumption D yields that the coecients π(π₯, π’) and π(π₯, π’) have uniformly linear growth in π₯, and of course they are π uniformly Lipschitz in π₯ as they are jointly Lipschitz. Thus {ππ‘,π₯ (π ), π β π°} π is bounded in πΏ for any nite π and the uniform integrability follows from the quadratic growth assumption (4.2) on π . It remains to apply Lemma 2.8. 25 Lemma 4.8. Let Assumption C hold true and let π΅ be an open neighborhood of (π‘, π₯) β [0, π ] × πͺ. For all π β π°π‘ there exists π¯ β π°π‘ such that ¯ π₯), π1[π‘,π ] + π¯1(π,π ] β π°(π‘, π (π )) where π is the rst exit time of (π , ππ‘,π₯ π β₯π‘ from π΅ . Proof. Let Λ π,π π (π ) π π‘,π₯ be the solution of (4.16) on π (π ) at time π . ππ‘,π₯ Λ π,π π (π ) ). π¯ := π1[π‘,π ] + 1(π,π ] π’ Λ(π π‘,π₯ integrable) initial condition [π, π ] with the (square Then the claim holds true for We have the following counterpart of Lemma 4.4. Lemma 4.9. (i) Let π΅ β [0, π ] × βπ be an open neighborhood of a point (π‘, π₯) β [0, π ]×πͺ such that π¯ (π‘, π₯) is nite and let π : π΅ β β be a continuous ¯ π₯) such function such that π¯ β€ π on π΅ . For all π > 0 there exists π β π°(π‘, that [ ] π π¯ (π‘, π₯) β€ πΈ π(π, ππ‘,π₯ (π )) + π, π (π )) where π is the rst exit time of (π , ππ‘,π₯ π β₯π‘ from π΅ . (ii) Let Assumptions C and D hold true and let π΅ β [0, π ] × πͺ be an open neighborhood of (π‘, π₯). For any π β π°π‘ and any continuous function π : π΅ β β satisfying π¯ β₯ π on π΅ , [ ] π π¯ (π‘, π₯) β₯ πΈ π(π, ππ‘,π₯ (π )) , π (π )) where π is the rst exit time of (π , ππ‘,π₯ π β₯π‘ from π΅ . Proof. In view of Lemma 4.1, part (i) is immediate from Theorem 3.1(i). Part (ii) follows from Theorem 3.1(ii) via Lemmata 4.7 and 4.8 . Proof of Theorem 4.6. The result follows from Lemma 4.9 by the arguments in the proof of Theorem 4.2. 4.3.1 Boundary Condition and Uniqueness In this section, we discuss the boundary condition and the uniqueness for the PDE in Theorem 4.6. We shall work under a slightly stronger condition on our setup. Assumption D'. linear growth in The coecients π₯, uniformly in π(π₯, π’) and π(π₯, π’) in the SDE (4.1) have π’. We also introduce the following regularity condition, which will be used to prove the comparison theorem. Denition 4.10. Then π€ is Consider a set of class β(πͺ) πͺ β βπ if the following hold for any 26 π€ : [0, π ] × πͺ β β. (π‘, π₯) β [0, π ) × βπͺ: and a function (i) There exist β : β+ β π > 0, an open neighborhood π΅ of π₯ in βπ and a function βπ such that lim inf πβ1 β£β(π)β£ < β πβ0 π¦ + β(π) + π(π) β πͺ (ii) There exists a function and for all (4.19) π¦ βπͺβ©π΅ π : β+ β β+ and π β (0, π). (4.20) such that lim π(π) = 0 and ( ) lim π€ π‘ + π(π), π₯ + β(π) = π€(π‘, π₯). (4.21) πβ0 (4.22) πβ0 π : β+ β βπ is any function of class π(π), then there exists π > 0 such that π¦ + β(π) + π(π) β πͺ for all π β (0, π). Note that (i) is a condition on the boundary of βπͺ ; it can be seen as a By (4.20) we mean that if variant of the interior cone condition where the cone is replaced by a more general shape. Condition (ii) essentially states that π€ is continuous along at least one curve approaching the boundary point through this shape. In Proposition 4.12 below, we indicate a sucient condition for β(πͺ), π¯β to be of class which is stated directly in terms of the given primitives. We shall see in its proof that Denition 4.10 is well adapted to the problem at hand (see also Remark A.4 below). Before that, let us state the uniqueness result. Proposition 4.11. Let π be continuous and let Assumptions C and D' hold true. Then π¯ has quadratic growth and the boundary condition is attained in the sense that π¯ β (π, β ) β€ π and π¯β (π, β ) β₯ π on πͺ. Assume in addition that π¯β is of class β(πͺ). Then (i) π¯ is continuous on [0, π ] × πͺ and admits a continuous extension to [0, π ] × πͺ, (ii) π¯ is the unique (discontinuous) viscosity solution of the state constraint problem ¯ π·π, π·2 π) = 0, π(π, β ) = π ββπ‘ π + π»(β , in the class of functions having polynomial growth and having a lower semicontinuous envelope of class β(πͺ). Proof. Recalling that π has quadratic growth (4.2), it follows from stan- dard estimates for the SDE (4.1) under Assumptions C and D' that π¯ has quadratic growth and satises the boundary condition. Assumption D' implies Assumption D and hence Theorem 4.6 yields that π¯ β and π¯β are sub- and supersolutions, respectively. sumption D' implies that ¯ π» Moreover, As- is continuous and satises Assumption E in 27 Appendix A; see Lemma A.3 below for the proof. If π¯β is of class β(πͺ), the comparison principle (Theorem A.1 and the subsequent remark, cf. Appendix A) yields that π¯ β = π¯β [0, π ]×πͺ; in particular, (i) holds. on Part (ii) also follows from the comparison result. We conclude this section with a sucient condition ensuring that is of class β(πͺ); π¯β the idea is that the volatility should degenerate so that the state process can be pushed away from the boundary. We remark that conditions in a similar spirit exist in the previous literature (e.g., [11, 10]); cf. Remark A.4 below. Proposition 4.12. Assume that π¯β is nite-valued on [0, π ] × πͺ and that πͺ, π and π satisfy the following conditions: (i) There exists a πΆ 1 -function πΏ , dened on a neighborhood of πͺ β βπ , such that π·πΏ is locally Lipschitz continuous and πΏ > 0 on πͺ, πΏ = 0 on βπͺ, πΏ < 0 outside πͺ. (ii) There exists a locally Lipschitz continuous mapping π’ Λ : βπ β π such that for all π₯ β πͺ there exist an open neighborhood π΅ of π₯ and π > 0 satisfying π(π§, π’ Λ(π§))β€ π·πΏ(π¦) β₯ π and π(π¦, π’ Λ(π¦)) = 0 for all π¦ β π΅β©πͺ and π§ β π΅. (4.23) Then π¯β is of class β(πͺ). Proof. Fix (π‘, π₯) β [0, π ) × βπͺ and let πΏ , π and π΅ π΅ is bounded. Consider π¦ β πͺ β© π΅ . assume that and be as above; we may Since π₯β² 7β π(π₯β² , π’ Λ(π₯β² )) is locally Lipschitz, the ordinary dierential equation β« π₯ Λ(π ) = π¦ + π ( ) π π₯ Λ(π), π’ Λ(Λ π₯(π)) ππ 0 has a unique solution π₯ Λπ¦ (π ) β πͺ For for π β [0, ππ₯ ), π₯ Λπ¦ on some interval π β (0, π], for π>0 [0, ππ¦ ). Then (4.23) ensures that small enough, for all π¦ β πͺ β© π΅. (4.24) we set β(π) := π₯ Λπ₯ (π) β π₯, π(π) := π. Then (4.21) is clearly satised. Using the continuity of π and π’ Λ, we see that πβ1 β£β(π)β£ β π(π₯, π’ Λ(π₯)), which implies (4.19). Moreover, using that π and π’ Λ are locally Lipschitz, we nd that π₯ Λπ₯ (π) β π₯ Λπ¦ (π) = π₯ β π¦ + π(π). 28 Together with (4.23), (4.24) and the local Lipschitz continuity of π, π’ Λ and π·πΏ , this implies that for π¦ β πͺ suciently close to π₯ and π > 0 small enough, ( ) πΏ π¦ + β(π) + π(π) ( ) =πΏ π¦βπ₯+π₯ Λπ₯ (π) + π(π) β« π ( )β€ ( ) = πΏ(π¦) + π π₯ Λπ₯ (π), π’ Λ(Λ π₯π₯ (π)) π·πΏ π¦ β π₯ + π₯ Λπ₯ (π) ππ + π(π) β« 0π ( )β€ ( ) Λπ¦ (π), π’ Λ(π₯ β π¦ + π₯ Λπ¦ (π)) π·πΏ π₯ Λπ¦ (π) ππ + π(π2 ) + π(π) = πΏ(π¦) + π π₯ β π¦ + π₯ 0 β₯ ππ + π(π), π > 0 small enough. This implies (4.20). (π , π¦) β [0, π ) × πͺ close to (π‘, π₯). For π > 0 small enough, ¯ + π(π), π₯ π π β π°(π Λπ¦ (π)) such that [ ( ππ )] ( ) πΈ π ππ +π(π),Λπ₯π¦ (π) (π ) β₯ π¯ π + π(π), π₯ Λπ¦ (π) β π. which is strictly positive for Consider can nd we Recall the degeneracy condition in (4.23). Setting π¯π := π’ Λ(Λ π₯π¦ )1[π ,π +π(π)] + 1(π +π(π),π ] π π , we obtain that [ ( π¯π )] π¯ (π , π¦) β₯ πΈ π ππ ,π¦ (π ) [ ( ππ )] = πΈ π ππ +π(π),Λ π₯π¦ (π) (π ) ( ) β₯ π¯ π + π(π), π₯ Λπ¦ (π) β π. π₯ Λπ¦ (π) β π₯ Λπ₯ (π) as π¦ β π₯, this leads to ( ) ( ) π¯β (π‘, π₯) β₯ π¯β π‘ + π(π), π₯ Λπ₯ (π) β π = π¯β π‘ + π(π), π₯ + β(π) β π Recalling that for π>0 small enough, which implies ( ) π¯β (π‘, π₯) β₯ lim sup π¯β π‘ + π(π), π₯ + β(π) πβ0 ( ) β₯ lim inf π¯β π‘ + π(π), π₯ + β(π) πβ0 Hence β₯ π¯β (π‘, π₯). ( ) limπβ0 π¯β π‘ + π(π), π₯ + β(π) = π¯β (π‘, π₯); i.e., (4.22) holds for π¯β . 4.3.2 On Closed State Constraints Recall that the value function π¯ considered above corresponds to the con- straint that the state processes remains in the open set πͺ. We can similarly consider the closed constraint; i.e., { π π (π ))] : π β π°π‘ , ππ‘,π₯ (π ) β πͺ π (π‘, π₯) := sup πΈ[π (ππ‘,π₯ 29 for all π β [π‘, π ], π -a.s. } set, the constraint function π¯ do not apply to π . Indeed, for the closed πΊ in Section 3 would not be u.s.c. and hence the The arguments used above for derivation of Theorem 3.1 fails; note that the upper semicontinuity is essential for the covering argument in the proof of Theorem 2.4(ii),(ii'). Moreover, the switching argument in the proof of Lemma 4.8 cannot be imitated since, given that the state process π ππ‘,π₯ βπͺ, it exit πͺ . hits the boundary know which trajectories of the state will actually is not possible to However, we shall see that, if a comparison principle holds, then the πͺ dynamic programming principle for the open constraint characterize the value function apply the PDE for π¯ π associated to πͺ. is enough to fully More precisely, we shall and its comparison principle to deduce that π = π¯ under certain conditions. Of course, the basic observation that this equality holds under suitable conditions is not new; see, e.g., [10]. We shall use the following assumption. Assumption C'. Assumption C holds with π’Λ dened on πͺ. Corollary 4.13. Let π be continuous, let Assumptions C' and D' hold true and assume that π¯β is of class β(πͺ). Then π = π¯ on [0, π ] × πͺ. We recall that π¯ admits a (unique) continuous extension to [0, π ] × πͺ under the stated conditions, so the assertion makes sense. Proof. The easier direction of the dynamic programming principle for π can be obtained as above, so the result of Lemma 4.9(i) still holds. It then follows by the same arguments as in the proof of Theorem 4.6 and Proposition 4.11 that the upper semicontinuous envelope satisfying β π (π, β ) β€ π on πͺ. π β is a viscosity subsolution of (4.18) As in the proof of Proposition 4.11, we can β π β€ π¯β on [0, π ] × πͺ by the apply the comparison principle of Theorem A.1 to deduce that [0, π ] × πͺ. On the other hand, we clearly have π¯ β€ π on denitions of these value functions. Therefore, we have β π¯β β€ π β β€ π β€ π¯β = π¯ β , where πβ denotes the lower semicontinuous envelope of π and the last equal- ity is due to Proposition 4.11. It follows that all these functions coincide. A Comparison for State Constraint Problems In this appendix we provide, by adapting the usual techniques, a fairly general comparison theorem for state constraint problems which is suitable for the applications in Proposition 4.11 and Corollary 4.13. In the following, β denotes a continuous mapping from to β which is nonincreasing in its third variable, πͺ βπ , and π > 0 is a xed constant. We consider the is a given open subset of equation ππ β βπ‘ π + β(β , π·π, π·2 π) = 0 30 βπ × βπ × ππ (A.1) and the following condition on Assumption E. β. πΌ > 0 such that ) lim inf β(π¦, π, π π ) β β(π₯, π, π π ) πβ0 ( ) ( ) β€ πΌ β£π₯ β π¦β£ 1 + β£πβ£ + π2 β£π₯ β π¦β£ + (1 + β£π₯β£)β£π β πβ£ + (1 + β£π₯β£2 )β£πβ£ There exists ( for all (π₯, π¦) β πͺ with β£π₯ β π¦β£ β€ 1 and for all (π, π, π) β βπ × βπ × π2π , β ππ × ππ and π β₯ 1 such that ( π ) π 0 β€ π΄π + ππ΄2π for all π > 0, 0 βπ π (π π , π π )π>0 where π΄π := π 2 ( πΌπ βπΌπ βπΌπ πΌπ ) + π. Theorem A.1. Let Assumption E hold true. Let π€1 be an u.s.c. viscosity subsolution on πͺ and let π€2 be an l.s.c. viscosity supersolution on πͺ of (A.1). If π€1 and π€2 have polynomial growth on πͺ and if π€2 is of class β(πͺ), then π€2 β₯ π€1 on {π } × πͺ Remark A.2. implies π€2 β₯ π€1 on [0, π ] × πͺ. Our result also applies to the equation ββπ‘ π + β(β , π·π, π·2 π) = 0, provided that (A.2) β is homogeneous of degree one with respect to its second and third argument, as it is the case for the Hamilton-Jacobi-Bellman operators in the body of this paper. Indeed, only if (π‘, π₯) 7β πππ‘ π€ π€1 is then a subsolution of (A.2) if and 1 (π‘, π₯) is a subsolution of (A.1), and similarly for the supersolution. Further extensions could also be considered but are beyond the scope of this paper. Proof of Theorem A.1. Assume that π€2 β₯ π€1 on {π } × πͺ. Let π β₯ 1 and πΆ > 0 be such that π€1 (π‘, π₯) β π€2 (π‘, π₯) β€ πΆ(1 + β£π₯β£π ) for all (π‘, π₯) β [0, π ] × πͺ. Assume for contradiction that sup(π€1 β π€2 ) > 0, then we can nd π > 0 and (π‘0 , π₯0 ) β [0, π ] × πͺ such that π := (π€1 β 2π β π€2 )(π‘0 , π₯0 ) = max (π€1 β 2π β π€2 ) > 0, (A.3) [0,π ]×πͺ where π(π‘, π₯) := ππβπ π‘ (1 + β£π₯β£2π ). Here π >0 is a xed constant which is large enough to ensure that π(π‘, π₯) := (A.4) 2 ( 2 ) β ππ(π‘, π₯) + βπ‘ π(π‘, π₯) + πΌ (1 + β£π₯β£)β£π·π(π‘, π₯)β£ + (1 + β£π₯β£ )β£π· π(π‘, π₯)β£ 31 [0, π ] × βπ , is nonpositive on where πΌ Note that by the assumption that (π‘0 , π₯0 ) β [0, π ) × πͺ. Case 1: π₯0 β βπͺ. For all is the constant from Assumption E. π€2 β₯ π€1 on {π } × πͺ, we must have π β₯ 1, there exist (π‘π , π₯π , π π , π¦ π ) β ([0, π ]×πͺ)2 satisfying Ξ¦π (π‘π , π₯π , π π , π¦ π ) = Ξ¦π , max (A.5) ([0,π ]×πͺ)2 where Ξ¦π (π‘, π₯, π , π¦) := π€1 (π‘, π₯) β π€2 (π , π¦) β Ξπ (π‘, π₯, π , π¦) and ) 1 ( Ξπ (π‘, π₯, π , π¦) := π2 β£π‘ + π(πβ1 ) β π β£2 + πβ£π₯ + β(πβ1 ) β π¦β£2 2 + β£π‘ β π‘0 β£2 + β£π₯ β π₯0 β£4 + π(π‘, π₯) + π(π , π¦), β π > 0. with π and given for π₯0 as in the statement of the Denition 4.10, and Note that (A.5), the assumption that π€2 satises (4.22), and (A.3) imply that ( ) Ξ¦π (π‘π , π₯π , π π , π¦ π ) β₯ Ξ¦π π‘0 , π₯0 , π‘0 + π(πβ1 ), π₯0 + β(πβ1 ) = (π€1 β 2π β π€2 )(π‘0 , π₯0 ) + π(1) = π + π(1) as π β β. (A.6) π€1 , π€2 and the denition of π, it folπ π π π lows that, after passing to a subsequence, (π‘ , π₯ , π , π¦ ) converges to some β β β β 2 (π‘ , π₯ , π‘ , π₯ ) β ([0, π ] × πͺ) . We then have Recalling the growth condition on π = (π€1 β 2π β π€2 )(π‘0 , π₯0 ) = max (π€1 β 2π β π€2 ) [0,π ]×πͺ β₯ (π€1 β 2π β π€2 )(π‘β , π₯β ) β β£π‘β β π‘0 β£2 β β£π₯β β π₯0 β£4 ) 1 ( β lim sup π2 β£π‘π + π(πβ1 ) β π π β£2 + πβ£π₯π + β(πβ1 ) β π¦ π β£2 πββ 2 β₯ lim inf Ξ¦π (π‘π , π₯π , π π , π¦ π ) πββ β₯ π, where (A.6) was used in the last step. After passing to a subsequence, we deduce that (π‘π , π₯π , π π , π¦ π ) β (π‘0 , π₯0 , π‘0 , π₯0 ), (A.7) π€1 (π‘π , π₯π ) β π€2 (π π , π¦ π ) β (π€1 β π€2 )(π‘0 , π₯0 ), π π Since = π‘π + π(πβ1 ) + π(πβ1 ), π¦π = π₯π (π‘0 , π₯0 ) β [0, π ) × βπͺ, it follows from β [0, π ) × πͺ for π large enough. (π π , π¦ π ) 32 + β(πβ1 ) + π(πβ1 ). (A.8) (A.9) (4.20), (4.21) and (A.9) that Let 2,+ π« πͺ π€1 2,β π« πͺ π€2 and be the closed parabolic super- and subjets as dened in [5, Section 8]. From the Crandall-Ishii lemma [5, Theorem 8.3] we obtain, for each π > 0, elements 2,+ (ππ , ππ , πππ ) β π« πͺ π€1 (π‘π , π₯π ) and 2,β (ππ , π π , πππ ) β π« πͺ π€2 (π π , π¦ π ) such that ππ = βπ‘ Ξπ (π‘π , π₯π , π π , π¦ π ), ππ = ββπ Ξπ (π‘π , π₯π , π π , π¦ π ), ππ = π·π₯ Ξπ (π‘π , π₯π , π π , π¦ π ), π π = βπ·π¦ Ξπ (π‘π , π₯π , π π , π¦ π ), ( πππ 0 0 βπππ ) β€ π΄π + ππ΄2π , π΄π = π·2 Ξπ (π‘π , π₯π , π π , π¦ π ); i.e., ) ( ) ( 2 π π πΌπ βπΌπ π· π(π‘ , π₯ ) + π(β£π₯π β π₯0 β£2 ) 0 . π΄π = ππ2 + βπΌπ πΌπ 0 π·2 π(π π , π¦ π ) where In view of the sub- and supersolution properties of (π π , π¦ π ) β [0, π ) × πͺ for π π€1 and π€2 , the fact that large, and Assumption E, we deduce that Ξπ := π(π€1 (π‘π , π₯π ) β π€2 (π π , π¦ π )) β€ 2(π‘π β π‘0 ) + βπ‘ π(π‘π , π₯π ) + βπ π(π π , π¦ π ) ( ) + πΌβ£π₯π β π¦ π β£ 1 + β£π π β£ + ππ2 β£π₯π β π¦ π β£ ( ) + πΌ (1 + β£π₯π β£)β£π π β ππ β£ + (1 + β£π₯π β£2 )β£ππ β£ , where π π := ( π·2 π(π‘π , π₯π ) + π(β£π₯π β π₯0 β£2 ) 0 0 π·2 π(π π , π¦ π ) By the denitions of ππ and ππ, ) . it follows that Ξπ β€ 2(π‘π β π‘0 ) + βπ‘ π(π‘π , π₯π ) + βπ π(π π , π¦ π ) ( ) + πΌβ£π₯π β π¦ π β£ 1 + β£π·π(π π , π¦ π )β£ + ππ2 β£π₯π + β(πβ1 ) β π¦ π β£ + ππ2 β£π₯π β π¦ π β£ ( ) + πΌ(1 + β£π₯π β£) 4β£π₯π β π₯0 β£3 + β£π·π(π‘π , π₯π )β£ + β£π·π(π π , π¦ π )β£ ( ) + πΌ(1 + β£π₯π β£2 ) β£π·2 π(π‘π , π₯π )β£ + β£π·2 π(π π , π¦ π )β£ + π(β£π₯π β π₯0 β£2 ) . Recalling (A.7)(A.9), letting πββ leads to ( )2 π(π€1 β π€2 )(π‘0 , π₯0 ) β€ 2βπ‘ π(π‘0 , π₯0 ) + πΌπ lim inf πβ(πβ1 ) πββ ( ) + 2πΌ (1 + β£π₯0 β£)β£π·π(π‘0 , π₯0 )β£ + (1 + β£π₯0 β£2 )β£π·2 π(π‘0 , π₯0 )β£ , 33 which, by (4.19) and the denition of π in (A.4), implies π(π€1 β 2π β π€2 )(π‘0 , π₯0 ) β€ 2π(π‘0 , π₯0 ) after letting π β 0. Since π > 0 has been chosen so that π β€ 0 on [0, π ]×βπ , this contradicts (A.3). Case 2: π₯0 β πͺ. This case is handled similarly by using ) 1 ( Ξπ (π‘, π₯, π , π¦) := π2 β£π‘ β π β£2 + β£π₯ β π¦β£2 + β£π‘ β π‘0 β£2 + β£π₯ β π₯0 β£4 + π(π‘, π₯) + π(π , π¦). 2 After taking a subsequence, the corresponding sequence of maximum points (π‘π , π₯π , π π , π¦ π )πβ₯1 again converges π large enough. The rest of the to (π‘0 , π₯0 , π‘0 , π₯0 ), so that π₯π , π¦ π β πͺ for proof follows the same arguments as in Case 1. Lemma A.3. ¯ dened in Under Assumption D', the operator π» ises Assumption E. Proof. (4.17) sat- The argument is standard. We rst observe that ( ) ¯ π’ (π¦, π, π π ) + πΏ ¯ π’ (π₯, π, π π ) = π(π₯, π’) β π(π¦, π’) β€ π + π(π₯, π’)β€ (π β π) βπΏ 1 β π + Ξ£ (π₯, π¦, π’)β€ Ξπ Ξ£π (π₯, π¦, π’), 2 1β€πβ€π where ( Ξ£(π₯, π¦, π’) := and Ξ£π π(π₯, π’) π(π¦, π’) denotes the πth column of ) and Ξ£. Since π ( Ξ := ππ 0 0 βπ π ) π is Lipschitz continuous and has uniformly linear growth by Assumption D', we have ( )β€ ( ) π(π₯, π’) β π(π¦, π’) π + π(π₯, π’)β€ (π β π) β€ πΌ β£π₯ β π¦β£β£πβ£ + (1 + β£π₯β£)β£π β πβ£ for some constant πΌ > 0. Recall from the statement of Assumption E the condition that π Ξ β€ π΄π + Since π ππ΄2π , where π΄π := π 2 ( πΌπ βπΌπ βπΌπ πΌπ ) + π. is Lipschitz continuous and satises Assumption D', it follows that Ξ£π (π₯, π¦, π’)β€ Ξπ Ξ£π (π₯, π¦, π’) β€ π2 β£π π (π₯, π’) β π π (π¦, π’)β£2 ( ) + πΌ 1 + β£π₯β£2 + β£π¦β£2 β£πβ£ + π(π), πΌ, where π π denotes the πth column of π . We conclude by using the Lipschitz continuity of π and the condition that β£π₯ β π¦β£ β€ 1. possibly after enlarging 34 Remark A.4. We conclude with some comments on Theorem A.1 and Proposition 4.12, and related results in the literature. The main issue in proving comparison with state constraints is to avoid boundary points; i.e., that π¦π (see the proof of Theorem A.1) ends up on the boundary. One classical way to ensure this, is to use a perturbation of β£π₯ β π¦β£2 in a suitable inward direction, like the function β above. Moreover, this requires the supersolution to be continuous at the boundary points, along the direction of perturbation; cf. (4.22). In [11], the inner normal π(π₯) π₯ β βπͺ is used as π₯ (cf. (A.9)); therefore, at the boundary point an inward direction. In the proof, one is only close to the comparison result [11, Theorem 2.2] requires the existence of a truncated cone around π(π₯) which stays inside the domain, and the continuity of the subsolution along the directions that it generates. Our condition (4.20) is less restrictive than the corresponding requirement in [11]: we only need π β 7 β(π), cf. (4.22), rather than all lines function π appears because we consider parabolic the continuity along the curve in a neighborhood. The equations, whereas [11] focuses on the elliptic case. In Proposition 4.12 we give conditions (certainly not the most general possible) ensuring that the value function is of class β(πͺ). They should be compared to [11, Condition (A3)], which is used to verify the continuity assumption of [11, Theorem 2.2]. Our conditions are stronger in the sense that they are imposed around the boundary and not only at the boundary; on the other hand, we require πΆ 1 -regularity of the boundary whereas [11] 3 requires πΆ . In [10], a slightly dierent technique is used, based on ideas from [9]. First, it is assumed that at each boundary point π₯, there exists a xed control which kills the volatility at the neighboring boundary points and keeps a truncated cone around the drift in the domain. This is similar to our (4.23), except that our control is not xed; on the other hand, we assume that it π₯. Thus, the conditions in [10] are 2 not directly comparable to ours; e.g., if πͺ is the unit disk in β , π = [β1, 1], 2 π(π₯, π’) = βπ₯ and π(π₯, π’) = β£π₯ β π’β£πΌ2 , then (4.23) is satised (with πΏ given by the Euclidean distance near the boundary and π’ Λ(π₯) = π₯2 ), while [10, kills the volatility in a neighborhood of Condition (2.1)] is not. Second, in [10], the state constraint problem is transformed so as to introduce a Neumann-type boundary condition and construct a suitable test function which, as a function of its rst component, turns out to be a uniformly strict supersolution of the Neumann boundary condition. The construction in [10] heavily relies on the assumption that the coecients are bounded; cf. the beginning of [10, Section 3] and the proof of [10, Theorem 3.1]. 35 References [1] B. Bouchard, R. Elie, and C. Imbert. Optimal control under stochastic target constraints. SIAM J. Control Optim., 48(5):35013531, 2010. [2] B. Bouchard, R. Elie, and N. Touzi. Stochastic target problems with controlled loss. SIAM J. Control Optim., 48(5):31233150, 2009. [3] B. Bouchard and N. Touzi. Weak dynamic programming principle for viscosity solutions. SIAM J. Control Optim., 49(3):948962, 2011. [4] B. Bouchard and T.N. Vu. A stochastic target approach for P&L matching problems. Mathematics of Operation Research, 37(3):526558, 2012. [5] M. Crandall, H. Ishii, and P.-L. Lions. User's guide to viscosity solutions of second order partial dierential equations. Bull. Amer. Math. Soc., 27(1):167, 1992. [6] W. H. Fleming and H. M. Soner. Controlled Markov Processes and Viscosity Solutions. Springer, New York, 2nd edition, 2006. [7] M. I. Freidlin. Functional Integration and Partial Dierential Equations. Princeton University Press, 1985. [8] H. Föllmer and P. Leukert. Ecient hedging: cost versus shortfall risk. Finance and Stochastics, 4(2):117146, 2000. [9] H. Ishii and S. Koike. A new formulation of state constraint problems for rst-order PDEs. SIAM J. Control Optim., 34(2):554571, 1996. [10] H. Ishii and P. Loreti. A class of stochastic optimal control problems with state constraint. Indiana Univ. Math. J., 51(5):11671196, 2002. [11] M. A. Katsoulakis. Viscosity solutions of second order fully nonlinear elliptic equations with state constraints. Indiana Univ. Math. J., 43(2):493519, 1994. [12] J.-M. Lasry and P.-L. Lions. Nonlinear elliptic equations with singular boundary conditions and stochastic control with state constraints. I. The model problem. Math. Ann., 283(4):583630, 1989. [13] H. M. Soner. Optimal control with state-space constraint. I. SIAM J. Control Optim., 24(3):552561, 1986. [14] H. M. Soner. Optimal control with state-space constraint. II. SIAM J. Control Optim., 24(6):11101122, 1986. [15] H. M. Soner and N. Touzi. Dynamic programming for stochastic target problems and geometric ows. J. Eur. Math. Soc. (JEMS), 4(3):201236, 2002. [16] P. Soravia. Optimality principles and representation formulas for viscosity solutions of Hamilton-Jacobi equations. I. Equations of unbounded and degenerate control problems without uniqueness. Adv. Dierential Equations, 4(2):275296, 1999. [17] P. Soravia. Optimality principles and representation formulas for viscosity solutions of Hamilton-Jacobi equations. II. Equations of control problems with state constraints. Dierential Integral Equations, 12(2):275293, 1999. [18] J. Yong and X. Y. Zhou. Stochastic Controls. Hamiltonian Systems and HJB Equations. Springer, New York, 1999. 36
© Copyright 2026 Paperzz