• Keine Ergebnisse gefunden

1.2 Notation

1.2.2 Approximation of stochastic processes

φ(XπtN) +

N−1

j=i

f(tj,Sπtj,Ybtn−1,πj ,Zbtn−1,πj )∆j

¸ , Zbd,tn,π

i = Pd,i

·∆Wd,i

i µ

φ(XπtN) +

N−1

j=i+1

f(tj,Sπtj,Ybtn−1,πj ,Zbn−1,πtj )∆j

¶¸

, d=1, . . . ,D,

initialized again at(Yb0,π,Zb0,π) = (0, 0).

At this stage the advantage of the forward approximation scheme reveals: Theorem 11 of Bender and Denk [2] specifies the moderate error occurring when approximating(Ytn,πi ,Zn,πti )with(Ybtn,πi ,Zbtn,πi ). In the forward scheme this error is bounded by a constant times the worst projection error occurring during the iterations. Consequently, it does not explode if the mesh grid size tends to zero as it is the case for the backward schemes. For more details, see the discussion in Bender and Denk [2], pp. 1802-1803.

In a final step Bender and Denk [2] replace the theoretical projections Pd,i by simulation based least-squares estimators and derive at last in their Theorem 15 that this estimator convergesP-almost surely to the approximation coming from the theoretical projection. Overall, they obtain convergence in probability for their final estimator towards the solution of the FBSDE.

1.2 Notation

1.2.1 Function spaces

As usual in the theory of BSDEs we deal with the following function spaces:

L2(F)- the space ofF-measurable random variablesXsuch thatE h

|X|2 i

<∞,

L2F³

Ω,C([0,T]),Rd´

- the space of(Ft)t≥0-adaptedRd-valued continuous processes X such that Eh

supt∈[0,T]|Xt|2i

<∞,

L2

³ 0,T;Rd

´

- the space of(Ft)t≥0-adaptedRd-valued continuous processes Xsuch that we have EhRT

0 |Xt|2dt i

<∞and

• M[0,T]:=L2F(Ω,C([0,T]),Rn)×L2(0,T;Rn)equipped with the norm

||(Y(·),Z(·))||M[0,T]:=

µ E

· sup

t∈[0,T]

|Yt|2

¸ +E

· Z T

0

|Zt|2dt

¸¶1/2 .

Pardoux and Peng [41] showed that inM[0,T]there is a unique solution to BSDEs satisfying the above assumptions.

1.2.2 Approximation of stochastic processes

In order to get the intuition behind the notation of the different discretizations of the occurring stochastic processes we introduce them in the following in full detail.

Time discretization

In any approximation scheme we will consider, the first stage of approximation is with respect to time.

That is we introduce a fixed partitionπ : 0 = t0 < . . . < tN = T of the interval [0,T] and compute

4 1.2. Notation

approximations of the solution of the FBSDE at the partition pointsti,i = 0, . . . ,N. For the solution of the backward part we furthermore use an iterative Picard-type approach and label these iterations with n∈N0. Hence, writing(Stπi,Ytn,πi ,Ztn,πi )indicates the time discretized solution of the FSBDE at timetiand iterationngiven the partitionπ. Proceeding this way, we have to introduce discretizations of increments of the Brownian motion with respect toπdenoted by∆Wi =Wti+1−Wti, i.e. we use forward increments.

Chapter 2 introduces a family of FSBDEs which is parameterized by a further stochastic processhwhich is again chosen once and then fixed throughout the whole calculations. We thus write(Sh,πti ,Yth,n,πi ,Zth,n,πi ) for the time discretized solution of the modified FSBDE at timetiand iterationngiven the partitionπ.

The choice h 0 thereby corresponds to the original discretization of Bender and Denk [2]. As our parametrization represents a change of measure, we also have to consider Brownian increments under a further measure and denote them by∆Wih=Wthi+1−Wthito distinguish them from the former.

In order to ease notation at a later stage we drop the superindiceshandπfor the time discrete solution of the FBSDE, i.e. instead of(Sth,πi ,Yth,n,πi ,Zh,n,πti )we simply write(Sti,Ytni,Ztni). We can justify this impre-ciseness not only because of the fewer indices but also because we do not change in the following steps the partition and the processhand hold them fixed.

Another variant of the equation withh 0 is studied in Chapter 3. Here we focus on drivers f, which are bounded by some constantR. As consequence, we will derive that under mild manipulations of the scheme of Bender and Denk our time discrete approximations of the solution of the backward part are bounded. To remind the reader of this property we write(Sti,Ytn,Ri ,Zn,Rti )suppressing the dependency on the time partitionπ.

In any setting, there will be Borel-functions such that the time discrete approximations of the solution of the backward SDE can be written as functions of a forward Markov process, which contains as first com-ponents the (discrete) forward diffusion and the other comcom-ponents depend on the shape of the terminal condition. We denote this process byXti. It turns out that these deterministic functions only depend on the partition point and the number of the Picard-iteration, such that we will write in Chapter 2,Ytni =yni(Xti), Znti =zni(Xti). In Chapter 3 we hereby ignore the influence of the boundRand also writeYtn,Ri =yni(Xti), Zn,Rti =zni(Xti). We emphasize that these functions are not the same across chapters, but there is no danger of mixing them up, because within one chapter we only deal with one set of functions.

Projections on finite-dimensional spaces

In Chapter 2 conditional expectations are further replaced by orthogonal projections on finite-dimensional spaces. We indicate this step by a hat, i.e. we write(Ybtni,Zbtni)for the projection of the time discretized solution of the modified BSDE at timeti and iterationn given the partitionπ on a fixed chosen finite-dimensional subspace.

Monte Carlo simulations

The final approximation step of our proceeding in Chapter 2 replaces the orthogonal projections on finite-dimensional subspaces by an estimator coming from a simulation based least-squares approach. For this purpose we need L independent Monte Carlo simulations of the occurring forward processes. In full detail, we have to simulate in analogy to Bender and Denk [2] the Brownian increments and the forward Markov process and denote them in Chapter 2 by∆λWih and λXti respectively, for λ = 1, . . . ,L and i=0, . . . ,N. Thus it is natural to write(λYbtni,λZbnti)for the resulting estimators for the discretized solution of the backward equation.

Our approach in Chapter 3 directly passes from the time discretization to a simulation based least-squares procedure. For this purpose a whole set of further only imaginary simulations is required. We need for each time pointtiin the partition extra simulations of the Brownian increments and the forward Markov process running until the end of the time horizon of the equation. These new processes are independent conditional to the information up to ti and in the mean time are identically generated as the already existing discrete time processes. To be able to distinguish these sets of processes we signify the imaginary ones with bars, i.e. ∆λWjandλXitj denote these processes at timetj. The additional superindex for the

1.2. Notation 5

discrete Markov process indicates that the additional feature starts at timeti. We will comment on these so-called ’ghost samples’ later on in more detail.

Further notation

A lot of other notation is used in the sequel, see also the index at the end of the thesis, however, it is not helpful to introduce it here. We will do so at the appropriate places and turn now to a variance reduced version of the algorithm of Bender and Denk [2].

Chapter2

Importance sampling

The content of this chapter is already published in Bender and Moseler [4]. We only supplemented some comments and explanations to further clarify our proceeding. The aim of this chapter is to introduce importance sampling for BSDEs. That is we develop a variance reduction method for BSDEs via a change of measure, whose basic idea is borrowed from option pricing.

2.1 Modified forward scheme

We now explain the starting point for the algorithm developed later on. Consider the following family of decoupled FBSDEs parameterized by some measurable, bounded and adapted processh:[0,T]−→RD:

dSth = ³

b(t,Sth) +σ(t,Sth)ht´

dt+σ(t,Sth)dWt, dYth = ³

−f(t,Sth,Yth,Zth) + (Zth)>ht

´

dt+ZthdWt, S0h = s0, YTh=φ(XTh).

where>denotes the transposition of a matrix. We denote(S,Y,Z) := (S0,Y0,Z0), the solution of the original FBSDE withh≡0.

The first observation is that the initial value of the backward part does not depend onh. In fact, defining a new measureQhbydQhhTdPwhere

Ψht =exp

½

Z t

0

h>udWu1 2

Z t

0

|hu|2du

¾ ,

we can apply the Girsanov theorem, to deduce that the law of (Sh,Yh,Zh) under Qh is the same as that of(S,Y,Z)underP. In particular, the constants(Y0,Z0)and (Y0h,Z0h)coincide. We mention that, however, the path of the processes at later time points(Sh,Yh,Zh)and(S,Y,Z)differ. Nonetheless, in many applications, e.g. in option pricing problems, one is mainly interested in estimatingY0. Having the different representations forY0at hand, we aim at reducing the variance of Monte Carlo estimators forY0by a judicious choice ofh. This turns out to generalize the importance sampling technique from calculating expectations to nonlinear BSDEs.

We now introduce the time discretized analog to the Picard-type iteration scheme with importance sam-pling induced by some processh. As it is natural that the choice ofhwill vary with the partitionπ, we do assume from now on that the partitionπis fixed. At first we specify the class of processes which we will consider in the sequel.

7

8 2.1. Modified forward scheme

A 6. The discretized process h is given by

hti =eh(ti,∆W0, . . . ,∆Wi−1)

for some bounded deterministic functioneh:π×RD×. . .×RD−→RD. The bound of h will be denoted Ch. The modified forward scheme is then given by

∆Wih,π = ∆Wi+htii, i=0, . . . ,N−1, ∆WNh,π=0, Ψh,π,jti = exp

½

i−1

k=j

h>tk∆Wk1 2

i−1

k=j

|htk|2k

¾

, j=0, . . . ,N−1, i=j, . . . ,N, Xth,π0 = X0,

Xth,πi = uπ(ti,Xth,πi−1,∆Wi−1h,π), i=1, . . . ,N, and, fori=0, . . . ,N,d=1, . . . ,D

Yth,n,πi = E

·

Ψh,π,itN φ(Xh,πtN ) +

N−1

j=i

Ψh,π,itj f(tj,Sh,πtj ,Yth,n−1,πj ,Zh,n−1,πtj )∆j

¯¯

¯¯Fti

¸

, (2.1)

Zh,n,πd,t

i = E

·∆Wd,ih,π

i µ

Ψh,π,itN φ(Xth,πN ) +

N−1

j=i+1

Ψh,π,itj f(tj,Sh,πtj ,Yth,n−1,πj ,Zth,n−1,πj )∆j

¶¯¯

¯¯Fti

¸

, (2.2)

initialized at(Yth,0,πi ,Zh,0,πti ) = (0, 0). For the special caseh 0, we are just back in the forward scheme discussed by Bender and Denk [2]. Note that, by construction, the firstMcomponents ofXth,π

i coincide withSth,π

i defined via the Euler-Maruyama scheme Sh,πt0 = s0,

Sh,πti+1 = ³

b(ti,Sh,πti ) +σ(ti,Sth,πi )hti

´

i+σ(ti,Sth,πi )∆Wi, i=0, . . . ,N−1.

Defining a new measureQh,πbydQh,πh,π,0tN dPthe Girsanov theorem implies that the process Wth,π=Wt+

N−1

j=0

htj(tj+1∧t−tj∧t),

is Brownian motion underQh,π. Consequently,∆Wh,πare Brownian increments under this measure. This implies that(Xh,π,Fti)is a Markovian process underQh,π and that the transition probabilities of Xh,π underQh,πare the same as those ofXπunderP.

The following theorem shows that, in this Markovian setting, the conditional expectations in the above iteration scheme actually simplify to regressions onXth,πi . On the one hand this is crucial for the Monte Carlo algorithm described in the next section, on the other hand it also allows us to derive some conver-gence results for the modified scheme in an elegant way.

Theorem 2.1.1. Under the standing assumptions there are deterministic functions yn,πi and zn,πi not depending on h such that

Yth,n,πi =yn,πi (Xth,πi ), Zth,n,πi =zn,πi (Xh,πti ).

2.1. Modified forward scheme 9

Proof. We proceed with a double induction, working forward in Picard-iterations and backward in time.

The claim is true forn =0,i =0, . . . ,N, since by definitionYth,0,πi =0 =Zh,0,πd,t

where we first use the martingale property ofΨh,π,itj , the fifth equality is due to the induction hypothesis and the sixth one is true because(Xh,πti ,Fti)is Markovian under the measureQh,π. Finally, the function

10 2.1. Modified forward scheme

Since the regression functions do not depend on the choice ofh and Xh,πt0 = X0, we can conclude that the error made by approximating(Y0,Z0)with(Yth,n,π0 ,Zth,n,π0 )is independent ofh. Hence, we can simply chooseh≡0 for which case the error estimate was already derived in Theorem 1.1.1.

Corollary 2.1.2. There are constants C and C(independent of h) such that for all h

|Yth,n,π0 − Y0|2+|Zh,n,πt0 − Z0|2≤CE h

|XT−XtπN|2 i

+C|π|+C µ1

2+C|π|

n , where Cis the same constant as in Theorem 1.1.1.

Remark 2.1.3. Another way to prove this result is to rewrite the iteration scheme under the new measure Qh,π. Since(Sh,π,Yh,n,π,Zh,n,π)has the same law under the new measure as(Sπ,Yn,π,Zn,π)has underP we can derive the above error estimate.

So far we analyzed the approximation error which is due to the choice ofhand the partitionπ. From now on, we assume that these objects are chosen fixed and we concentrate on other features of the scheme.

Hence, we drop in the following the superindiceshandπ, such that we write for the discretized solution at timetjof the modified FBSDE no more(Sh,πtj ,Yth,n,πj ,Zth,n,πj )but(Stj,Ytnj,Ztnj). Accordingly, we make use of the notationsΨitj andXtj instead ofΨh,π,itj and Xh,πti respectively. As last simplification we introduce

∆Wihas substitute for∆Wih,π.

We now add a further assumption for the remainder of this chapter, which guarantees thatΨ0tiYtni and Ψ0tiZnti are square-integrable underP. This assumption turns out to be essential in order to avoid infinite variances within the Monte Carlo implementation.

A 7. For i=0, . . . ,N−1 E

·µ

Ψ0tNφ(XtN) +

N−1

j=i

Ψ0tjf(tj,Stj, 0, 0)∆j

2¸

<∞. (2.3)

Remark 2.1.4. If the driver f does not depend onSt andφ(XtN) L2+ε(F)for someε > 0 the above condition is satisfied, sinceΨ0tj ∈Lp(F)for anyp>0 andj=0, . . . ,N. This is seen by

Eh

0tj|pi

= E

· exp

½

i−1

j=0

p h>tj∆Wj1 2

i−1

j=0

p2|htj|2j+1

2(p2−p)

i−1

j=0

|htj|2j

¾¸

exp

½1

2(p2−p)Ch2T

¾ E

· exp

½

i−1

j=0

p h>tj∆Wj1 2

i−1

j=0

|p htj|2j

¾¸

= exp

½1

2(p2−p)Ch2T

¾

and Hölder’s inequality finally implies condition (2.3).

For the first level of the Picard-iteration the square-integrability is now straightforward:

Lemma 2.1.5. It holds that0tiYt1i0tiZt1i)∈L2(F)for every i=0, . . . ,N.

Proof. SinceΨitj0tj0ti andΨ0ti isFti-measurable we obtain fori=0, . . . ,N:

Ψ0tiYtni = E

·

Ψ0tNφ(XtN) +

N−1

j=i

Ψ0tjf(tj,Stj,Ytn−1j ,Ztn−1j )∆j

¯¯

¯¯Fti

¸

, (2.4)

Ψ0tiZnd,t

i = E

·∆Wd,ih

i µ

Ψ0tNφ(XtN) +

N−1

j=i+1

Ψ0tjf(tj,Stj,Ytn−1j ,Zn−1tj )∆j

¶¯¯

¯¯Fti

¸

. (2.5)

2.1. Modified forward scheme 11

Consequently forn=1 E

and by Hölder’s inequality Eh

In order to derive the analog result forn>1 we now state some a priori estimates generalizing Lemma 7 in Bender and Denk [2].

Lemma 2.1.6. SupposeΓandγare positive real numbers, yι,zι,ι=1, 2are adapted processes and

j is an{Ftj}-martingale, we can rewrite, noting (2.4) and (2.5), Ψ0tiYt(ι)i = E

12 2.1. Modified forward scheme

In a first step we show that

N−1

2.1. Modified forward scheme 13 Secondly, we shall derive

0≤i≤Nmax λiE

To this end we estimate E

Multiplying withλi, an iterated application of the above inequality yields (2.7).

Finally, we put (2.6) - (2.7) together, and, noting thatZisD-dimensional, we deduce

0≤i≤Nmax λiEh With this result at hand we can conclude:

Corollary 2.1.7. For every i=0, . . . ,N and n∈Nwe have0tiYtni0tiZnti)∈L2(F).

Proof. Considering(Yn,Zn)and(Yn−1,Zn−1)we are in the situation of Lemma 2.1.6 withy(1) = Yn−1, y(2) =Yn−2,z(1) =Zn−1andz(2) =Zn−2. Hence, choosingγ=8DK2(T+1)andΓ=4K2(T+1)(2D(γ+