• Keine Ergebnisse gefunden

eΓTR2(1+T)2D

N−1

i=0

1

i Z

|x|≥R0

ix2dN(0,∆i)

= eΓTR2(1+T)2D 1

N−1

i=0

1 (∆i)3/2

Z

|x|≥R0

i

x2ex

2 2∆idx

= eΓTR2(1+T)2D 1

N−1

i=0

1 (∆i)3/2

p∆i Z

|y|≥R0iy2ey

2 2 dy

= eΓTR2(1+T)2D 1

N−1

i=0

Z

|y|≥R0

y2ey

2 2 dy

= 2NeΓTR2(1+T)2D 1

e−R20/4 Z

R0 y2ey

2

2+R20/4dy

2NeΓTR2(1+T)2D 1

e−R20/4 Z

R0

y2ey

2 4 dy

2NeΓTR2(1+T)2D 1

e−R20/4C

= CNR2exp(−R20/4),

whereN(µ,σ2)denotes the cumulative distribution function of Gaussian distributed random variables with meanµand varianceσ2.

Now, we are able to provide the aggregate error which is generated by this kind of truncation:

Theorem 3.1.7. There is a constant C such that for any n∈N

0≤i≤N−1max E h

|Ytni −Ytn,Ri |2 i

+

N−1

i=0

E h

|Ztni −Zn,Rti |2 i

i≤CNR2exp(−R20/4), given|π|is small enough.

Proof. Corollary 3.1.4, Lemma 3.1.5 and Lemma 3.1.6 yield

0≤i≤N−1max E h

|Ytni −Ytn,Ri |2 i

+

N−1

i=0

E h

|Ztni −Zn,Rti |2 i

i

max

0≤i≤N−1λiEh

|Ytni −Ytn,R

i |2i

+

N−1

i=0

λiEh

|Znti−Zn,Rt

i |2i

i

n−1

k=0

µΓ 4|π|+1

2

k

CNR2exp(−R20/4)

CNR2exp(−R20/4)

for some other constantCprovided|π|is small enough.

3.2 Projection approach in the case of Markov processes

The next step is the basis for a least-squares Monte Carlo procedure and is easily proven with the results of the second chapter:

30 3.2. Projection approach in the case of Markov processes

Lemma 3.2.1. There are deterministic functions yni and zd,ni (i=0, . . . ,N,n=0, 1, . . . ,d=1, . . . ,D)such that Ytn,Ri =yni(Xti)and Zd,tn,R

i =zd,ni (Xti).

We define zni = (z1,ni , . . . ,zD,ni )>.

Proof. We can proceed as in Lemma 2.1.1 and only have to take into account that sinceXti and∆Wd,iare independent the same is true forXti and[∆Wd,i]wi.

Note, that here the regression functions also depend onR, however we suppress this feature for simplicity in the sequel.

Now, we choose as in the last chapter fori=0, . . . ,N−1 and anyd=0, . . . ,Din each caseKd,i determin-istic basis functionspd,i,k(·),k =1, . . . ,Kd,i such thatpd,i(Xti)is square-integrable. We thereby used the notation

pd,i(·) = (pd,i,1(·), . . . ,pd,i,Kd,i(·))>.

In the following proofs we make use of a lot of structures, which we now describe in detail starting with the projection spaces:

Projection spaces and matrices:

Usually, as approximation, projections to the finite-dimensional subspaces Pd,i =:n

α·pd,i(·),α∈RKd,io

are considered. Thereby, we use for the approximation of yni the basis functions p0,i and for the ap-proximation ofzd,ni the functions pd,i withd = 1, . . . ,D. Due to our special situation we apply a slight modification, which we describe in detail later on.

In order to calculate the different projection coefficients we needL>maxd,iKd,iindependent Monte Carlo simulations ofXti,i =0, . . . ,Nand∆Wi,i =0, . . . ,N−1. We denote them again byλXti λ = 1, . . . ,L, i=0, . . . ,Nand∆λWi,λ=1, . . . ,L,i=0, . . . ,N−1 respectively.

To simplify the notation we writepd,i,k(λXti) = pλd,i,kand pλd,i = (pλd,i,1, . . . ,pλd,i,K

d,i)>ford =0, 1, . . . ,D.

We define ford =0, 1, . . . ,Dthe matrixBd,iL with dimensionL×Kd,iwhich just contains(pλd,i)>as rows.

The rank of this matrix is denoted byKLd,iwhich is smaller or equal toKd,iand random.

As the unknown functions yni and zd,ni are bounded by Cy and Cπz respectively, their approximations should satisfy this property as well and we therefore approximateyni with an element of

[P0,i]y=:

n

·p0,i]y(·),α∈RK0,i o

andzd,ni ford=1, . . . ,Dwith an element of [Pd,i]z =:n

·pd,i]z(·),α∈RKd,io ,

where[·]yand[·]ztruncate atCyandCπz respectively, i.e. for any real-valued functionξwe set [ξ]y(x):= (−Cy∨ξ(x))∧Cy,

[ξ]z(x):= (−Czπ∨ξ(x))∧Czπ.

’Ghost samples’:

For the convergence proof we need extra simulations ofλXtj, j = 0, . . . ,Nand the simulated Brownian

3.2. Projection approach in the case of Markov processes 31

increments∆λWj,j=0, . . . ,N−1. More precisely, we create a further set of independent (from everything else) Brownian increments denoted by∆λWj,j=0, . . . ,N−1,λ=1, . . . ,Land thereby constructNsets of discrete Markov processesλXi,i=0, . . . ,N−1 such that forλ=1, . . . ,L,j=i+1, . . . ,Nthe random variablesλXitj andλXtj conditioned toλXti are independent and identically generated andλXitj = λXtj

forλ=1, . . . ,Landj=0, . . . ,i.

If one applies the Euler-Maruyama scheme for the approximation of the forward diffusion this means for the first components ofλXtj andλXitj:

λStj+1 = λStj +b(tj,λStj)∆j+σ(tj,λStj)∆λWj, j=0, . . . ,N−1,

λSitj = λStj, j=0, . . . ,i,

λSitj+1 = λSitj+b(tj,λSitj)∆j+σ(tj,λSitj)∆λWj, j=i, . . . ,N−1.

The other components are constructed by

λXitj+1 =uπ(tj+1,λXitj,∆λWj), j=i, . . . ,N−1.

Discretized norms:

We introduce the following discretized norms: For aRD0-valued functionξ= (ξ1, . . . ,ξD0)>we define for i=0, . . . ,N

kξki,L :=

vu ut1

L

L λ=1

D0

d=1

d(λXti)|2 and fork=0, . . . ,N−1,i=k, . . . ,N

kξki, ¯Lk :=

vu ut1

L

L λ=1

D0 d=1

d(λXkti)|2.

Algorithm and projection coefficients:

Our algorithm now works as follows: It is initialized withy0,Li = 0 = z0,Li and for any n Nwith yn,LN (λXtN) =φ(λXtN)andzn,LN =0.

Afterwards, we calculate the following projection coefficients:

αn,L0,i =arginf

α

1 L

L λ=1

¯¯

¯¯

½

φ(λXtN) +

N−1

j=i

f(tj,λStj,yn−1,Lj (λXtj),zn−1,Lj (λXtj))∆j

¾

−α·pλ0,i

¯¯

¯¯

2

(3.4)

and ford=1, . . . ,D αn,Ld,i =arginf

α

1 L

L λ=1

¯¯

¯¯[∆λWd,i]wi

i

½

φ(λXtN) +

N−1 j=i+1

f(tj,λStj,yn−1,Lj (λXtj),zn−1,Lj (λXtj))∆j

¾

−α·pλd,i

¯¯

¯¯

2

.(3.5) We emphasize that these coefficients are despite the same notation not equal to those of Chapter 2. How-ever, since in the remainder of this chapter we only will use the here defined version, no confusion should arise.

32 3.2. Projection approach in the case of Markov processes

Our approximations then finally are

yin,L(λXti) = [αn,L0,i ·p0,i]y(λXti), zd,n,Li (λXti) = [αn,Ld,i ·pd,i]z(λXti).

We additionally need the following projection coefficients to prove convergence:

eαn,L0,i =arginf the functionsyn−1,Lj and zn−1,Lj are fixed and we use for the estimation ofαn,Ld,i the original simulations

λXti,λ=1, . . . ,Land∆λWj j=i, . . . ,N−1,λ=1, . . . ,L, whereas the estimation ofeαn,Ld,i is based on the

Orthonormality of the basis functions

The definition of the above projection coefficients is not always unique. Especially, we have to specify which solution we choose if the matrices(BLd,i)>Bd,iL ford =0, 1, . . . ,Dandi =0, . . . ,N−1 are singular.

As Gobet et al. [22] we choose an approach due to the singular value decomposition of a matrix, which we shortly sketch in the appendix based on the description in Golub and Van Loan [24]. Proceeding this way, we can suppose that for anyd =0, 1, . . . ,Dandi =0, . . . ,N−1 the identity 1L(Bd,iL )>Bd,iL = IdKL solve problems (3.4) and (3.5) respectively.

3.2. Projection approach in the case of Markov processes 33

σ-algebras, conditional expectations and conditional probabilities:

We define theσ-algebras

FiL :=σ(∆λWj,λ=1, . . . ,L,j=0, . . . ,i−1),

FiL,k :=σ(∆λWj,λ=1, . . . ,L,j=0, . . . ,N−1,∆λWk, . . . ,∆λWi−1,λ=1, . . . ,L)fori=k+1, . . . ,N, FkL,k :=FL:=σ(∆λWj,λ=1, . . . ,L,j=0, . . . ,N−1)

and the corresponding syntax for the conditional expectations and conditional probabilities. We introduce the redundant parameterkin the third definition to harmonize notation in the sequel.

Further abbreviations:

We define to shorten the notation:

λfin := f(ti,λSti,yni(λXti),zni(λXti)),

λfin,L := f(ti,λSti,yn,Li (λXti),zn,Li (λXti)),

λfn,ki := f(ti,λSkti,yni(λXkti),zni(λXkti)), i=k, . . . ,N−1,

λfn,L,ki := f(ti,λSkti,yn,Li (λXkti),zn,Li (λXkti)) i=k, . . . ,N−1, yλ,n,L,ki := EiL,kh

φ(λXktN) +

N−1

j=i

λfn−1,L,kjji

= EiL,k h

yλ,n,L,ki+1 i

+ λfn−1,L,kii, i=k, . . . ,N−1, . For functionsξd:RM0 R,d=0, . . . ,Dwe write furthermore

λfi0,..,D) := f(ti,λSti,ξ0(λXti),(ξd(λXti))d=1,..,D),

λfki0,..,D) := f(ti,λSkti,ξ0(λXkti),(ξd(λXkti))d=1,..,D).

Occurring regular and exception sets:

Following events play a decisive role in the subsequent proofs. Forn∈N,d=1, . . . ,D,i=0, . . . ,N−1 andβ>0 to be chosen later on we define:

A0,in,L =

n

k{eαn,L0,i −αn,L0,i }p0,ik2i,L <iβ o

,

An,Ld,i = n

k{eαn,Ld,i −αn,Ld,i }pd,ik2i,L <iβo , An,Ly,ik =

n

∀ξ∈[P0,i]y−yn−1i :kξki,L

k− kξki,L<β/2i o

,

An,Lz,ik =



∀ξ∈

 [P1,i]z

... [PD,i]z

−zn−1i :kξki,L

k− kξki,L <iβ/2



, and forn∈N,i=1, . . . ,N−1

An,Lz i−1,Li =

½µ1 L

L λ=1

|yni(λXi−1ti )−yiλ,n,L,i−1|2

1/2

µ1

L

L λ=1

|yin(λXiti)−yλ,n,L,ii |2

1/2

<iβ+1

¾ .

34 3.2. Projection approach in the case of Markov processes

After this bunch of definitions we now turn to the mathematics. The key tool for the convergence proof is the following observation:

Lemma 3.2.2. Let k =0, . . . ,N−1,Γ >0andλi, i =0, . . . ,N−1withλ0=1andλi+1= (1+Γ∆ii for

where the second equality is due to the conditional independence ofλXitj andλXtj, see e.g. Chow and Teicher [13], Corollary 7.3.2. Moreover, we obtain

yni(λXkti)−yλ,n,L,ki

Via Young’s and Jensen’s inequality and because of the Lipschitz continuity of f we gain for anyΓ>0:

E

3.2. Projection approach in the case of Markov processes 35

Repeated application of Young’s inequality and multiplication withλiyields λiE It is quite bothering that the discretized norms are evaluated at the ghost samples which will be overcome with the help of the exception sets. For the first inequality below we additionally use the following property of non-negative numbersx,y:

x22(x−y)2++2y2.

36 3.2. Projection approach in the case of Markov processes

We start with the convergence proof with a result for the first part of the solution of the BSDE:

Proposition 3.2.3. For any n=2, 3, . . .,Γ,β>0andλi, i=0, . . . ,N withλ0=1andλi+1= (1+Γ∆iifor Due to the equality

yni(λXti) = E

and the minimality of the coefficientβen,L0,i we obtain that its conditional expectationEL(eβn,L0,i)minimizes 1

L

L λ=1

|yni(λXti)−α·pλ0,i|2.

The Pythagorean theorem then implies:

E

3.2. Projection approach in the case of Markov processes 37

The second term of (3.6) can be estimated by the contraction property of projections:

3E and the first term of (3.7) is an error term:

3Eh

k{EL(eαn,L0,i )eαn,L0,i}p0,ik2i,Li

=3T2,in,L. (3.9)

The second term of (3.7) remains which can be tackled again by the contraction property:

3E Multiplying withλiand transition to the maximum yields:

0≤i≤N−1max λiEh Lemma 3.2.2 then implies the assertion fork=i:

0≤i≤N−1max λiE

38 3.2. Projection approach in the case of Markov processes

Next, we consider the corresponding error for theZpart:

Proposition 3.2.4. For any n=2, 3, . . .,Γ,β,µ>0andλi, i=0, . . . ,N withλ0 =1andλi+1= (1+Γ∆ii sincezd,ni is bounded and[·]zis Lipschitz continuous:

E

which is due to the conditional identical and conditional independent distribution of the involved random variablesλXtj andλXitj, and because of the minimality ofeβn,Ld,i, we know thatEL(βen,Ld,i)minimizes

The Pythagorean theorem and the triangle inequality then yield:

E

3.2. Projection approach in the case of Markov processes 39

The second term of (3.12) is estimated withβ>0 and the contraction property as follows:

3Eh

k{eαn,Ld,i −αn,Ld,i} ·pd,ik2i,Li

=3E

·

k{eαn,Ld,i −αn,Ld,i } ·pd,ik2i,L1{k{eαn,L

d,i−αn,Ld,i}·pd,ik2i,L<∆βi}

¸

+3E

·

k{eαn,Ld,i −αn,Ld,i } ·pd,ik2i,L1{k{eαn,L

d,i−αn,Ld,i}·pd,ik2i,L≥∆βi}

¸

3∆iβ +3E

·1 L

L λ=1

¯¯

¯¯[∆λWd,i]wi

i

½

φ(λXitN) +

N−1

j=i+1

λfn−1,L,ijj

¾

[∆λWd,i]wi

i

½

φ(λXtN) +

N−1

j=i+1

λfjn−1,Lj

¾¯¯

¯¯

2

·1

{k{eαn,Ld,i−αn,Ld,i}·pd,ik2i,L≥∆βi}

¸

3∆iβ+12R2(1+T)2R20

iP([An,Ld,i]c). (3.13)

The first summand of (3.12) directly yields an error term:

3Eh

k{EL(eαn,Ld,i)eαn,Ld,i } ·pd,ik2i,Li

=3T4,d,in,L, (3.14)

and we are left with the second summand of (3.11). The contraction property of projections yields:

3Eh

k{EL(βen,Ld,i)−EL(eαn,Ld,i)} ·pd,ik2i,Li

3E

·1 L

L λ=1

¯¯

¯¯EL

·[∆λWd,i]wi

i

½

φ(λXitN) +

N−1

j=i+1

λfn−1,ijj−φ(λXitN)

N−1

j=i+1

λfn−1,L,ijj

¾¸¯¯

¯¯

2¸ .

The above inequality can be transformed with the help of the tower property of conditional expectations and the measurability ofλWd,ito:

3E h

k{EL(βen,Ld,i)−EL(eαn,Ld,i)} ·pd,ik2i,L i

3E

·1 L

L λ=1

¯¯

¯¯EL

·[∆λWd,i]wi

i

½ Ei+1L,i

· N−1

j=i+1

(λfn−1,ij λfn−1,L,ij )∆j

¸¾¸¯¯

¯¯

2¸

. (3.15)

As[∆λWd,i]wiis independent ofFLand the truncation is symmetric to zero, we haveEL£

[∆λWd,i]wiV¤

=0

40 3.2. Projection approach in the case of Markov processes

for anyFL-measurable random variableV. We apply this property to term (3.15) and obtain:

3E where the first inequality is due to the Cauchy-Schwarz inequality.

We add in term (3.16) similar terms however containing the ghost samples starting at timeti−1such that if summing them up we obtain a telescoping sum. Definingyn0(λX−1t0 ) := 0 =: yλ,n,L,−10 simplifies the notation at this juncture. For term (3.17) we exploit the Cauchy-Schwarz inequality and get forµ>0:

3

3.2. Projection approach in the case of Markov processes 41

where we used in the last inequality the Lipschitz continuity of f, Young’s inequality and the identity

λXiti = λXti.

Fori = 0, (3.19) - (3.20) is negative and we can skip this term. To keep notation as easy as possible we thus defineAn,Lz −1,L0 =∅. Since fori =1, . . . ,N−1 the second factor of (3.19) - (3.20) is nonnegative we can estimate this expression further by:

E

Hence, adding (3.13), (3.14), (3.18), (3.21), (3.22), (3.23) we obtain for theD-dimensional processZ E Multiplying withλiiand summing up from 0 toN−1 finally yields:

N−1

42 3.2. Projection approach in the case of Markov processes

The assertion is then implied by Lemma 3.2.2.

The whole error originating by the Monte Carlo simulation in then-th Picard iteration can be estimated by:

Proof. Gathering the both preceding propositions yields forn=2, 3, . . .

0≤i≤N−1max λiE

3.2. Projection approach in the case of Markov processes 43

Proceeding exactly as before we get an upper bound for the error occurring in the first Picard iteration:

Lemma 3.2.6. Taking into account the notation of Proposition 3.2.5 it holds:

0≤i≤N−1max λiEh

Proof. Proceeding in the same way as in Proposition 3.2.3 the boundedness and Lipschitz continuity of [·]yyields fori=0, . . . ,N−1

and because of the minimality of the coefficient βe1,L0,i we know that EL[βe1,L0,i] minimizes the expression

1LLλ=1|y1i(λXti)−α·p0,iλ|2. Thus, the Pythagorean theorem and the contraction property imply:

44 3.2. Projection approach in the case of Markov processes

Analogly, we obtain for the second part of the solution:

Eh

and due to the minimality of the coefficientβe1,Ld,i its conditional expectationEL[βe1,Ld,i]minimizes the expres-sion1LLλ=1|zid,1(λXti)−α·pλd,i|2. Hence, the Pythagorean theorem and the contraction property imply: As consequence we can phrase:

Corollary 3.2.7. Taking into account the definitions of Proposition 3.2.5 we obtain for any n∈N:

0≤i≤N−1max Eh for a number of simulationsLexceeding all bounds. The next section is dedicated to this task.