• Keine Ergebnisse gefunden

Limit behavior of G with one rate change point

In this section, we extend Theorem 3.4 allowing for one rate change point, while testing the null hypothesis of variance homogeneity. Assuming a process with at most one rate change point, the processGcan be shown to converge against a limit process Le (Theorem 3.6), which

3. The MFT for testing variance homogeneity

is, likeL, a zero-mean 2h-dependent Gaussian process with unit variance (Corollary 3.7). It differs fromL only in the covariance in the 3h-neighborhood of a change pointc (see Section 5.2 and Fig. 5.4 C,D).

Theorem 3.6. One rate change point: Convergence of G

LetΞ(n) ∈M (Def. 3.3) with at most one rate change and no variance change, as follows. Let Ξ11, σ12, ν12),Ξ22, σ22, ν22)∈R withµ1 6=µ2, σ1222. For c∈(0, T]and n= 1,2, . . . let Ξ(n) := Ξ1|[0,nc]+ Ξ2|(nc,nT], (3.9) meaning that Ξ(n) fulfills H0. Assume a consistent estimator ˆc of c with

|ˆc−c|=oP(1/√

n) (3.10)

where oP(·) is the small o-notation with respect to convergence in probability. Let G(n) be the filtered derivative process associated withΞ(n) using the empirical means µˆˆc1, µˆˆc2 estimated in the intervals [0,ˆc) and [ˆc, T]. Then with Le from (3.11), as n→ ∞, we have

G(n) −−→d L,e

where −→d denotes weak convergence in the Skorokhod topology. The marginals Leh,t of the limit process Le equalL outside the h-neighborhood of c and are given by

Leh,t=









L√h,t, |t−c|> h,

riν2)2/(µ2h2)(Wt+h−Wc)+

riν1)2/(µ1h2)(Wc−Wt)−

µ1ν12/h2(Wt−Wt−h)

s(1)t , t∈[c−h, c],

µ2ν22/h2(Wt+h−Wt)−

leν2)2/(µ2h2)(Wt−Wc)−

leν1)2/(µ1h2)(Wc−Wt−h)

s(1)t , t∈(c, c+h],

(3.11) for a standard Brownian motion (Wt)t≥0. The functions µri := µri,h,t, µle :=µle,h,t are the limits of the empirical means µˆri,µˆle and are given by µri,h,t:=µ1 for t≤c−h, µri,h,t:=µ2 for t > cand

µri,h,t:= h

(c−t)/µ1+ (t+h−c)/µ2

, (3.12)

fort ∈(c−h, c] and analogously for µle. The true order of scaling

(s(n)t )2

t∈τh

is defined by nh/µ12

1 for t < c−h, by nh/µ22

2 for t > c+h and for |t−c| ≤ h by the following linear interpolation

(s(n)t )2 := (s(n)h,t)2 :=

1 n

µ1ν12

h +(c−t)h2µ1riν1)2+(t+h−c)h2µ2riν2)2

, if c−h≤t≤c,

1 n

(c−(t−h))

h2µ1leν1)2+(t−c)h2µ2leν2)2+ µ2hν22

, if c < t≤c+h.

(3.13) Sketch of proof: Again, we sketch the proof and refer for the detailed proof version to Section 4.2.

The key ingredients are the Anscombe-Donsker-Theorem and continuous mapping. In addition to the proof of Theorem 3.4, a change point in the rate requires separate considerations for

20

3. The MFT for testing variance homogeneity

different intervals in the neighborhood of a change point. Like in the proof of Theorem 3.4 we first assume known process parameters and use the modified filtered derivative process Γ

Γ(n)t = Γ(n)ri,t −Γ(n)le,t, (3.14) which is comparable to the process Γ(n)t defined in equation (3.6). The detailed definitions are given in Section 4.2. Moreover, we decompose the limit processLe∼Leri−Lele, where∼ denotes equality in distribution and refer again for detailed definitions to Section 4.2. The first step of the proof will be to show convergence of the processes

Γe(n)ri :=

Nn(t+h)−Nnt−1 nh/µri

t

·Γ(n)ri and Γele against

Leri,Lele

using the Donsker-Ascombe-Theorem and continuous mapping.

With Lemma 4.5, we can then conclude

Γ(n)ri(n)le d

−→

Leri,Lele

, and using continuous mapping again yields fort∈τh

Γ(n)= Γ(n)ri −Γ(n)le −−→d Leri−Lele∼L.e

In step two of the proof, we first replace the true means µ1, µ2 in the numerator by their estimators to define the process ˆΓ(n) and show

Γˆ(n)−Γ(n)−−→P (0)t. (3.15) Then, we use Lemma 4.6 to substitute the scalings(1)t used in ˆΓ by the estimator ˆs(n)t to prove the assertion.

We show thatLe is a Gaussian process with zero mean and unit variance.

Corollary 3.7. Marginal distribution of Le

Let Le be defined as in (3.11). For all t∈τh it holds Leh,t ∼N(0,1).

Proof: As the increments of a standard Brownian motion (Wt)t≥0are independent and Gaussian distributed,Le is Gaussian distributed. Moreover, the increments have zero expectation and thus the zero mean follows by the linearity of expectation. To show the unit variance, we analyse three cases. We use the independence of increments and the property that for 0≤s≤t: Var(Wt−Ws) =t−s. For |t−c|> h the assertion follows directly. Now, let t∈[c−h, c]. We obtain

Var(Leh,t) =

µ2riν22

µ2h2(t+h−c) + µµ2riν21

1h2(c−t) +µh1ν212h

µ1ν12

h + hri2νµ11)2(c−t) +hri2νµ22)2(t+h−c) = 1.

The caset∈(c, c+h] is shown analogously.

Note that analogous results to Theorem 3.6 hold if there are several rate change points with pairwise distances each larger than 2h as shown in Figure 3.3 for two rate change points.

3. The MFT for testing variance homogeneity

In theh-neighborhood of each rate change point the marginals of the corresponding limit process are similar to the marginals ofLe aroundc and in between the marginals are identical to the marginals ofL. In case of rate change points with distance smaller than or equal to 2h, the structure of the limit process becomes more complicated but is still comparable to the structure ofL. Furthermore, note that the proof of Theorem 3.6 is based on a Functionale Central Limit Theorem and a consistent estimator ofst. Therefore the result can be shown not only for renewal processes but also for a subclass of renewal processes with varying variance (RPVV) as introduced in Messer et al. (2014).

c1 c1+h

c1h c2h c2c2+h

~L

L~

L L L

marginals of limit process similar to

Figure 3.3: Limit process in the case of two rate cps with distance larger than 2h. The marginals of the limit process resemble in the h-neighborhoods of the two rate change pointsc1, c2 the marginals of the processL˜ around its rate change point, and outside they are identical to the marginals of L.

As the marginals of L andLe differ only in the h-neighborhood of cand both processes are 2h-dependent, their covariance structures differ in the 3h-neighborhood of the rate change pointc, which is illustrated in Figure 3.4. Our simulations in Section 5.2 suggest that the differences between L and Le are typically small with respect to the 95%-quantile of their absolute maxima. We therefore suggest to use the parameter independent limit processLalso in the situation of potential rate change points for the derivation of the rejection threshold in the statistical test. The simulations in Section 5 show that the MFT using L instead of Le keeps the asymptotic significance level for most combinations ofµand σ even for the case of multiple unknown rate change points.

c c+h

ch c+3h c3h

impact of

~L

c−h L~

c+h

Figure 3.4: Impact of a rate change point on the covariance of L.e The processes L and Le are 2h-dependent as visualized by orange arrows in the graph. As the marginals of L and Le differ in the h-neighborhood of the rate change point c (colored orange), the covariance of Le differs in a 3h-neighborhood of the rate change point from the covariance of L.

22

Chapter 4

Proofs of Theorems 3.4 and 3.6

In this section, we prove the key theoretical results of the first part of this thesis in detail. The proof of Theorem 3.4 in Section 4.1 uses the Anscombe-Donsker-Theorem, continuous mapping and the consistency of the estimator ˆs2. In Section 4.2, we prove Theorem 3.6 basically using the same ideas as for the proof of Theorem 3.4.

We specify the sets of indices

le:={Nn(t−h)+ 2, Nn(t−h)+ 3, . . . , Nnt} and Iˆri :={Nnt+ 2, Nnt+ 3, . . . , Nn(t+h)}.

The Anscombe-Donsker invariance principle is of central importance for the proofs. We state it here following Gut (2009, Theorem 2.1, p.158).

Theorem 4.1. Anscombe-Donsker-Theorem

Let(ξk)k≥1 be i.i.d. random variables with zero mean and variance σ2 <∞, ( ˜N(t))t≥0 be a nondecreasing, right-continuous family of positive, integer valued random variables and set

Zn(t) := 1 σ√

n

N˜nt

X

k=1

ξk.

Suppose that

N˜(t) t

−−−→a.s.

t→∞ κ (0< κ <∞).

Then we have in (D[0,∞), dsk)

Zn

√κ

−−−→d t→∞ W, with W as standard Brownian motion.

Proof: Compare, e.g., Gut (2009).

4. Proofs of Theorems 3.4 and 3.6

4.1 Proof of Theorem 3.4

Recall the sketch of proof directly following the statement of Theorem 3.4. The processes Γ(n)le and Γ(n)ri in equation (3.6) are in detail defined as

Γ(n)ri,t := 1 Step 1: weak process convergence for known parameters

RecallVi= (ξi−µ)2 for the life times (ξi)i≥1 of a point process Ξ with a known meanµ. We apply the Anscombe-Donsker-Theorem to the process defined by

Yt(n):= 1 Here,W denotes a standard Brownian motion.

Letϕ: (D[0,∞), dSK)→(D[h, T −h]×D[h, T −h], dSK⊗dSK) be defined by

This function is continuous. Mapping √

µY(n) via ϕ, the first component is given by

By Slutsky’s theorem this also implies

 see Lemma A.3.2 in Messer et al. (2014). Thus by Slutsky’s theorem,

(n)ri,t)t∈τh d asn→ ∞ by continuous mapping which is the assertion for a known mean µ.

24

4. Proofs of Theorems 3.4 and 3.6

Note that the expectationE[V1] vanishes, since it appears in both summands.

Step 2: replacement of parameters by their estimators

In a second step we use the estimated mean ˆµnT. We show that for

Thus, the same arguments as in step 1 can be applied to show the assertion G(n) −→ L.

Additionally, by Slutsky’s theorem and Corollary 4.4 (given below), we can exchange the factor (nh/2µν2)1/2 with the estimator 1/ˆs(nt, nh) with the convergence holding true.

To show (4.1), we rewrite ˆYt(n)=Yt(n)+R(n)t , with R given by This decomposition holds true since

Nnt convergence in (4.1) follows by Slutsky’s theorem.

It suffices to show that supt∈[0,T]R(n)t vanishes in probability. Recall σ2=Var(ξ1) and set

Then, by the Anscombe-Donsker-Theorem we find in (D[0, T], dSK) that

(Zt(n))t∈[0,T] → (Wt)t∈[0,T] weakly as n → ∞, such that supt∈[0,T]Zt(n) → supt∈[0,T]Wt

weakly. Applying the reflection principle, it is known that the distribution of supt∈[0,T]Wt directly derives from the normal distribution (e.g., Billingsley, 1968). Now, we first focus on the second summand ofR(n)t and show that its square root vanishes

sup

4. Proofs of Theorems 3.4 and 3.6

asn→ ∞. This holds true since in (D[0, T], dSK) it holds almost surely that (Nnt)t∼(nt/µ)t asn→ ∞, compare, e.g., Lemma A.3.2 in Messer et al. (2014). Thus, the first factor within the supreme itself vanishes such that the entire expression tends to zero. By continuous mapping theorem, this holds true for the squared expression, such that the second summand inR(n)t uniformly tends to zero in probability. For the first summand inR(n)t , we decompose (ˆµnT−µˆnt) = (ˆµnT−µ)−(ˆµnt−µ) and apply the same argument as before to both summands.

As a resultR uniformly vanishes in probability.

Now, we state the consistency of the variance estimator ˆsin Corollary 4.4. Therefore, we need the consistency of the local estimators forµand ν as stated in the following lemmata.

Lemma 4.2. Let Ξbe an element ofR with mean µ. Let T >0,h∈(0, T /2] and µˆle and µˆri be defined as the empirical means of the life times in the left and the right window, respectively.

Then it holds in (D[h, T −h], d||·||)) almost surely as n→ ∞ that

(ˆµle)t∈τh−→(µ)t∈τh and (ˆµri)t∈τh−→(µ)t∈τh. Proof: This is a weaker version of Lemma A.15 in Messer et al. (2014).

Lemma 4.3. Let Ξ be an element ofR with ν2 =Var((ξ1−µ)2). Let T >0, h ∈ (0, T /2]

and νˆle2 and νˆri2 be defined as in (3.5) using the estimated global mean ˆ

µ(i)= ˆµnT = (1/NnT)PNnT

i=1 ξi ∀i≥1. Then it holds in(D[h, T −h], d||·||))almost surely as n→ ∞ that

(ˆνle2)t∈τh−→(ν2)t∈τh and (ˆνri2)t∈τh−→(ν2)t∈τh. (4.2) Proof: For a known mean µ, i.e., Vi instead of ˆVi convergences (4.2) are shown using the same techniques as for the consistencies of (ˆµ)t and (ˆσ2)tin Messer et al. (2014). A complete proof can be found in Albert (2014). To extend the result to the estimated global mean ˆµnT we derive

ˆ

νri2 = 1

Nn(t+h)−Nnt−1

Nn(t+h)

X

i=Nnt+2

((ξi−µˆnT)2−σˆri2)2

= 1

Nn(t+h)−Nnt−1

Nn(t+h)

X

i=Nnt+2

i2−2ξiµˆnT + ˆµ2nT −σˆri2)2

=

Nn(t+h)

P

i=Nnt+2

i4−4ξi3µˆnT + 6ξi2µˆ2nT −2ξi2σˆri2 −4ξiµˆ3nT + 4ξiµˆnTσˆri2 + ˆµ4nT −2ˆµ2nTσˆri2 + ˆσ4ri)

Nn(t+h)−Nnt−1 .

(4.3) Using the summand withξ3iµˆnT as an example, we explain that the difference between the summands with the estimators forµandσ in the latter display and with the known values vanishes a.s. asymptotically forn→ ∞. It holds

1

Nn(t+h)−Nnt−1

Nn(t+h)

X

i=Nnt+2

ξ3iµˆnT − 1

Nn(t+h)−Nnt−1

Nn(t+h)

X

i=Nnt+2

ξi3µ

= 1

Nn(t+h)−Nnt−1

Nn(t+h)

X

i=Nnt+2

ξ3i(ˆµnT −µ). (4.4)

26

4. Proofs of Theorems 3.4 and 3.6

The strong consistency (ˆµnT)t∈τh → (µ)t∈τh a.s. asn → ∞ follows directly by Lemma 4.2 (note that in our situation regarding the fixed point in timeT is sufficient) and thus it follows that (ˆµnT −µ)t∈τh vanishes a.s. Moreover, withξi∈L4 and the exact same techniques as in the consistency proof of ˆσri (Lemma A.16 in Messer et al., 2014) it can be shown that almost surely asn→ ∞

1

Nn(t+h)−Nnt−1

Nn(t+h)

X

i=Nnt+2

ξi3

t∈τh

−→(c)t∈τh

for some constant c >0. This implies that the expression in (4.4) vanishes a.s. in (D[h, T −h], d||·||)).

Similar arguments using also the strong consistency of ˆσri (where the replacement of the estimator ˆµri by the estimator ˆµnT neither changes the correctness of Lemma A.16 in Messer et al. (2014) nor the arguments in its proof) hold for all other summands in equation (4.3).

Consequently, the assertion follows by standard application of Slutsky’s theorem.

Lemma 4.3 directly implies the consistency of the variance estimator ˆs2.

Corollary 4.4. Let Ξ∈R with ν2 =Var((ξ1−µ)2). Let T >0, h∈(0, T /2] and ˆs2(t, h) be defined as in (3.4). Then it holds in (D[τh], d||·||) almost surely as n→ ∞ that

nsˆ2(nt, nh)

t∈τh−→

2 h/µ

t∈τh

.

Proof: Recall equation (3.4)

ˆ

s2t := νˆri2 h/ˆµri

+ νˆle2 h/ˆµle.

The assertion follows from the consistency of ˆµri,µˆle (Lemma 4.2) and ˆνri2,νˆle2 (Lemma 4.3) by application of Slutsky’s theorem.