• Keine Ergebnisse gefunden

Solution to Series 3

N/A
N/A
Protected

Academic year: 2022

Aktie "Solution to Series 3"

Copied!
5
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

Solution to Series 3

1. a) Plotting and calculating the mean:

> mu <- mean(t.yields)

> plot(t.yields)

> abline(h = mu, lty=3)

Time

t.yields

0 10 20 30 40 50 60 70

305070

We can regard this time series as being stationary.

b) Plotting the ACF:

> acf(t.yields, plot = TRUE)

0 5 10 15

−0.40.41.0

Lag

ACF

The correlogram shows us that for lags k ≥ 3, the estimated autocorrelations ρbhki do not differ significantly from 0. The first of these autocorrelations is negative; as the time series oscillates very noticeably, this negativity is not at all surprising.

Looking at lagged scatterplots:

> lag.plot(t.yields, lag = 6, layout = c(2, 3), do.lines = FALSE)

lag 1

t.yields305070

20 40 60 80

lag 2

t.yields

lag 3

t.yields

20 40 60 80

lag 4

t.yields

lag 5

t.yields

20 40 60 80

lag 6

t.yields 305070

In the lagged scatterplot with lag 1 the pairs [xt, xt+1] show the negative linear relationship we expected from the correlogram. For lag 2, however, the lagged scatterplot shows up a positive linear relationship, and for lagk≥4we see no further correlation. The pairs[xt, xt+3] (lagged scatterplot at lag 3) still have a slightly negative connection, but the correlogram tells us that we can assume ρh3b i = 0.

(2)

c) The variance of the arithmetic meanµbis

Var(µ) =b 1

n2 γ(0) n+ 2

n−1

X

k=1

(n−k)ρ(k)

! .

From the correlogram in Part b) we see that the estimated autocorrelations bρhki do not differ significantly from 0 for lagsk≥3. Thus we can set all the autocorrelationsρ(k)fork≥3to 0. We obtain

Var(µ) =b 1 n2 γ(0)

n+ 2(n−1)ρ(1) + 2(n−2)ρ(2) .

To estimate the variance ofµ, we replaceb γ(0), ρ(1) andρ(2)by their estimates.

R code:

> n <- length(t.yields)

> gamma0 <- var(t.yields) * (n - 1)/n

> rho <- acf(t.yields, plot=F)$acf

> Var.mu <- n^(-2) * gamma0 * (n + 2*sum((n - 1:2)*rho[2:3])) This yields an estimated variance ofVar(d µ) = 1.643.b

The bounds of an approximate 95% confidence interval for the mean yield are then given by

µb±1.96·se(µ) =b µb±1.96· q

Var(d µ)b .

In our case, we get a confidence interval of[48.62,53.64].

If we assume independence, the variance ofµbis estimated as

Var(d µ) =b 1 n2

n

X

s=1

Var(Xd s) = bγ(0)

n = 1.997.

Under this independence assumption, therefore, an approximate 95% confidence interval for the mean yield is given by

[µb−1.96·se(µ),b µb+ 1.96·se(µ)] = [48.36,b 53.90].

Thus the correct specification of the independence structure here leads to a confidence interval which is10%narrower.

d) Only the first partial autocorrelation is significantly different from zero. Thus we can assume that our time series is an AR(1) process.

5 10 15

−0.40.0

Lag

Partial ACF

e) For an AR(1) model, the Yule-Walker equations simplify to

ρ(1) =b αb·ρ(0),b ρ(0) = 1b

σ2can be estimated by bσ2=bσX2 ·(1−αb2). Here, we getαb=ρ(1) =b −0.390 andσb2= 120.266.

Determining parameters with the help of R:

> r.yw <- ar(t.yields, method = "yw", order.max = 1)

> r.yw$ar [1] -0.3898783

> r.yw$var.pred [1] 122.0345

(3)

2. a) The experimental setup leads us to expect a period of 2 seconds. Since our measurements are spaced apart by 0.15 seconds, each 2-second period covers 2/0.15 = 13.3 measurements. This period, however, is subject to fluctuations, which are visible in both the time series plot and the correlogram of ordinary autocorrelations:

Time

ts.forceA

0 50 100 150 200 250

−0.50.5

0 5 10 15 20

−0.50.5

Lag

ACF

Series ts.forceA

5 10 15 20

−0.40.20.8

Lag

Partial ACF

Series ts.forceA

b) The PACF (see Part a)) is clearly significant for lags up to 6; also lags 9, 10 and 17 are slightly significant. We could therefore use an AR model of order 6, 9 or 17. Order 17 seems quite high (and hence difficult to interpret), so orders6or9would be preferred. However, we cannot see by eye whether order6or9 is really sufficient; this can be done with a residual analysis, see Part c).

The AIC plot yields similar proposes:

0 5 10 15 20

0200400600

Order

AIC

MLE Yule−Walker Burg

There is a big jump of the AIC at order6, and a smaller one at order9; hence one of these orders would be plausible to fit the given time series. However, the minimum of the AIC is attained at order p= 10(MLE),p= 17(Yule-Walker) orp= 17(Burg), respectively.

c) Fitting the AR model:

> p <- 9

> ar.force <- arima(ts.forceA, order = c(p, 0, 0), method = "ML") The residuals of this model look as follows:

> par(mfrow = c(2, 2), mar = c(3, 3, 2, 0.1))

> plot(ar.force$residuals, ylab = "residuals")

> acf(ar.force$residuals, type = "partial", plot = TRUE, main = "")

> plot(ts.forceA - ar.force$residuals, ar.force$residuals, xlab = "fitted values", ylab = "residuals", main = "Tukey-Anscombe plot")

> qqnorm(ar.force$residuals)

> qqline(ar.force$residuals)

(4)

Time

residuals

0 50 100 150 200 250

−0.3−0.10.10.3

5 10 15 20

−0.100.000.10

Lag

Partial ACF

●●

●●

● ●

● ●

● ●

●●

−0.5 0.0 0.5

−0.3−0.10.10.3

Tukey−Anscombe plot

fitted values

residuals

●●

●●●●

●●

●●

●●

●●

●●

● ●

●●

●●

●●

●●

●●

●●

−3 −2 −1 0 1 2 3

−0.3−0.10.10.3

Normal Q−Q Plot

Theoretical Quantiles

Sample Quantiles

The model with order 9 is acceptable. Its residuals are normally distributed, they have constant variance, and the correlogram as well as the Tukey-Anscombe plot do not indicate any dependence.

If we take order p= 6 instead of order p = 9, the residuals look different. They still show some (weak) correlation, indicating that the order is not sufficient for this time series:

Time

residuals

0 50 100 150 200 250

−0.3−0.10.1

5 10 15 20

−0.100.000.10

Lag

Partial ACF

●●

● ●

−0.5 0.0 0.5

−0.3−0.10.1

Tukey−Anscombe plot

fitted values

residuals

● ●

●●

●●●●

●●

●●

●●

●●

●●

●●

●●

● ●

●●

●●

● ●

●●

−3 −2 −1 0 1 2 3

−0.3−0.10.1

Normal Q−Q Plot

Theoretical Quantiles

Sample Quantiles

d) > force.pred <- predict(ar.force, n.ahead = 40)

> plot(window(ts.force, start = 250), ylab = "")

> lines(force.pred$pred, lty = 2)

(5)

> lines(force.pred$pred + 1.96*force.pred$se, lty = 3)

> lines(force.pred$pred - 1.96*force.pred$se, lty = 3)

Time

250 260 270 280 290 300 310 320

−0.50.00.5

Referenzen

ÄHNLICHE DOKUMENTE

a certain graph, is shown, and he wants to understand what it means — this corre- sponds to reception, though it involves the understanding of a non-linguistic sign;

Most reactions generated by the three different risk groups were related to the categories emotions, risk feedback valence, expectedness, and future lifestyle change, cf..

Die Produzenten von E-Only-Journals müssen sich den Herausforderungen der Langzeitar- chivierung elektronischer Ressourcen in verstärktem Maße stellen, da hier keine Archivie-

Then, fit the data using R and interpret your results.. A plot also shows that the new treatment improves response values after surgery, whereas the rates are unchanged with a

b) Due to the partial autocorrelations present, an AR(2) model for the residuals makes sense. Note that the ordinary autocorrelations make up a dampened sine curve, a property

d) All of the three methods seem to overestimate this particular time series a little. Probably exponential smoothing works best.. e)

This means if we calculate exp(r.pred$pred), we get the predicted values on the original scale. In Part b), however, the formation of differences needs reversing (by taking

1. a) An experimenter wishes to compare four treatments in blocks of two runs. Find a BIBD with six blocks. Find a BIBD with seven blocks. Analyze these data in a split plot anova..