• Keine Ergebnisse gefunden

A Tools

The description of statistical methods in this appendix is - with kind permission - mainly taken from KAUKER (1998). Unless other references are given in the text it is based on the publication by VON STORCH AND ZWIERS (1998).

A.1 Empirical Orthogonal Functions

Principal Component Analysis or Empisical Orthogonal Function (EOF) Analysis as it is called in geosciences was introduced into meteorology by LORENZ (1956). Its puspose is to derive dominant pattems of variability from a statistical field, which in practice is equivalent to a data reduction. This is achieved by converting data Sets with many degrees of freedom into data sets with less degrees of freedom, while keeping the most of their variability. A key aspect for use in geosciences is that EOFs allow for an analysis of interdependencies within investigated data sets.

A principal component analysis identifies a series of orthogonal pattems

2,

called the Ernpirical Orthogonal Functions (EOFs), which for any number K minirnize the mean- squared esi-OS

where t usually represents time and

X

is a multivariate random vector. Because of the orthogonality of the EOFs, the optimal coefficients a k ( t ) are given by the projection of the EOF Z k onto the random vector

2 :

These coefficients are called EOF coefficients, or principal components. They are pair- wise uncosselated.

It can be shown that the EOFs are the eigenvectors of the sample covariance matrix and that the EOF coefficients are statistically independent if the involved distributions are Gaussian.

If the EOFs are norrnalized, the total variance of the vector time series

2

may be decomposed into independent contributions from the EOFs

i.e., the EOFs form a orthogonal basis of the random vector

2.

A.2 Canonical Correlation Analysis

In certain problems, it is useful to identify pairs of pattems in two observed or simulated random vectors simultaneously. The Caizonical Correlation Analysis (CCA) is a tech- nique that is based On optimizing the cosselation between patterns (HOTELLING, 1936).

The CCA decomposes two random vectors X f t ) and ? ( t ) (2?(t) and Y ( t ) are a~zomalies, i.e., the time means have been subtracted prior to the analysis) into K patterns

k = l

where f l , and

qk

are the kth CCA patterns and a k ( t ) and

Mt)

are the kth CCA time series. Note that in general X ( t ) and Y ( t ) and therefore f l k ( t ) and { ( t ) have different spatial dimensions. The pattems and time series fulfill the following constraints:

The coefficients a k ( t ) and à Ÿ k ( t are optimal in the sense that they minimize the equations

The correlations between a k and 0 . 1 , between à Ÿ and à Ÿ l and between ai, and ß are Zero for all k # 1 .

eil and à Ÿ have the maximal possible correlation.

o;2 and ß have the next highest cosselation by being orthogonal to the forrner pair, and so On.

An important caveat to keep in mind is the method's intrinsic tendency to retum overesti- mated cosselation coefficients from a finite sample. Often a EOF truncation is perforrned to reduce the dimensionality of the calculation. The results may depend On the EOF truncation of the data.

A.3 Repression Analvsis

A.3 Regression Analysis

Regression Analysis is used here to investigate the link between two concurrent events as well as between time-lagged events. The approach used is to calculate a regression for every (spatial) point xj between a time series a ( t ) and a random vector X ( x , t ) . The

4

analysis aims at determining a time-independent vector Y ( x ) - the regression coefficient - under the constraint of minimizing

To calculate lagged regressions the discrete time steps of a ( t ) and X ( x , t) are delayed against each other. The correlation coefficient of this regression is deterrnined by:

where a - ( z ( x , t ) ) denotes the standard deviation of X ( x , t ) .

A.4 Principal Oscillation Patterns

The basic idea is to identify a linear system with a few degrees of freedom from a complex System. Consider a discretized linear system

X ( t

+

1) = A

.

X ( t )

+

noise, (A.10)

where

2

is a n-dimensional random vector. The system matrix A is a time-independent n X n matrix. Multiplication of equation (A.lO) from the right-hand side by the transposed of

X(t)

and taking expectations leads to

A = C o v ( X ( t

+

l ) , Y ( t ) )

.

C o v - ' ( 2 ( t ) , X ( t ) ) , (A. 11) i.e., A is the product of the sample lag+l covariance matrix and the inverse of the sample covariance.

The eigenvectors

fi

of A are called the Principal Oscillation Patterns (POPS). In gen- eral, A is not syrnmetric, i.e., its eigenvalues A are complex. The discussion is restricted to the case of complex eigenvalues. If A is not degenerated, the eigenvectors form a linear basis

(A. 12) with z j ( t ) being the POP coefficients. Inserting equation (A.12) into equation (A.lO) yields (the index j is omitted)

z ( t

+

1) -@'= Az(t)

.@'+

noise, (A. 13)

so that, if z ( 0 ) = 1,

Figure A.l: Schematic diagram of the time evolution of POP coefficients z ( t ) with an initial value z(0) = ( 0 , l ) . z rotates counter-clockwise in one period T around the origin. The e-folding time

T , for which z ( t ) = 1/e, is marked by an open circle. From V O N STORCH ET AL. (1990).

The contribution P ( t ) of the POP p t o the process X ( t ) is the real part of z ( t )

5

with * representing the complex conjugate Operator. Splitting pinto the real part p" and imaginary partpi and writing 2 z ( t ) = z r ( t ) - i z i ( t ) yields (setting z ( 0 ) = 1)

P{f)

= z ( t ) r q T + z @ y . f ? (A. 16)

= pt(cos(rlt)

-

p" - sin(r]t)

.pi

), (A.17) with

A

= pexp(ir]). The so-called reconstructed POP of chapter 6 is the contribution P ( t ) of the first POP, in practice "reconstructed" from the real and imaginary parts of the POP and their coefficients as in equation (A. 16).

The geometric meaning of equation (A.17) is a trajectory of a spiral in the space spanned by p' and

6'

(Fig. A.1) with the period T = 271-/n and the e-foldiizg time T = - l / l n ( p ) . The trajectory passes throughp" a n d g i in the consecutive order

. Pr -+ -5'

->

-F

->

5'

4 P,?

. . .

(A. 18) The e-folding time or damping time r is the time in which the amplitude is damped to 1 / e th of the initial amplitude.

The Pattern coefficients z,(t) are given as the projection of the random vector

X

onto

the adjointpatterns

gA,

which are the normalized eigenvectors of the adjoint matrix AA:

The adjoint patterns can be estimated from the adjoint matrix A A . This is not always a stable procedure as the eigenvectors

G A

of the matrix A~ have to be orthogonal to the eigenvectors p, i.e., have to satisfy the constraint

< >=

Sã Because some of the POPS represent noise, it is often advisable to estimate the patterns

PA

by minimizing

A.4 Principal Oscillation Patterns

An alternative is not to derive adjoint Patterns but to derive the coefficients z by a least- Square fit of the data by minimizing

The auto-spectrum F.; of the POP coefficient z ( t ) is a function the eigenvalue A and of the auto-spectrum of the noise I?,, (see equation A.13):

If the noise spectrum F,, is white, the temporal statistics of the signal are determined by the eigenvalue A. In this case, the width of the spectrum depends only on p . The smaller the value of p the broader the spectrum. The spectrum has a single maximum,

rz

= (1 - P)-', at W = 7.

Appendix B