• Keine Ergebnisse gefunden

Truncated Maximum Likelihood, Goodness of Fit Tests and Tail Analysis

N/A
N/A
Protected

Academic year: 2022

Aktie "Truncated Maximum Likelihood, Goodness of Fit Tests and Tail Analysis"

Copied!
43
0
0

Wird geladen.... (Jetzt Volltext ansehen)

Volltext

(1)

Goodness of Fit Tests and Tail Analysis

Christian Gourieroux Joanna Jasiaky This version: March 11, 1998 (First draft: September 28, 1997)

CRESTandCEPREMAP,e-mail:gouriero@ensae.fr.

y

YorkUniversity,e-mail:jasiakj@yorku.ca

(2)

Abstract

Truncated Maximum Likelihood, Goodness of Fit Tests and Tail Analysis.

We propose a new method of tail analysis for data featuring a high degree of lep- tokurtosis. Heavy tails can typically be found in nancial series, like for example, the stock returns or durations between arrivals of trades. In our framework, the shape of tails can be assessed by tting some selected pseudo-models to extremely valued observations in the sample. A global examination of the density tails is performed using a sequence of truncated pseudo-maximum likelihood estimators, called the tail parameter function (t.p.f.). In practice, data often exhibit local irregularities in the tail behaviour. To detect and approximate these patterns we introduce the local parameter function (l.p.f.), a method of tail analysis involving a selected interval of extreme observations. An immediate extension of the pseudo-value based approach to density analysis yields a new procedure for testing the goodness of t. We also de- velop a new nonparametric estimator of the density function. The method is applied to unequally spaced high frequency data. We study an intradaily series of returns on the Alcatel stock, one among the most frequently traded stocks on the Paris Stock Exchange.

Keywords: Goodness of t test, tail analysis, truncated maximum likelihood, kernel, Hill estimator.

(3)

Resume

Maximum de vraisemblance tronque, tests d'adequation et analyse de queues.

Nous proposons une nouvelle methode d'analyse des queues de distribution. Un e et leptokurtique important est generalement mis en evidence sur la plupart des series nancieres, qu'il s'agisse de series de rendements ou de durees entre echanges. Dans notre approche la forme des queues est etudiee en ajustant divers pseudo-models aux donnees extr^emes de l'echantillon. Un examen global des densites de queues peut ^etre e ectue en utilisant une suite d'estimateurs du pseudo-maximum de vraisemblance tronque, appelee fonction de parametre de queue t.p.f.]. Une approche alternative repose sur des analyses locales de la queue l.p.f.]. Ces analyses permettent aussi d'introduire des tests d'ajustement et un nouvel estimateur a noyau de la densite.

L'approche est appliquee a une serie de rendements du titre Alcatel, l'un de plus echanges a la bourse de Paris.

Mots cles: Tests d'adequation, analyse de queues, maximum de vraisemblance tronque, noyau, estimateur de Hill.

(4)

1 Introduction

Since the seminal work of Mandelbrot (1963) on cotton prices, the presence of heavy tails in the marginal distributions of nancial returns has been conrmed by a large number of empirical studies. This phenomenon has inspired a number of researchers seeking to provide a plausible explanation. Mandelbrot advanced a hypothesis of an underlying stable non-normal distribution of returns in a random walk model. Clark (1973) and Harris (1989) suggested that returns are generated by a mixture of normal distributions, whith the rate of new information arrival acting as a stochastic mixing variable. The heavy tails may also be viewed as a consequence of conditional heteroscedasticity, due to time varying volatility of returns Engle (1982)]. In the limiting case of strong persistence and presence of a unit root in the volatility equation, the marginal variance may even not exist Nelson (1990)]. The presence of heavy tails was also documented in the conditional distributions of returns for some nancial assets. This feature is crucial for risk management and should be accounted for in the risk control rules adopted by nancial institutions. Indeed, standard rules such as the Value at Risk ] are implicitely based on a mean-variance approach, and are inappropriate when data frequently admit extreme values. The leptokurtosis of conditional distributions may also explain some other characteristics of the series of nancial returns such as the long range of temporal dependence Resnick, Samorodnitsky (1997)].

In general, we say that a probability distribution features a heavy tail if asymptotically it behaves like the Pareto distributionPY > y] =y; L(y) >0 y >1, where L is slowly varying:

limt!1L(ty)L(t) = 1. The statistical methods for tail analysis focus on the estimation of the tail index . In the case of i.i.d. observations, the estimators are generally computed from the order statistics: Y(1) ::: Y(n). Two well-known estimators of are the Hill estimator Hill (1975)]:

1^ = 1k

k

X

i=1

log Y(i) Y(k +1)

and the de Haan's moment estimator de Haan (1970)]:

1^ = 1k

k

X

i=1

log Y(i) Y(k +1)

r:

In this paper we propose a new method of tail analysis. We introduce a parametric family of distributionsF =ff(y) 2gand determine the value of theparameter such thatf(y) is at the minimum distance from the empirical distribution of the observedy's, for values suciently largey c, or for y lying in a neighbourhood of largec. The tail behavior of the distribution is revealed by the behavior ofwith respect to c.

(5)

The formulas of estimators and test statistics, and their properties are derived under the as- sumption of i.i.d. observationsY1:::Yn. We present the asymptotic results under general regu- larity conditions.

In section 2 we introduce a partition of the real line into non overlapping intervals and allow the parameter to vary independently between the intervals. The comparison of the full sample maximum likelihood estimators of with the maximum likelihood estimators truncated over the intervals yields various goodness of t tests. In section 3 we consider the pseudo maximum likeli- hood (henceforth P.M.L.) estimation of theparameter from the data truncated byc. The limits of the P.M.L. estimators dene a function ofc, called the tail parameter function (t.p.f.). We show that the tail parameter function is constant only for a distribution in the F family. We discuss conditions under which the t.p.f. indicates the true underlying distribution. In section 4, we intro- duce an estimator of the t.p.f., derive its asymptotic distributional properties and discuss the links with the Hill estimator. In section 5 we develop a local pseudo-maximum likelihood method for estimation of density functions. The basic idea consists in smoothing the estimator of censored over an interval using a kernel. Next, the log-derivative of the unknown density function at a given point is approximated by the log-derivative of the pseudo density evaluated at the local estimator of. The empirical results are discussed in section 6, where various methods proposed in the paper are applied to a series of nancial returns. Some technical issues are are explained in Appendices 1,2 and 3. Section 7 concludes the paper.

2 Truncated Maximum Likelihood and Goodness of Fit Tests

2.1 Decomposition of the Log-Likelihood Function

Let us consider a family F of distributions on Y 2 R, parametrized by 2 Rp. The distributions inF have positive p.d.f. f(:) with respect to a dominating measure. We denote byF(:) andS(:) the corresponding cumulative function and survivor function, respectively.

For a random variable Y with the distribution function f(:), andA1:::AK a partition of

Y, we introduce the qualitative variableZ indicating the element of the partition containingY: Z = (Z1:::ZK)0 = 1A1(Y):::1AK(Y)]0 (2.1) where 1A(Y) = 1, ifY 2A and 1A(Y) = 0, otherwise.

The marginal distribution of the vectorZ is multinomialM(1 F(A1):::F(AK)), with the probability function PZ =z] =h(z). We denote by fk(y) the conditional density function ofY givenZk = 1, (i.e. Y 2Ak).

(6)

By applying the Bayes theorem we obtain the following decomposition of the conditional density function ofY:

log f(y) = logh(z) +XK

k =1

zk logfk(y) (2.2) which implies:

;

@2logf(y)

@@0 =;@2log h(z)

@@0 ;

K

X

k =1

zk @2logfk(y )

@@0 : (2.3)

The expectations of all terms in (2.3) with respect to the distribution f(:) yield the following decomposition of the information matrix:

JY() =JZ() +XK

k =1

JYjk() (2.4)

where

JY() = E;@2logf(Y)

@@0

JZ() = E;@2logh(Z)

@@0

JYjk() = E

;Zk@2log fk(Y)

@@0

:

Let us also introduce the matrix:

JYjZ() =JY();JZ() =XK

k =1

JYjk() (2.5)

which is the information matrix based on the conditional distribution ofY givenZ.

2.2 Extended Model

The model F may be nested in a larger modelF , allowing for an independent variation of the parameters appearing in the marginal distribution ofZ and in the conditional distributions ofZ givenY 2Ak. We now consider the modelF , where

the marginal distribution of Z ish(z),

the conditional distribution ofY givenY 2Ak isfk(yk),k= 1:::K,

1:::K are unconstrained with values in Rp.

(7)

The parameters of this extended model can consistently be estimated from a sampleY1:::Yn of i.i.d. variables by the maximum likelihood. The log-likelihood function to be maximized:

LY(1:::K) =Xn

i=1

logh(zi) +XK

k =1

(Xn

i=1

zk ilogfk(yik)) (2.6)

consists of two additive terms involving separately the parametersand. This allows for distinct optimizations with respect to these parameters and ensures an asymptotic independence of the estimators of and1:::K.

Property 2.1

:

LetY1:::Ynbe i.i.d. variables with a distribution inF . Under standard regularity conditions, the unconstrained maximum likelihood estimators ~~1:::~K are consistent, asymptotically normal, and asymptotically independent:

pn~;] !d N0JZ();1]

pn~k;k] !d N0JYjk(k);1] k= 1:::K:

2.3 Goodness of Fit Tests

In this subsection we present a procedure involving various pseudo-true value based tests to evaluate the t of the model. Let us introduce the hypotheses: H =f1:::K, unconstrained g, H0 =

f1=:::=Kg,H0=f1=:::=K =g.

Under the hypothesis H0, the common parameter value of and can be estimated by the maximum likelihood based on the entire sample ofnobservations:

^= argmax

n

X

i=1

log f(yi): (2.7)

UnderH0, this estimator is consistent and asymptotically normal:

pn(^;)!d N0JY();1]: (2.8) Under the hypothesis H0, two estimation methods can be considered. One consists again on maximizing the likelihood function and yields the estimator of =k k= 1:::K. Alternatively, we can use the equivalent approximation:

(8)

~ = XK

k =1

JYjk(~k)];1XK

k =1

JYjk(~k)~k: (2.9)

Under the hypothesis H0, ~ and ~ are consistent estimators of , asymptotically normal and independent, with the limiting variance matrices:

Vaspn(~;)] = JZ();1 (2.10) Vaspn(~ ;)] = JYjZ();1: (2.11)

Let us now consider the sequence of nested hypotheses:

H0H0 H =ff 2F g

and introduce test statistics H0jH, H0jH0, H0

jH for testingH0 against H, H0 against H0, H0 againstH respectively.

The test statistics can be formulated as the following likelihood ratios:

H0jH = 2LY(^);LY(~ ~1:::~K)]

H0jH0 = 2LZ(^);LZ(~)]

H0

jH = 2LYjZ(^);LYjZ(~1:::~K)]

whereLYLZLYjZ are the log-likelihood functions corresponding to the observations ofYZ and to the conditional model of Ygiven Z respectively. Standard results on likelihood ratio tests Gourieroux, Monfort (1997)] imply the following property:

Property 2.2

:

Under the null hypothesisH0, the statisticsH0jH0 andH0

jH are asymptotically independent, with asymptotic distributions: 2(p) and2((K;1)p), respectively.

The statisticH0jH =H0jH0 +H0

jH follows asymptotically a2(Kp) distribution.

It is possible to replace these test statistics by their asymptotic equivalents which are easier to compute. Such equivalent statistics are:

(9)

H0jH0 = (^;~)0^V(^) + ^V(^ )];1(^;~) (2.12) H0

jH = XK

k =1

(~k;~)0(^V(~k));1(~k;~ ): (2.13) The extended modelF may itself be nested in a more general modelF , where the marginal dis- tribution ofZis multinomial with parametersp1:::pKconstrained only by pk 08kPKk =1pk= 1, and where the conditional distributions ofY givenZ are as inF . The likelihood ratio statistic for testingH =ff 2F gagainstH =ff 2F gis asymptotically equivalent to the standard chi-square statistic of the goodness of t test. Under the null hypothesis H0, the statistic is asymptotically independent ofH0jH0,H0

jH,H0jH and follows the asymptotic2(K;p;1) dis- tribution. In some sense the statisticsH0jH0,H0

jH are complementary to the standard chi-square goodness of t test.

3 The Tail Parameter Function

In section 2 we showed the pseudo-true value based test statistics to assess the t of the model.

We now extend this approach and introduce a new method of tail analysis which instead of the pseudo-true values based on the entire sample involves tail based pseudo-true values.

3.1 Denition

Let us consider a parametric family of density functions: F =ff(:)2Rpg, with respect to the Lebesgue measure on Y = R+ = (0+1. We assume that the density functions in this family are positive.

For any probability distribution, with density function f0(:), we can write the optimization problem:

max

2

E0

1Yc log f(Y) S(c )

(3.1)

wherec2R+ andE0 is the expectation with respect tof0. We assume:

Assumption

A.1.

The optimization problem (3.1) has a unique solution(cf0F) for anyc 0.

The parameter value(cf0F) is such that the parametrized distributionf(:) provides the best approximation of the tail of the true distribution dened by the levelc.

Denition 3.1

Let us consider a distribution f0 satisfying A.1. The tail parameter function associated withf0 is the mapping(: f0F) fromR+ on .

(10)

For notational convenience we further denote the tail parameter function (t.p.f.) by (: f0), omiting F. Under standard regularity conditions, (cf0) is the limit of a pseudo maximum likelihood estimator based on the pseudo model F and a sample Y1:::Yn with distributionf0, and left truncated byc:

(c f0) = lim

n!1

^n(c) (3.2)

where:

^n(c) =argmax

2 n

X

i=1

f1yiclog f(yi)

S(c)g: (3.3)

3.2 Characterization of a Distribution of the Pseudo Model

In this subsection we show that the t.p.f. is constant only for distributions inF.

Property 3.2

:

Let us assume that the density functions are rst order dierentiable with respect toand y, and that the parameteris identiable. The densityf0 belongs to the familyF if and only if the tail parameter function is constant over c.

Proof

:

Necessary Condition

Iff0 belongs toF, there exists a unique parameter value0such thatf0(:) =f(:0), and the t.p.f. is(cf0) =08c:

Sucient Condition

Let the common solution of the optimization problems:

= argmax

E01Y>c log f(Y ) S(c ) ]

= argmax

Z

1

c

logf(y)f0(y)dy;S0(c)logS(c)]: The rst order conditions are:

Z

1

c

@logf(y )

@ f0(y)dy;S0(c)@logS(c )

@ = 0 8c:

By dierentiating again with respect to the truncation level c we get:

;

@logf(c )

@ f0(c);dS0(c)

dc @logS(c )

@ ;S0(c)@2logS(c )

@c@ = 0 8c:

f0(c)

;

@log f(c )

@ +@logS(c )

@

;S0(c)@2logS(c )

@c@ = 0 8c:

(11)

f0(c)

S0(c) = @2logS(c )

@c@j

;

@logf(c )

@j +@logS(c )

@j

;1 8c j = 1:::p:

The hazard function corresponding to f0 is uniquely dened by f(c ) and since the density functionf(c ) also satises the last equation, we conclude thatf0(:) =f(: ).

Q.E.D.

3.3 Identifying a Distribution by its Tail Parameter Function

In this subsection we focus on two important questions:

(i) Does there exist pseudo-familiesF such that(: f0) characterizes anyf0?

(ii) For a given pseudo-family F, what condition satisfy the distributions identiable by their t.p.f. ?

We give below some preliminary answers to these questions.

Property 3.3

:

If Y is a positive variable and the pseudo-family is the family of exponential distributions:

F=ff(y) =exp(;y) 2R+g, then

(i) the t.p.f. is the expected residual life : (cf0) =E0((Y ;c)jY > c) (ii) the t.p.f. characterizes the distribution.

Proof

:

(i) The optimization problem max

fE01Y>c log];E01Y>c(Y ;c)]g involves the rst order condition:

1E0(1Y>c);E01Y>c(Y ;c)] = 0 and yields the solution:

(cf0) =E0((Y ;c)jY > c): (ii) Conversely we get:

(cf0) = E01Y>c(Y ;c)]

E0(1Y>c)

=

R

1

c (y;c)f0(y)dy S0(c)

=

R

1

c S0(y)dy S0(c) :

(12)

Therefore (cf0);1 = ;dcd log Rc1S0(y)dy], and, when (:f0) is known, the survivor function can be found by integration.

Q.E.D.

Property 3.4

:

There exists a one to one relationship between the distributionf0and its t.p.f. in a neighbour- hood of the parametric pseudo-modelF.

Proof

: The rst order condition dening(cf0) is:

8c:

Z

1

c

f0(y) @

@logfy(cf0)]dy;S0(c)@logS

@ c(cf0)] = 0:

Let us consider a distribution f0 close tof(: ) and introduce (cf0) = (cf0);cf(:0)]

=(cf0);0.

A rst order expansion yields:

8c:

Z

1

c

f0(y);f(y0)]@

@logf(y0)dy;S0(c);S(c0)] @

@logS(c0) + (cf0)fZ 1

c

f(y0) @2

@2logf(y0)dy;S0(c) @2

@2logS(c0)g= 0:

Consider now a given discrepancy function (:f0). By dierentiating the previous relation with respect toc we obtain a dierential equation of the type:

f0(c);f(c0)]A(c);S0(c);S(c0)]B(c) +C(c) = 0

where ABC are unknown functions. This is a rst order linear dierential equation for the functionS0(:);S(: 0). The set of solutions is ane of dimension 1, and the integration constant is uniquely determined by the initial condition: S0(;1);S(;1 0) = 1;1 = 0.

Q.E.D.

The previous property is a local identiability condition of the true density by its t.p.f.

3.4 The Form of the Tail Parameter Function in a Neighborhood of the Pseudo-Family

In this section we show how to evaluate the discrepancy function (:f0) introduced in the Property 3.4, in a neighborhood of the pseudo-family. The property 3.5 below follows directly from the expansion of the rst order condition and corresponds to some standard properties of the maximum likelihood estimators under local alternatives.

Property 3.5

:

(13)

Let us consider a parametric family: ~F=ff(:)()2Bg, and assumeF =ff(:) = f(:0)2 g, andf0=f(:0) with small . Then (c f0) =0 + ~J(c), where ~J(c) is the cross term of the information matrix corresponding to the model F truncated at c and evaluated at (00).

4 Tail Analysis

In this section we present the t.p.f. method of tail analysis. We begin with the null hypothesis concerning the tail of the distribution:

H~0=f9c0:f0(y) =f0(y0) 8y c0g:

This hypothesis may be tested by applying the procedure presented in section 2, based on a partition of the real line. Else, for an increasing sequence ofc it can be veried if the consistent estimator of the t.p.f. remains approximately constant for largec. The second approach is similar to some classical methods of tail analysis. It provides exploratory plotting techniques, but should be used with caution. Its limitation results from a strong correlation of the estimated values of the t.p.f. at dierent points, due to the overlapping intervals.

4.1 Statistics Based on the Truncated Pseudo Maximum Likelihood

Let us consider a sample of i.i.d. variables Y1:::Yn with an unknown distribution f0. We are interested in testing the null hypothesis ~H0. Since the t.p.f. is constant forc c0, it is natural to expect that a consistent estimator of(: f0) is constant as well. Such an estimator is:

^n(c) =argmax

n

X

i=1

1yic logf(yi)];logS(c)Xn

i=1

1yic (4.1)

cvarying. This functional estimator is a stepwise function with jumps at each observed value. To see that, let us introduce the order statistics corresponding to the observations:

y(1) y(2) :::: y(n): The jumps arise at the valuesy(k +1),kvarying, and we get:

^nk = ^n(y(k +1)) (4.2)

= argmax

k

X

i=1

logf(y(i) );logS(y(k +1))k: (4.3)

(14)

Example 4.1: When the pseudo-family F is exponential family F = ff(y) = exp(;y), 2R+g, the estimator is:

^1 nk = 1k

k

X

i=1

y(i);y(k +1)]:

Example 4.2 : When the pseudo-family is Pareto family, F =ff(y) =y;;1 2 R+g, we get:

^1 nk = 1k

k

X

i=1

log

y(i) y(k +1)

which is the Hill estimator Hill (1975)], with properties in the i.i.d. case discussed by Mason (1982).

4.2 Asymptotic Properties of the Estimated t.p.f. under the Null Hy- pothesis

The asymptotic behavior of ^n(c1):::^n(cJ) for a nite set of truncation points: cj j = 1:::J with cj c08j follows from the standard results on the maximum likelihood estimators. We denote byJ(c0) the information matrix of the model truncated by c.

Property 4.1

:

Under the null hypothesis ~H0:

(i) the estimator ^n(c1):::^n(cJ)]0 converges to (0:::0) and is asymptotically normal:

pn^n(c1);0:::^n(cJ);0]0!d N(0), where =A;1BA;1, A=

2

6

4

J(c10) 0 0 ... J(cJ0)

3

7

5

B= J(sup(ci cj)0)]: (ii) In particular, for a truncation pointc we get:

pn^n(c);0]!d N0J(c0);1]:

5 Local P.M.L. Analysis of the Density Function

5.1 The Estimation Method

An alternative application of the approach based on approximating the unknown density function f0 by f(: ~k) on the interval Ak (see, section 2) is a non parametric estimation of the density

(15)

functionf0. Let us consider an intervalA= c;h c+h]. The approximation of theparameter is the solution of:

~ch = argmax

E01c;h<Y<c+hlog f(Y )];E01c;h<Y<c+h] log

Z

1c;h<y<c+hf(y)dy

= argmax

E0

1

2h1c;h<Y<c+h logf(Y)

;E0

1

2h1c;h<Y<c+h

log

Z 1

2h1c;h<y<c+hf(y)dy ~ch=argmax

E0

1

hK Y ;c h

log f(Y)

;E0

1

hK Y ;c h log

Z 1

hK y;c h

f(y)dy (5.1) whereK(u) = 121;11](u).

From the last formula, we derive a non parametric estimation method for the density function f0. We consider a kernel satisfying the following assumptions:

Assumptions concerning the kernel:

A.2

RK(u)du= 1

A.3

RK(u)udu= 0

A.4

RK(u)u2du=2, (exists).

We propose to use the estimator dened by : ~n(c) =argmax

"

n

X

i=1

hK1 yi;c h

log f(yi);Xn

i=1

hK1 yi;c h

logZ 1

hK y;c h

f(y )dy

#

(5.2) to estimate the p.d.f. f0from the pseudo distribution after substituting ~n(c) for. This approach is based on a local approximation of the p.d.f. by an element of the pseudo family F and is an analogue of the locally weighted least square regression developped by Cleveland (1979) and Hardle (1990) in the context of non parametric estimation of the regression function see also Gourieroux, Monfort, Tenreiro (1994) for the extension to kernel M-estimators].

5.2 The Gaussian Pseudo-family

As an illustration of the method described above we examine a special case where the estimator ~n(c) admits an explicit form and can easily be interpreted. Let us consider a gaussian pseudo- family with the mean :

(16)

f(y) =(y;)

where is the p.d.f. of the standard normal distribution. Let us also use a gaussian kernel:

K(:) =(:). We get:

Z 1

h y;c h

f(y)dy = Z 1

h c;y h

(y;)dy

= 1

p1 +h2 c;

p1 +h2

and the estimator is:

~n(c) = argmax

n

X

i=1

h 1 yi;c h

;

12 log 2;(yi;)2 2

; n

X

i=1

h 1 yi;c h

;

12 log 2;12 log(1 +h2); (c;)2 2(1 +h2)

: It is equal to :

~n(c) = 1 +h2 h2

P

n

i=1 1

h;yih;cyi

P

n

i=1 1

h;yi;ch ;h12c:

We nd that:

~n(c) = 1 +h2

h2 m~n(c); 1 h2c where :

m~n(c) =

;P

n

i=1 1

h;yi;ch yi

;P

n

i=1 1

h;yih;c

is the Nadaraya - Watson estimator of the regression function: m(c) =EYjY =c] =c, Nadaraya (1964), Watson (1964) ].

Moreover:

~n(c);c = 1 +h2

h2 ( ~mn(c);c)

= 1 +h2 h2

P

n

i=1 1

h(yi;c);yi;ch

P

n

i=1 1

h;yi;ch

= (1 +h2)@c@

P

n

i=1 1

h;yih;c

P

n

i=1 1

h;yi;ch

= (1 +h2)@log

@c f~n(c)

(17)

where ~fn(c) is the (gaussian) kernel estimator of the density functionf0. Then, the asymptotic properties of ~n(c);c are deduced directly from the asymptotic properties of ~fn(c) and @f~@cn(c) Silverman (1986)]. In particular ~n(c) converges toc+@@clogf0(c) when ntends to innity andh tends to zero at appropriate rates.

5.3 The Local Parameter Function

Before presenting the regularity conditions ensuring the consistency of the local pseudo-maximum likelihood estimator when the number of observationsn tends to innity whereas the bandwidth htends to zero, we derive the possible limits for ~ch whenhtends to zero.

Property 5.1

:

Let us assume the conditions A1-A4 and

A.5

The density functionsf(y ) andf0(y) are positive and third order dierentiable with respect to y.

A.6

For h small and any c, the following integrals exist: R K(u)logf(c+uh)f0(c+uh)du,

R K(u)f0(c+uh)du, RK(u)f(c+uh)du, and are twice dierentiable under the integral sign with respect toh.

i) Whenhtends to zero, the objective function : Ah() =E0hK1 Y ;c

h

logf(Y)

;E0hK1 Y ;c h

log

Z 1

hK y;c h

f(y)dy is equivalent to :

Ah() = 2h2 2

"

2@logf(c)

@y @logf0(c)

@y ;

@logf(c)

@y

2

#

+o(h2):

ii) The local parameter function (l.p.f.) ~(cf0) = limh!0~ch is the solution of the equation:

@logf(c ~(cf0))

@y = @logf0(c)

@y :

Proof

:

i) is shown in Appendix 1.

ii) The objective function is a locally quadratic function with respect to

@log f(c)

@y

which implies the rst order condition:

(18)

@logf(c)

@y =@log f0(c)

@y : Q.E.D.

We deduce from the previous property the following corollary:

Corollary 5.1

The local parameter function characterizes the distribution.

Proof:

Indeed, if ~(:f0) is known, we also know the log-derivative of the density function, since

@logf0(c)

@y = @logf(c ~(cf0))

@y 8c and by integration we nd the density functionf0itself.

Q.E.D.

Corollary 5.2

The local parameter function is constant if and only if the distribution belongs to the pseudo-family

F.

Proof:

Iff0(y) =f(y0), we see immediately that ~(cf0)) =0 is constant. The converse part results from Corollary 5.1.

Q.E.D.

The Corollary 5.1 may be used to derive the functional estimators of the log-derivative of the density function. Indeed if ~n(c) is a consistent functional estimator of ~(cf0), then

@log f(c ~n(c))

@y

is a consistent functional estimator of @log@yf0(c), whereas @log@yf(c) is continuous with respect to . Also, the Corollary 5.2 may be used for testing the goodness of t in the family F.

Example 5.1:

(19)

For the gaussian familyF=ff(y) =(y;) 2Rg, we have:

@logf(c)

@y = @log(c;)

@y =;c:

We deduce that:

~(cf0) =c+@logf0(c)

@y : Q.E.D.

5.4 Local Parameter Function and Tail Index

In this section we show that the local parameter function can be used to nd the tail index >0 of the true unknown distribution f0. From the Karamata's theorem Karamata (1962), Resnick (1987), Corollary 1.12], the true distribution is such that : 1;F0(y) =y; L(y) y >1, with >0, if and only if :

1;F0(y) =c(y)exp;Z y

1

(t)

t dt (5.3)

where thecand functions satisfy:

lim

y!1

c(y) =c >0 lim

y!1

(y) = >0: (5.4)

Let us assume thatcand are twice dierentiable and satisfy the additional limiting conditions:

lim

y!1

yc0(y) = lim

y!1

y2c0(y) = lim

y!1

y0(y) = 0: (5.5)

We get from (5.3):

;y@logf0(y)

@y =(y) + 1 + yc0(y)(y) +c(y)y0(y);yc0(y);y2c00(y) c(y)(y);yc0(y) which implies:

lim

y!1

;y@logf0(y)

@y =+ 1: (5.6)

(20)

Therefore the tail index may be directly derived from the limiting behaviour of;y@log@yf0(y). Let us now consider the asymptotic behaviour of the local parameter function, which clearly depends on the tail behaviour of the pseudo-familyf(y).

i) If any distribution f(y) admits a tail index() in a one to one continuous relationship with, we deduce from the equation dening ~(cf0) that:

;c@fc ~(cf0)]

@y =;c@logf0(c)

@y

and by taking the limit whenc tends to innity we nd that:

lim

c!1

~(cf0)] + 1 = lim

c!1

~(cf0)] + 1 =+ 1:

Therefore the local parameter function tends to a limit such that the tail index of the estimated pseudo-distribution coincides with the index of the true distribution.

ii) For a family of distributions with thin tails, the term ~(cf0) may diverge to innity when ctends to innity. The examples below illustrate this feature.

Example 5.2: If the pseudo-family is the gaussian family indexed by the mean, we get:

;y@logf(y)

@y =y(y;) and by Property 5.1:

~(yf0) =y;1yy@logf0(y)

@y y

for largey.

5.5 Asymptotic Properties of the Local Pseudo-Maximum Likelihood Estimator

The asymptotic properties of the local P.M.L. estimator ofare derived along the following lines.

We rst nd the asymptotically equivalent formula of the objective function and of the estimator, which only depend on a limited number of kernel estimators. Then, we deduce the properties of the local P.M.L. estimator from the properties of these basic kernel estimators. We only detail the additional assumptions which are necessary for the asymptotic equivalence to hold, since the set

(21)

of assumptions for the existence and asymptotic normality of the basic kernel estimators are quite standard see Parakasa-Rao (1983), Bosq- Lecoutre(1987), Hardle (1990)].

Property 5.2

The local pseudo-maximum likelihood estimator ~n(c) exists and is a strongly consistent estimator of the local parameter function ~(cf0) under A.1 - A.6 and the following additional assumptions:

A.7

: the parameter set is an open set

A.8

: there exists a unique solution in of the equality:

@logf(c)

@y =@log f0(c)

@y

A.9

: the following kernel estimators are strongly consistent:

(i) n1Pni=1h1K;yi;ch a:s:! f0(c)

(ii) 1nPni=1h1K;yi;ch ;yi;ch 2a:s:! 2f0(c)

(iii) nh1 Pni=1 1hK;yih;c;yih;ca:s:! 2f0(c)@log@yf0(c)

A.10

In any neighbourhood of, the third order derivative @3logf(y)]=@y3 is dominated by a functiona(y) such thaty3a(y) is integrable.

Proof:

See Appendix 2(i).

Property 5.3

Under assumptions A.1 - A.10 the local pseudo-maximum likelihood estimator is asymptotically equivalent to the solution ~~n(c) of the equation:

@log fc ~~n(c)]

@y = 12h12( ~mn(c);c) where:

m~n(c) =

P

n

i=1K;yi;ch yi

P

n

i=1K;yi;ch

is the Nadaraya - Watson estimator ofm(c) =E(YjY =c) =c based on the kernelK.

Proof:

(22)

See Appendix 2(ii).

Therefore the asymptotic distribution of ~n(c) may be deduced from the standard properties of m~n(c);c. We only consider below the pointwise convergence in distribution A functional theorem for ~n(:) may only be derived if there exists a functional theorem for ~mn(c) at second order].

Under standard regularity conditions Prakasa-Rao (1983), Hardle (1990), Bosq, Lecoutre (1987)] the numerator and denominator of 1=h2( ~mn(c);c) have the following asymptotic proper- ties:

A.11

: Ifn!1,h!0,nh3!1,nh5!0, we have the limiting distribution:

pnh3hnh13Pni=1K;yih;c(yi;c);2@f0(c)@y i

pnhnh1 Pni=1K;yi;ch ;f0(c)

!

d

! N 0f0(c)

R

u2K2(u)du R uK2(u)du

R uK2(u)du RK2(u)du

:

The formulas of the rst and second order asymptotic moments are easily veried see Appendix 3].

The rate of convergence of the numerator is slower than the rate of convergence of the denominator since we study the degenerate case, where the Nadaraya-Watson estimator is applied to a regression with regressor equal to the regressand. We deduce that the asymptotic distribution of

pnh3 h12 ~mn(c);c];2@logf0(c)

@y

coincides with the asymptotic distribution of

pnh3f01(c)

1 nh3

n

X

i=1

K yi;c h

(yi;c);2@f0(c)

@y

!

i.e. Nh0 f0(c)1 Ru2K2(u)dui.

We deduce directly by the -method the asymptotic distribution of the local pseudo maximum likelihood estimator and of the log derivative of the true p.d.f.

Property 5.4

Under assumptions A.1 - A.11 we have:

i)

pnh3

@logf(c ~~n(c))

@y ;@logf0(c)

@y

!

d

!N0 4f10(c)

Z u2K2(u)du:

(23)

ii)

pnh3~~n(c);~(cf0)!d N

2

40

@2logf(c ~(cf0))

@y@

!

;2 1 4f0(c)

Z

u2K2(u)du

3

5: Remark 5. : The functional estimator of the log-derivative @log@yf0(c) may be compared to the standard one:

@log ^f0(c)

@y =

P

n

i=1 1

hK0;yih;c

P

n

i=1 1

hK;yi;ch : The rate of convergence is identical and the asymptotic distribution is:

pnh3

@log ^f0(c)

@y ;@logf0(c)

@y

!

d

!N0 4f10(c)

Z K0(u)2du:

The asymptotic distributions of the two kernel estimators of the density function are identical if

jK0(u)j=juK(u)j, in particular for a gaussian kernel.

6 Empirical Results

We examine a series of returns for the Alcatel stock covering the period of July and August 1995. Alcatel belongs to the most frequently traded stocks on the Paris Stock Exchange (Paris Bourse). On average the shares of Alcatel are exchanged every 52 seconds. The data consist of 20502 observations recorded intradaily with an accuracy of 1 second. Figure 1 shows the dynamic pattern of the Alcatel series. Due to irregular spacing, the conventional time scale is replaced in Figure 1 by an axis ordered by indices of subsequent trades, or equivalently numbers of observations in the sample. The Alcatel returns feature a tendency for clustering and a time varying volatility, while potential deviations of the mean return from zero can visually not be distinguished. Figure 2 displays the marginal density of returns and indicates the examined tail area. The density function is centered at 4.259E-6, while the variance and standard deviation are 9.771E-7 and 0.000989, respectively. The distribution is slightly asymmetric, with the skewness coecient -0.00246. The high kurtosis 5.329689 is due to the presence of heavy tails stretched between the extreme values of -0.00813 and 0.007255. 90 % of the probability mass is concentrated between -0.0017 and 0.001738.

The interquartile range 0.000454 is 100 times smaller than the overall range 0.01538. The shape of tails, although kernel smoothed in Figure 2, suggests some local irregularities in the rate of tail decay. Slight lobes can clearly be distinguished in both tails. In the right tail we observe a higher probability of returns taking values between 0.0012 and 0.0013 compared to the probability of those of a slightly smaller size, i.e. between 0.0010 and 0.0012. In the left tail we recorded relatively more returns between -0.0014 and -0.0012 than those taking marginally higher values.

(24)

We perform the tail analysis using three pseudo-models:

1) a gaussian pseudo-model, 2) a Pareto pseudo-model,

3) an exponential pseudo-model applied to Box-Cox transformed data.

In each case, we apply the methods proposed in the paper. The rst method consists in comput- ing the pseudo maximum likelihood estimator over the interval cj1) for an increasing sequence (cj) j = 1:::n. The second method involves maximization of the pseudo likelihood function over disjoint intervals cj cj+1] j = 1:::n. These procedures yield estimators of the t.p.f. and the l.p.f., respectively. We present below the optimization criteria for the estimation of the t.p.f for the three pseudo-models considered.

1)

gaussian family :

The t.p.f. estimator consists of a sequence of solutions (c2c), c=cj j= 1:::nof the following maximization problems:

max

n

X

i=1

1yi>cj ;1

2 log2;12 log2;(yi;)2

22 ;log ;c

forcj j= 1:::n.

To obtain a concave objective function, we introduce the change of parameters: a= 1= b==.

2)

Pareto family :

For logYi(1) the t.p.f is obtained by solving:

max

n

X

i=1

1yi>cjlog;(logyi;cj)]: The estimator is:

1^ =

P

n

i=11yi>cj(logyi;cj)

P

n

i=11yi>cj :

3)

Box - Cox transformed exponential variables :

We introduce a transformed variable X = (Yi;1)=(1 ). The survivor function is dened by the probability:

S(y) = exp

; yi;1

while the density of the pseudo model is:

f(y) =yi;1exp

; yi;1

:

Referenzen

ÄHNLICHE DOKUMENTE

Scottish Vowel Length Rule (SVLR), prosodic timing, sound change, dialect contact, the Voicing Effect, real-time change, Scottish English, Glaswegian vernacular..

In this Chapter, we attempt to generalize the theory of score tests. The situation is similar to the one in estimation theory. There is a classical estimation method based on the use

Die hier vorgestellten Arbeiten lassen sich unter drei breit gefasste Themenfelder subsumieren: Die ersten drei Beiträge thematisieren die Bedeutung des Kontextes für

(eds.), German Reunification: A Multinational History (London: Routledge, 2016); Gehler/Graf, Europa und die deutsche Einheit... Injection of Nuance Our report injects nuance

By compar- ing with the results for a more tranquil period of time, we conclude that in the context of the financial crisis the Romanian stock market became more sensitive to the

Example countries are highlighted in panel (b) of Figure 3, with some changing positions: Russia overtakes the United States for example, and China moves up the Lorenz curve

Concerning the one investment portfolio, whereas the London Stock Exchange (LSE) was undoubtedly leader in terms of capital markets as a whole - including money

The model incorporates four key factors that give rise to populist stabilization cycles: (i) various coalitions are formed and later coalitions are shifted (ii) some of the