SEMIPARAMETRIC PARTIALLY LINEAR
ERRORS-IN-VARIABLES MODEL
Hua Liang, Wolfgang Hardle and Raymond J. Carroll April 4, 1997
Abstract
We consider the partially linear model relating a response Y to predictors (X T) with mean function XT +g(T) when the X's are measured with additive error. The semiparametric likelihood estimate of Severini and Staniswalis (1994) leads to biased estimates of both the parameterand the functiong( ) when measurement error is ignored. We derive a simple mod- ication of their estimator which is a semiparametric version of the usual parametric correction for attenuation. The resulting estimator of is shown to be consistent and its asymptotic dis- tribution theory is derived. Consistent standard error estimates using sandwich{type ideas are also developed.
Key Words and Phrases:
Errors-in-Variables Functional Relations Measurement Error Non- parametric Likelihood Orthogonal Regression Partially Linear Model Semiparametric Models Structural Relations.Short title:
Partially Linear Models and Measurement Error.AMS 1991 subject classication:
Primary: 62J99, 62H12, 62E25, 62F10 Secondary: 62H25, 62F10, 62F12, 60F05.Hua Liang is Associate Professor of Statistics, Institute of Systems Science, Chinese Academy of Sciences, Bei- jing 100080, China. Wolfgang Hardle is Professor of Econometrics, at the Institut fur Statistik und Okonometrie, Humboldt-Universitat zu Berlin, D-10178 Berlin, Germany. Raymond J. Carroll is Professor of Statistics, Nutrition and Toxicology, Department of Statistics, Texas A&M University, College Station, TX 77843{3143. This research was supported in part by Sonderforschungsbereich 373 \Quantikation und Simulation Okonomischer Prozesse". The rst author was supported by Alexander von Humboldt Foundation. The third author was supported by a grant from the National Cancer Institute (CA 57030) and by the Alexander von Humboldt Foundation. The authors would like to thank Stefan Prot for his comments.
Consider the semiparametric partially linear model based on a sample of size
n
,Y
i =X
Ti+g
(T
i) +i (1) whereX
iis a possibly vector{values covariate,T
i is a scalar covariate, the functiong
( ) is unknown, and the model errors i are independent with conditional mean zero given the covariates. The partially linear model was introduced by Engle, et al. (1986) to study the eect of weather on electricity demand, and further studied by Heckman (1986), Chen (1988), Speckman (1988), Cuzick (1992a,b), Hua & Hardle (1997) and Severini & Staniswalis (1994).We are interested in estimation of the unknown parameter
and unknown functiong
( ) in model (1) when the covariatesX
are measured with error, and instead of observingX
, we observeW
i=X
i+U
i (2)where the measurement errors
U
i are independent and identically distributed, independent of (Y
iX
iT
i), with mean zero and covariance matrix uu. We will assume that uu is known, taking up the case that it is estimated in section 5. The measurement error literature has been surveyed by Fuller (1987) and Carroll, et al. (1995).If the
X
's are observable, estimation of at ordinary rates of convergence can be obtained by a local{likelihood algorithm, as follows. For every xed , letg
b(T
) be an estimator ofg
(T
). For example, in the Severini and Staniswalis implementation,g
b(T
) maximizes a weighted likelihood assuming that the model errorsi are homoscedastic and normally distributed, with the weights be- ing kernel weights with symmetric kernel density functionK
( ) and bandwidthh
. Having obtainedb
g
(T
), is estimated by a least squares operation:minimize Xn
i=1
n
Y
i;X
Ti;bg
(T
i)o2:
In this particular case, the estimate for
can be determined explicitly by a projected least squares algorithm. Letg
byh( ) and bg
xh( ) be the kernel regressions with bandwidthh
ofY
andX
onT
, respectively. Then bx="
n
X
i=1
f
X
i;bg
xh(T
i)gfX
i;g
bxh(T
i)gT#
;1Xn i=1
f
X
i;bg
xh(T
i)gfY
i;g
byh(T
i)g:
(3) One of the important features of the estimator (3) is that it does not require undersmoothing, so that bandwidths of the usual orderh
n
1=5 lead to the resultn
1=2(n ) Normal(0B
;1CB
;1) (4)where
B
is the covariance matrix ofX
;E
(X
jT
) andC
is the covariance matrix offX
;E
(X
jT
)g. The least squares form of (3) can be used to show that if one ignores measurement error and replacesX
byW
, the resulting estimate is inconsistent for . The form though suggests even more. It is well{known that in linear regression, inconsistency caused by measurement error can be overcome by applying the so{called \correction for attenuation". In our context, this suggests that we use the estimator bn="
n
X
i=1
f
W
i;g
bwh(T
i)gfW
i;bg
wh(T
i)gT ;n
uu#
;1Xn i=1
f
W
i;bg
wh(T
i)gfY
i;g
byh(T
i)g:
(5) The estimator (5) can be derived in much the same way as the Severini{Staniswalis estimator. For every , letg
b(T
) maximize the weighted likelihood ignoring measurement error, and then form via a negatively penalized operation:minimize Xn
i=1
n
Y
i;W
Ti;bg
(T
i)o2;Tuu:
(6) The negative sign in the second term in (6) looks odd until one remembers that the eect of measurement error is attenuation, i.e., to underestimate in absolute value when it is scalar, and thus one must correct for attenuation by making larger, not by shrinking it further towards zero.In this paper, we analyze the estimate (5), showing that it is consistent, asymptotically normally distributed with a variance dierent from (4). Just as in the Severini{Staniswalis algorithm, in kernel weighting ordinary bandwidths of order
h
n
;1=5 may be used.The outline of the paper is as follows. In Section 2, we dene the weighting scheme to be used and hence the estimators of
andg
( ). Section 3 is the statement of the main results for, while the results forg
( ) are stated in Section 4. Section 5 states the corresponding results for the measurement error variance uu estimated. Section 6 gives a numerical illustration. Final remarks are given in Section 7. All proofs are delayed until the appendix.2 DEFINITION OF THE ESTIMATORS
For technical convenience we will assume that the
T
i are conned to the interval 01]. Throughout, we shall employC
(0< C <
1) to denote some constant not depending onn
but may assume dierent values at each appearance. In our proofs and statement of results, we will let theX
's be unknown xed constants, a situation which is commonly called the functional relation, see Kendall& Stuart (1992) and Anderson (1984). The results apply immediately to the case that the
X
's are independent random variables, see Section 7.Let
!
ni(t
) =!
ni(t T
1:::T
n) be probability weight functions depending only on the design pointsT
1:::T
n:
For example!
ni(t
) = 1h
nZ si
si;1
K
t
;s h
nds
1i
n
(7)where
s
0 = 0s
n= 1 ands
i = (1=
2)(T
i+T
i+1)1i
n
;1:
Hereh
n is a sequence of bandwidth parameters which tends to zero asn
! 1 andK
( ) is a kernel function, which is supported to have compact support and to satisfysupp(
K
) = ;11]supjK
(x
)jC <
1ZK
(u
)du
= 1 andK
(u
) =K
(;u
):
In the paper, for any a sequence of variables or functions (
S
1:::S
n), we always denoteS
T = (S
1:::S
n)S
ei =S
i ;Pnj=1!
nj(T
i)S
j SeT = (S
e1::: S
en):
For example, W
fT = (W
f1::: W
fn) W
fi =W
i;Pnj=1!
nj(T
i)W
jeg
i=g
(T
i);Pnk=1!
nk(T
i)g
(T
k) Ge = (g
e1::: g
en)T:
g
e1::: g
en)T:
The fact that
g
(t
) =E
(Y
i;X
TijT
=t
) =E
(Y
i;W
TijT
=t
) suggestsb
g
n(t
) = Xnj=1
!
nj(t
)(Y
j;W
Tjbn) (8)as the estimator of
g
(t
).In some cases, it may be reasonable to assume that the model errors
i are homoscedastic with common variance 2. In this event, sinceE
fY
i;X
Ti;g
(T
i)g2= 2andE
fY
i;W
Ti;g
(T
i)g2=E
fY
i;X
Ti;g
(T
i)g2+Tuu, we deneb
2n=
n
;1Xni=1(
Y
ei;W
fTibn)2;bTnuubn:
(9) as the estimator of 2.3 MAIN RESULTS
We make the following assumptions.
Assumption 1.1.
There exist functionsh
j( ) dened on 01] such that thej
th component ofX
i, namelyX
ij, satisesX
ij =h
j(T
i) +V
ij, 1i
n
, 1j
p
, whereV
ij is a sequence of real numbers which satisfy limn!1n
;1Pni=1V
i = 0 and limn!1n
;1Pni=1V
iV
Ti =B
is a positive denite matrix, whereV
i = (V
i1:::V
ip)T.Assumption 1.2. g( ) and h
j( ) are Lipschitz continuous of order 1.
Assumption 1.3.
Weight functions!
ni( ) satisfy:(
i
) max1in n
X
j=1
!
ni(T
j) =O
(1)(
ii
) max1ijn
!
ni(T
j) =O
(b
n)(
iii
) max1in n
X
j=1
!
nj(T
i)I
(jT
j;T
ij> c
n) =O
(c
n) whereb
n=n
;4=5c
n=n
;1=5logn:
Our two main results concern the limit distributions of the estimate of
and 2.THEOREM 3.1.
Suppose Assumptions 1.1-1.3 hold andE
(4 +kU
k4)<
1. Then bn is an asymptotically normal estimator, i.e.n
1=2(bn;))N
(0B
;1;B
;1) where; =
E
h(;U
T)2fX
;E
(X
jT
)gfX
;E
(X
jT
)gTi+E
f(UU
T;uu)T(UU
T ;uu)g+E
(UU
T2):
Note that ;=E
(;U
T)2B
+E
f(UU
T;uu)T(UU
T;uu)g+ uu 2 if is homoscedastic and independent of (XT
).THEOREM 3.2.
Suppose the condition of Theorem 3.1 hold, and that the's are homoscedastic with variance 2, and independent of (XT
). Thenn
1=2(bn2; 2))N
(0 2) where 2 =E
f(;U
T)2;(Tuu+ 2)g2.Remarks
As described in the introduction, an important aspect of the results of Severini and Staniswalis is that their methods lead to asymptotically normal parameter estimates in kernel regression even with bandwidths of the usual order
h
nn
;1=5. The same holds for our estimators in general. For example, suppose that the design pointsT
i satisfy that there exist constantsM
1M
2 such thatM
1=n
mininj
T
i;T
i;1jmaxin j
T
i;T
i;1jM
2=n:
Then Assumptions 1.3(i)-(iii) are satised by simple verication.
It is relatively easy to estimate the covariance matrix of
bn. Let dim(X
) be the number of components ofX
. A consistent estimate ofB
is justf
n
;dim(X
)g;1Xni=1
f
W
i;g
bwh(T
i)gfW
i;g
bwh(T
i)gT;uu:
In the general case, one can use (30) to construct a consistent sandwich{type estimate of ;, namely
n
;1Xni=1
n
W
fi(Y
ei;W
fTibn) + uubnon
W
fi(Y
ei;W
fTibn) + uubnoT
:
In the homoscedastic case, namely that
is independent of (XTU
) with variance 2, and withU
being normally distributed, a dierent formula can be used. Let C() =E
f(UU
T; uu)T(UU
T;uu)g. Then a consistent estimate of ;is(bn2+
bTnuubn)B
bn+bn2uu+C(bn):
In the classical functional model, instead of obtaining an estimate of uu through replication, it is instead assumed that the ratio of uu to 2 is known. Without loss of generality, we set this ratio equal to the identity matrix. The resulting analogue of the parametric estimators to the partially linear model is to solve the following minimization problem:
n
X
i=1
j
Y
ei;W
fTip1 +k
k2j2 = min!where here and in the sequelk k denotes the Euclidean norm. One can use the techniques of this paper to show that this estimator is consistent and asymptotically normally distributed.
The asymptotic variance of the estimate of
in this case whenis independent of (XT
) can be shown to equalB
;1"
(1 +k
k2)2 2B
+E
f(;U
T)2;1;T1g 1 +kk2#
B
;1 where ;1 = (1 +kk2)U
+ (;U
T):
4 ASYMPTOTIC RESULTS FOR THENONPARAMETRICPART
THEOREM 4.1.
Suppose Assumptions 1.1-1.3 hold and!
ni(t
) are Lipschitz continuous of order 1 for alli
= 1:::n:
IfE
(4+kU
k4)<
1. Then for xedT
i, the asymptotic bias and asymptotic variance ofg
bn(t
) are respectively,Pni=1!
ni(t
)g
(T
i);g
(t
) andPni=1!
ni2 (t
)(Tuu+ 2). These are all of orderO
(n
;2=5) for kernel estimators.If the (
X
iT
i) are random, then the bias and variance formulae are the usual ones for nonpara- metric kernel regression.Although in some cases the measurement error covariance matrix uu has been established by independent experiments, in others it is unknown and must be estimated. The usual method of doing so is by partial replication, so that we observe
W
ij =X
i+U
ijj
= 1:::m
i.We consider here only the usual case that
m
i 2, and assume that a fraction of the data has such replicates. LetW
i be the sample mean of the replicates. Then a consistent, unbiased method of moments estimate for uu isbuu =
Pni=1Pmj=1i (
W
ij;W
i)(W
ij;W
i)TPni=1(
m
i;1):
The estimator changes only slightly to accommodate the replicates, becoming
bn ="
n
X
i=1
n
W
i;bg
wh(T
i)onW
i;bg
wh(T
i)oT;n
(1;=
2)buu#
;1
n
X
i=1
n
W
i;bg
wh(T
i)ofY
i;g
byh(T
i)g (10) whereg
bwh( ) is the kernel regression of theW
i's onT
i.Using the techniques in the appendix, one can show that the limit distribution of (10) is Normal(0
B
;1;2B
;1), with;2 = (1;
)E
h(;U
T)2fX
;E
(X
jT
)gfX
;E
(X
jT
)gTi +E
h(;U
T)2fX
;E
(X
jT
)gfX
;E
(X
jT
)gTi+(1;
)E
hfUU
T;(1;=
2)uugTfUU
T;(1;=
2)uug+UU
T2i+
E
hfUU
T ;(1;=
2)uugTfUU
T;(1;=
2)uug+UU
T2i:
(11) In (11),U
refers to the mean of twoU
's. In the case that is independent of (XT
), the sum of the rst two terms simplies to f 2+T(1;=
2)uugB
.Standard error estimates can also be derived. A consistent estimate of
B
isB
bn=fn
;dim(X
)g;1Xni=1
n
W
i;g
bwh(T
i)onW
i;g
bwh(T
i)oT;(1;=
2)buu:
Estimates of ;2 are also easily developed. In the homoscedastic case with normal errors, the sum rst two terms is estimated by (bn2 + (1;
=
2)bTnbuubn)B
bn. The sum of the last two terms is a deterministic function of ( 2uu), and these estimates are simply substituted into the formula.A general sandwich{type estimator is developed as follows. Dene
=n
;1Pni=1m
;1i , and deneR
i=W
fi(Y
ei;W
fTi bn) +buubn=m
i+ (=
)(m
i;1)n(1=
2)(W
i1;W
i2)(W
i1;W
i2)T ;buuo
:
Kernel fit: SBP on patient Age
40 50 60
Patient Age
125130135140Average SBP
Figure 1: Estimate of the function
g
(T
) in the Framingham data ignoring measurement error.Then a consistent estimate of ;2 is the sample covariance matrix of the
R
i's.6 NUMERICAL EXAMPLE
To illustrate the method, we consider data from the Framingham Heart Study. We considered
n
= 1615 males withY
being their average blood pressure in a x 2{year period,T
being their age andW
being the logarithm of the observed cholesterol level, for which there are two replicates.We did two analyses. In the rst, we used both cholesterol measurements, so that in the notation of Section 5,
= 1. In this analysis, there is not a great deal of measurement error. Thus, in our second analysis, which is given for illustrative purposes, we used only the rst cholesterol measurement, but xed the measurement error variance at the value obtained in the rst analysis, in which case = 0. For nonparametric tting, we chose the bandwidth using crossvalidation to predict the response. In precise, we compute the square error using a geometric sequence of 191 bandwidths ranging in 120]. The optimal bandwidth is selected to minimize the square error among these 191 candidates. An analysis ignoring measurement error found some curvature inT
, see Figure 1 for the estimate ofg
(T
).As below mention, we will consider four case, using XploRe4 (See Hardle, et al. (1995)) to calculate each case. Our results are as follows. First consider the case that the measurement error was estimated, and both cholesterol values were used to estimate uu. The estimator of
ignoring measurement error was 9:
438, with estimated standard error 0:
187. When we accounted for measurement error, the estimate increased slightly tob= 12:
540, and the standard error increased to 0:
195.In the second analysis, we xed the measurement error variance and used only the rst choles- terol value. The estimator of
ignoring measurement error was 10:
744, with estimated standard er- ror 0:
492. When we accounted for measurement error, the estimate increased slightly tob= 13:
690, and the standard error increased to 0:
495.7 DISCUSSION
Our results have been phrased as if the
X
's were xed constants. If they are random variables, the proofs simplify and the same results are obtained, with nowV
i =X
i;E
(X
ijT
i).The nonparametric regression estimator (8) is based on locally weighted averages. In the random
X
context, the same results apply if (8) is replaced by a locally linear kernel regression estimator.If we ignore measurement error, the estimator of
is given by (3) but with the unobservedX
replaced by the observedW
. This diers from the correction for attenuation estimator (5) by a simple factor which is the inverse of the reliability matrix (Gleser, 1992). In other words, the estimator which ignores measurement error is multiplied by the inverse of the reliability matrix to produce a consistent estimate of . This same algorithm is widely employed in parametric measurement error problems for generalized linear models, where it is often known as an example of regression calibration (see Carroll, et al., 1995, for discussion and references). The use of regression calibration in our semiparametric context thus appears to hold promise when (1) is replaced by a semiparametric generalized linear model.We have treated the case that the parametric part
X
of the model has measurement error and the nonparametric partT
is measured exactly. An interesting problem is to interchange the roles ofX
andT
, so that the parametric part is measured exactly and the nonparametric part is measured with error, i.e.,E
(Y
jXT
) =T
+g
(X
). Fan and Truong (1993) have shown in this case that with normally distributed measurement error, the nonparametric functiong
( ) can be estimated only at logarithmic rates, and not with raten
;2=5. We conjecture even so thatis estimable at parametric rates, but this remains an open problem.Anderson, T.W. (1984). Estimating Linear Statistical Relationships. Annals of Statistics, 12, 1-45.
Carroll, R.J., Ruppert, D. and Stefanski, L.A. (1995), Nonlinear Measurement Error Models. Chap- man and Hall, New York.
Chen, H. (1988). Convergence Rates for Parametric Components in a Partly Linear Model. Annals of Statistics, 16, 136-146.
Cuzick, J. (1992a). Semiparametric Additive Regression. Journal of the Royal Statistical Society, Series B, 54, 831-843.
Cuzick, J. (1992b). E"cient Estimates in Semiparametric Additive Regression Models with Un- known Error Distribution. Annals of Statistics, 20, 1129-1136.
Engle, R. F., Granger, C.W.J., Rice,J. and Weiss, A. (1986). Semiparametric estimates of the relation between weather and electricity sales. Journal of the American Statistical Association,
81
, 310-320.Fan, J. & Truong, Y. K. (1993). Nonparametric regression with errors in variables. Annals of Statistics, 21, 1900{1925.
Fuller, W. A. (1987). Measurement Error Models. Wiley, New York.
Gleser, L.J. (1992). The importance of assessing measurement reliability in multivariate regression.
Journal of the American Statistical Association, 87, 696-707.
Hardle,W., Klinke, S. and Turlach, B.A. (1995). XploRe: An Interactive Statistical Computing Enviroment. Springer-Verlag,
Heckman, N. E. (1986). Spline smoothing in partly linear models. Journal of the Royal Statistical Society, Series B, 48, 244-248.
Kendall, M. and Stuart, A. (1992). The Advanced Theory of Statistics 2, 4th ed, Charles Gri"n, London.
Liang, H. and Hardle, w. (1997). Asymptotic normality of parametric part in partially linear heteroscedastic regression models. manuscript.
Speckman, P. (1988). Kernel Smoothing in Partial Linear Models. Journal of the Royal Statistical Society, Series B, 50, 413-436.
Severini, T.A. and Staniswalis, J.G. (1994). Quasilikelihood Estimation in Semiparametric Models.
Journal of the American Statistical Association, 89, 501-511.
8 APPENDIX
In this appendix, we prove several lemmas required. Lemma A.1 provides bounds for
h
j(T
i);Pnk=1
!
nk(T
i)h
j(T
k) andg
(T
i);Pnk=1!
nk(T
i)g
(T
k). The proof is immediate.Lemma A.1.
Suppose that Assumptions 1.1 and 1.3 (iii) hold. Then max1inj
G
j(T
i);Xnk=1
!
nk(T
i)G
j(T
k)j=O
(c
n) forj
= 0:::p
whereG
0( ) =g
( ) andG
l( ) =h
l( ) forl
= 1:::p:
Lemma A.2.
If Assumptions 1.1-1.3 hold. Thennlim!1
n
;1X
eTX
e =B
Proof.
Denoteh
ns(T
i) =h
s(T
i);Pnk=1!
nk(T
i)X
ks. It follows fromX
js=h
s(T
j) +V
js that the (sm
) element ofX
eTX
e (sm
= 1:::p
) isn
X
j=1
X
ejsX
ejm = Xnj=1
V
jsV
jm+Xnj=1
h
ns(T
j)V
jm+Xnj=1
h
nm(T
j)V
js+Xnj=1
h
ns(T
j)h
nm(T
j)def= Xn
j=1
V
jsV
jm+X3q=1
R
(nsmq)The strong law of large number implies that limn!11
=n
Pni=1V
iV
Ti =B
and Lemma A.1 meansR
(3)nsm =o
(n
), which together with the Cauchy-Schwarz inequality show thatR
(1)nsm =o
(n
) andR
(2)nsm=o
(n
):
This completes the proof of the lemma.Lemma A.3.
(Bennett's inequality) Let ;1:::
;n be independent random variables with zero means and bounded ranges: j;ijM:
Then for each>
0P
fjXni=1;ij
>
g2exp(
;
2."
2
(Xn
i=1
var
(;i) +M
)#)
:
Denote
0j = jI
(jjjn
1=4) and 00j = j ;0jj
= 1:::n:
We next establish several results for nonparametric regression.Lemma A.4.
Assume that Assumption 1.3 holds andE
= 0 andE
4<
1Then max1inj n
X
k=1
!
nk(T
i)kj=O
P(n
;1=4log;1=2n
): Proof. LetM
=Cb
nn
1=4. Lemma A.3 implies
P
0
@max
1inj n
X
j=1
!
nj(T
i)(0j;E
0j)j> C
1n
;1=4log;1=2n
1
A
n
X
i=1
P
8
<
: j
n
X
j=1
!
nj(T
i)(0j;E
0j)j> C
1n
;1=4log;1=2n
9
=
2
n
exp(
;
C
1n
;1=2log;1n
Pnj=1
!
nj2 (T
i)E
2j+ 2b
nlog;1=2n
)
2
n
expf;C
12C
logn
g;!0 for some largeC
1>
0which implies that
max
1inj n
X
j=1
!
nj(T
i)(0j ;E
0j)j=O
P(n
;1=4log;1=2n
):
(12) On the other hand, we know thatmax
1in n
X
j=1
!
nj(T
i)jE
00jj max1kn max
1inj
!
nk(T
i)jXnj=1
n
;1E
jjj4
Cn
;2=3E
jj4=O
(n
;1=2) =O
P(n
;1=4log;1=2n
) (13)n
X
j=1
E
j00jjn
;1Xnj=1
E
4j =E
4:
(14)Moreover, the Hartman-Winter theorem entails that
n
X
j=1(j
00jj;E
j00jj) =O
2
6
4 8
<
:
n
X
j=1
E
j00jj2loglog0
@
n
X
j=1
E
j00jj21
A 9
=
1=23
7
5=
O
P(n
1=4log log1=2n
):
(15) It follows from (14) and (15) thatn
X
j=1j
00jj=O
P(n
1=4loglog1=2n
) (16) andmax
1inj n
X
j=1
!
nj(T
i)00jj max1kinj
!
nk(T
i)jXnj=1
j
00jj=O
(n
;5=12loglog1=2n
) =O
P(n
;1=4log;1=2n
) Combining the results of (12), (13) with (15), we obtainmax
1inj n
X
k=1
!
nk(T
i)kj=O
P(n
;1=4log;1=2n
):
This completes the proof of Lemma A.4.Lemma A.5.
Assume that Assumptions 1.1-1.3hold. IfE
= 0 andE
4<
1:
ThenI
n=o
P(n
1=2), whereI
n =Pni=1Pj6=i!
nj(T
i)(0j;E
0j)(0i;E
0i):
Proof.
Letj
n = hn
2=3log2n
i (a
] denotes the integer portion ofa:
)A
j = nh(j;1)jn ni+ 1
:::
hjn jn
io
A
cj =f12:::n
g;A
j andA
ji =A
j;fi
g:
Observe thatI
n = Xjnj=1
X
i2Aj
X
k2Aj i
!
nk(T
i)(0k;E
0k)(0i;E
0i) +Xjnj=1
X
i2Aj
X
k2Acj
!
nk(T
i)(0k;E
0k)(0i;E
0i)def= Xjn
j
U
nj+Xjnj