• No results found

A simulation method for skewness correction

N/A
N/A
Protected

Academic year: 2022

Share "A simulation method for skewness correction"

Copied!
40
0
0

Loading.... (view fulltext now)

Full text

(1)

U.U.D.M. Project Report 2008:22

Examensarbete i matematisk statistik, 30 hp Handledare och examinator: Silvelyn Zwanzig December 2008

Department of Mathematics

A simulation method for skewness correction

Måns Eriksson

(2)
(3)

Abstract

Let X1, . . . , Xn be i.i.d. random variables with known variance and skewness.

A one-sided confidence interval for the mean with approximate confidence level α can be constructed using normal approximation. For skew distributions the actual confidence level will then be α + o(1). We propose a method for obtaining confidence intervals with confidence level α+o(n−1/2) using skewness correcting pseudo-random variables. The method is compared with a known method; Edgeworth correction.

(4)

h h

(5)

Acknowledgements

I would like to thank my advisor Silvelyn Zwanzig for introducing me to the subject, for mathematical and stylistic guidance and for always encouraging me.

I would also like to thank my friends and teachers at and around the department of mathematics for having inspired me to study mathematics, and for continuing to inspire me.

(6)

h h

(7)

Contents

1 Introduction 7

1.1 Skewness . . . . 7

1.2 Setting and notation . . . . 7

2 The Edgeworth expansion 8 2.1 Definition and formal conditions . . . . 8

2.2 Derivations . . . . 11

2.2.1 Edgeworth expansion for Sn . . . . 11

2.2.2 Edgeworth expansions for more general statistics . . . . 14

2.2.3 Edgeworth expansion for Tn . . . . 15

2.2.4 Some remarks; skewness correction . . . . 16

2.3 Cornish-Fisher expansions for quantiles . . . . 16

3 Methods for skewness correction 17 3.1 Coverages of confidence intervals . . . . 17

3.2 Edgeworth correction . . . . 18

3.3 The bootstrap . . . . 19

3.3.1 The bootstrap and Sn . . . . 19

3.3.2 Bootstrap confidence intervals . . . . 20

3.4 A new simulation method . . . . 21

3.4.1 Skewness correction through addition of a random variable . . . . 21

3.4.2 Simulation procedure . . . . 22

4 Comparison 23 4.1 Coverages of confidence intervals . . . . 23

4.2 Criteria for the comparison . . . . 24

4.3 Comparisons of the upper limits . . . . 24

4.4 Simulation results . . . . 26

4.5 Discussion . . . . 26

A Appendix: Skewness and kurtosis 28 A.1 The skewness of a sum of random variables . . . . 28

A.2 The kurtosis of a sum of random variables . . . . 30

B Appendix: P (ˆθnew ≤ ˆθEcorr) 31

C Appendix: Simulation results 36

(8)

h h

(9)

1 Introduction

1.1 Skewness

The notion of skewness has been a part of statistics for a long time. It dates back to the 19th century and most notably to an article by Karl Pearson from 1895 ([26]). Skew distributions are found in all areas of applications, ranging from finance to biology and physics. It has been seen both in theory and in practice that deviations from normality in the form of skewness might have effects too big to ignore on the validity and performance of many statistical methods and procedures.

This thesis discusses skewness in the context of the central limit theorem and normal approximation, in particular applied to confidence intervals. Some methods for skewness correction are discussed and a new simulation method is proposed. We assume that the concept of skewness is known and refer to Appendix A for some basic facts about skewness.

1.2 Setting and notation

Throughout the thesis we assume that we have an i.i.d. univariate sample X1, . . . , Xn, with EX = µ, Var(X) = σ2 and E|X|3 < ∞, such that X satisfies Cram´er’s condition lim supt→∞|ϕ(t)| < 1, where ϕ is the characteristic function of X. At times we will also assume that EX4 < ∞. We use X to denote a generic Xi, that is, X is a random variable with the same distribution as the Xi. Thus, for instance, EX is the mean of the distribution of the observations; EX = EXi for all i.

The α-quantile vα of the distribution of some random variable X is defined to be such that P (X ≤ vα) = α. When X ∼ N (0, 1) we denote the quantile λα, that is, Φ(λα) = α.

We use An to denote a general statistic, Sn = n1/2( ¯X − µ)/σ to denote the stan- dardized sample mean and Tn= n1/2( ¯X − µ)/ˆσ to denote the studentized sample mean, where ˆσ2 = n1P(Xi− ¯X)2.

The skewness E(X − µ)33 of a random variable X is denoted Skew(X), γ or γX

if we need to distinguish between different random variables. The kurtosis of X, E(X − µ)44− 3, is denoted Kurt(X), κ or κX. Basic facts about skewness and kurtosis are stated in Appendix A.

As for asymptotic notation, for real-valued sequences an and bn we say that an= o(bn) if an/bn→ 0 as n → ∞ and an= O(bn) if an/bn is bounded as n → ∞.

Finally, we say that a sequence Xn of random variables is bounded in probability if, limc→∞lim supn→∞P (|Xn| > c) = 0. We write this as Xn = OP(1) and if, for some sequence an, anXn= OP(1) we write Xn= OP(1/an).

(10)

2 The Edgeworth expansion

In this section we introduce our main tool, the Edgeworth expansion. Later we will use it to determine the coverage of confidence intervals.

2.1 Definition and formal conditions

Theorem 1. Assume that X1, . . . , Xn is an i.i.d. sample from a univariate distribution, with mean µ, variance σ2 and E|X|j+2 < ∞, that satisfies lim supt→∞|ϕ(t)| < 1. Let Sn= n1/2( ¯X − µ)/σ. Then

P (Sn≤ x) = Φ(x) + n−1/2p1(x)φ(x) + . . . + n−j/2pj(x)φ(x) + o(n−j/2) (1) uniformly in x, where Φ(x) and φ(x) are the standard normal distribution function and density function and pk is a polynomial of degree 3k − 1. In particular

p1(x) = −1

6γ(x2− 1) and p2(x) = −x 1

24κ(x2− 3) + 1

72γ2(x4− 10x2+ 15) . Proof. A proof is given in Section 2.2.1. See also [7] and [8].

(1) is called an Edgeworth expansion for Sn.

The condition lim supt→∞|ϕ(t)| < 1 is known as Cram´er’s condition and was de- rived by Cram´er in [8]. Note that the condition holds whenever X is absolutely con- tinuous. This is an immediate consequence of the Riemann-Lebesgue lemma (Theo- rem 1.5 in Chapter 4 of [18]). Moreover, if we limit the expansion to P (Sn ≤ x) = Φ(x) + n−1/2p1(x)φ(x) + o(n−1/2), so that the remainder term is o(n−1/2), then it suf- fices that X has a non-lattice distribution, which was shown by Esseen in [16].

The Edgeworth expansion was first developed for the statistic Snbut has later been extended to other statistics.

Definition 1. Let Andenote a statistic. Then if

P (An≤ x) = Φ(x) + n−1/2a1(x)φ(x) + . . . + n−j/2aj(x)φ(x) + o(n−j/2), (2) where Φ(x) and φ(x) are the standard normal distribution function and density function and ak is a polynomial of degree 3k − 1, (2) is called the Edgeworth expansion for An.

If ak= 0 for all k < i the normal approximation of the distribution is said to be i:th order correct.

In general, the polynomials ak will depend on the moments of the statistic. We can therefore, in a sense, view the Edgeworth expansion as an extension of the central limit theorem, where information about the higher moments of the involved random variables is used to obtain a better approximation of the distribution function of An. The expan- sion gives an expression for the size of the remainder term depending on the sample size.

(11)

We might want to compare this to the Berry-Esseen theorem (see for instance [16] or Section 7.6 of [18]), which essentially, in our context, says that the error in the normal approximation is of order n−1/2.

The Edgeworth expansion for simple statistics was first introduced in papers by Cheby- shev in 1890 ([4]) and Edgeworth in 1894, 1905 and 1907 ([12, 13, 14]). The idea was made mathematically rigorous by Cram´er in 1928 ([7]) and Esseen in 1945 ([16]). The expansions and their applications for more general statistics were then developed in several papers, including [2], by various authors in the mid-1900’s.

A thorough treatment of the Edgeworth expansion is found in Chapter 2 of [22].

Chapter 13 of [10] gives a brief introduction to the Edgeworth expansion with some of the most important results and Section 17.7 of [8] is a standard reference for the case where the expansion for Snis considered.

Conditions for (2) to hold are given next. Although we will focus on expansions of quite simple statistics, the Edgeworth expansion can be used in very general circum- stances. We state a theorem by Bhattacharya and Ghosh ([2]) that provides conditions for the Edgeworth expansion (2) to hold in a general case and illustrate how the theorem relates to the expansion for Sn.

Let X, X1, X2, . . . , Xn be i.i.d. random column vectors in Rd with mean µ and let X = n¯ −1Pn

i=1Xi. Let A : Rd→ R be a function of the form AS= (g(x) − g(µ))/h(µ) or AT = (g(x) − g(µ))/h(x), where g and h are known, ˆθ = g( ¯X) is an estimator of the scalar θ = g(µ), h(µ)2 is the asymptotical variance of n1/2θ and h( ¯ˆ X) is an estimator of h(µ).

For t ∈ Rd, t = (t(1), t(2), . . . , t(d)), define ||t|| = ((t(1))2+ . . . + (t(d))2)1/2 and, for a random d-vector X, ϕ(t) = E(exp(iPd

j=1t(j)X(j)).

Theorem 2. Assume that A has j + 2 continuous derivates in a neighbourhood of µ = E(X) and that A(µ) = 0. Furthermore, assume that E(||X||j+2) < ∞ and that the characteristic function ϕ of X is such that lim sup||t||→∞|ϕ(t)| < 1. Let σA be the asymptotic standard deviation of n1/2A( ¯X) and assume that σA> 0. Then, with An= n1/2A( ¯X)/σA and for j ≥ 1,

P (An≤ x) = Φ(x) + n−1/2a1(x)φ(x) + . . . + n−j/2aj(x)φ(x) + o(n−j/2) (3) uniformly in x. ak is a polynomial of degree 3k − 1 with coefficients depending on A and on moments of X of order less than or equal k + 2. ak is odd for even k and even for odd k.

Proof. See [2].

Note that the above theorem is a summary of the results in Bhattacharya’s and Ghosh’s 1978 paper and thus not stated in one single theorem in the original work. Our summary largely resembles that in Chapter 2 of [22].

(12)

The class of functions that satisfy the conditions in Theorem 2 contains many func- tions of great interest. In particular, we can write any moment estimator, i.e. an estimator based on sample moments, as a function A of vector means in the same way that we will do below for the mean.

The Edgeworth expansion is sometimes written as an infinte series, P (An ≤ x) = Φ(x) + n−1/2a1(x)φ(x) + . . . + n−j/2aj(x)φ(x) + . . .. However, for the series to converge it is required, for an absolutely continuous random variable, that E

exp((X−µ)4·σ22)

< ∞;

a condition that fails even for exponentially distributed X (see [7]). Thus we prefer the stopped series used in (3), which also turns out to be more useful in practice.

Before we show how Theorem 2 relates to the expansion for Sn we state a slightly more general corollary.

Corollary 1. Let An be either AS = n1/2θ − θ)/σ or AT = n1/2θ − θ)/ˆσ, where θ is some unknown scalar parameter, ˆθ is an asymptotically unbiased estimator of θ, σ2 is the asymptotic variance of n1/2θ and ˆˆ σ2 is some consistent estimator of σ2. Then the first two polynomials in the Edgeworth expansion are

a1(x) = −

k1,2+1

6k3,1(x2− 1) and

a2(x) = −x

1

2(k2,2+ k1,22 ) + 1

24(k4,1+ 4k1,2k3,1)(x2− 3) + 1

72k3,12 (x4− 10x2+ 15)

 where the kj,i comes from an expansion of the j:th cumulant of An:

κj,n= n−(j−2)/2(kj,1+ n−1kj,2+ n−2kj,3+ . . .).

Proof. Details are given in Section 2.2.2.

Theorem 2 can be used to show that Sn as well as the studentized sample mean Tn= n−1/2( ¯X −µ)/ˆσ admit Edgeworth expansions. Assume that X1, . . . , Xnis a sample from a univariate distribution, that the unknown parameter θ is the mean µ of the distribution and that the distribution has variance σ2. Take d = 2, Xi = (Xi, Xi2)T and µ = E(X) = (EX, EX2)T and let g(x(1), x(2)) = x(1) and h(x(1), x(2)) = x(2)− (x(1))2.

Then g(µ) = µ and g( ¯X) = ¯X. Furthermore h(µ) = σ2 and h( ¯X) = n−1

n

X

i=1

Xi2 n−1

n

X

i=1

Xi2

= n−1

n

X

i=1

(Xi− ¯X)2 = ˆσ2

AS = (g(x) − g(µ))/h(µ) and AT = (g(x) − g(µ))/h(x) both fulfill the conditions in Theorem 2 and the asymptotic standard deviation σAof n1/2A( ¯X) is 1. Furthermore, the condition E(pX2+ (X2)2j+2) < ∞ can be reduced to E(|X|j+2) < ∞ and Cram´er’s condition reduces to lim supt→∞|ϕ(t)| < 1. Thus the conditions for existence of the expansion for Sn and Tn follow and turn out to be those in Theorem 1. Both statistics are of the form that is considered in Corollary 1 and the polynomials in their expansions can thus be found.

(13)

2.2 Derivations

Knowing under which conditions the Edgeworth expansion exists, we are ready to derive the expressions for the polynomials p1 and p2.

2.2.1 Edgeworth expansion for Sn

As before, let Sn = n1/2( ¯X − µ)/σ. Assuming that E|X|j+2 < ∞ and that X fulfills Cram´er’s condition, by Corollary 1

P (Sn≤ x) = Φ(x) + n−1/2p1(x)φ(x) + . . . + n−j/2pj(x)φ(x) + o(n−j/2). (4) To actually be able to make reasonable use of the expansion we need to find expressions for the polynomials pk. The case where j = 2 will prove to be of special interest to us, so although we consider general j we will in the end only derive explicit expressions for p1 and p2. Our exposition is mainly based on that in [8] and [22] but aims to be somewhat more thorough than those that we have seen in the existing literature.

Since we will assume the existence of moments of order higher than 2, Sn is asymp- totically N (0, 1)-distributed, i.e. Sn converges in distribution to S ∼ N (0, 1), and thus the characteristic function ϕnof Snconverges to e−t/2, the characteristic function of the standard normal distribution, as n tends to infinity. That is, as n → ∞

ϕn(t) = E(exp(itSn)) −→ E(exp(itS)) = e−t2/2 for −∞ < t < ∞ where S ∼ N (0, 1).

Recall that if Y1, . . . , Ynare i.i.d. and Sn= Y1+ . . . + Ynthen ϕSn(t) = (ϕY1(t))n(see Theorem 1.8 in Chapter 4 of [18]). Now, let Yi = (Xi−µ)/σ. Then Sn= n1/2( ¯X −µ)/σ = n1/2 1nPn

i1Yi = n−1/2Pn

i1Yi and thus ϕn(t) = E(exp(itSn)) = E



exp(itn−1/2

n

X

i=1

Yi)



=



ϕY(tn−1/2)

n

.

Next we define the cumulant generating function for Y as ln ϕY(t). A MacLaurin ex- pansion of ln ϕY(t) shows that if E(|Y |k) < ∞ then, after a rearrangement of the terms, we can write ln ϕY(t) on the form

ln ϕY(t) =

k

X

j=1

(it)j

j! κj + o(|t|k) as t → 0

for some {κj}. We call the coefficients κj the cumulants of Y ; in particular κj is the j:th cumulant. See Section 15.10 of [8] for details.

By Theorem 4.2 in Chapter 4 of [18] we have that ϕY(t) = 1 +Pk j=1

(it)j

j! EYj+ o(|t|k) as t → 0 when E(|Y |k) < ∞. If we for a moment don’t worry about the existence of

(14)

moments and convergence of the series we conclude that

X

j=1

(it)j

j! κj = ln ϕY(t) = ln 1 +

X

j=1

(it)j j! EYj

and by looking at the MacLaurin expansion of the right hand side (i.e. the MacLaurin expansion of the function ln(1 + x) where we replace x with P EYj(it)j/j!) it follows

that

X

j=1

(it)j j! κj =

X

k=1

(−1)k+11 k

X

j=1

(it)j j! EYj

k

. Comparing the coefficients of (it)j we find that

κ1 = EY = 0

κ2 = EY2− (EY )2= 1

κ3 = EY3− 3EY EY2+ 2(EY )3= EY3

κ4 = EY4− 3(EY )2− 4EY EY3+ 12(EY )2EY2− 6(EY )4= EY4

The expression for κj holds whenever E(|Y |j) < ∞. Note that the assumption that E(|X|j) < ∞ implies that E(|Y |j) < ∞.

Returning our attention to Sn, the relation ϕn(t) = (ϕY(tn−1/2))nand the fact that κ1 = 0 and κ2 = 1 now give us that

ϕn(t) =

expX

j=1

(itn−1/2)j j! κjn

= expX

j=1

n−(j−2)/2(it)j j! κj

=

exp



1

2t2+ n−1/21

3!κ3(it)3+ . . . + n−(j−2)/21

j!κj(it)j+ . . .



= e−t2/2exp



n−1/21

3!κ3(it)3+ . . . + n−(j−2)/21

j!κj(it)j+ . . .



= e−t2/2



1 + n−1/2r1(it) + n−1r2(it) + . . . + n−j/2rj(it) + . . .



where the last equality is obtained through the MacLaurin expansion ex= 1+x+x2/2!+

. . .

rj is a polynomial of degree 3j that depends on κ3, . . . , κj+2. By comparing the coefficients of n−j/2 on the last two lines we find that

r1(x) = 1 6κ3x3 and

r2(x) = 1

24κ4x4+ 1 72κ23x6. We can rewrite the expression for ϕn(t) above as

ϕn(t) = e−t2/2+ n−1/2r1(it)e−t2/2+ n−1r2(it)e−t2/2+ . . . + n−j/2rj(it)e−t2/2+ . . . (5)

(15)

Now, since ϕn(t) = R

x=−∞eitxdP (Sn ≤ X) and e−t2/2 = R

x=−∞eitxdΦ(x) it seems plausible that there is an inversion of (5) of the form

P (Sn≤ x) = Φ(x) + n−1/2R1(x) + . . . + n−j/2Rj(x) + . . . where Rk is a function such thatR

−∞eitxdRk(x) = rk(it)e−t2/2 so that ϕn(t) =

Z x=−∞

eitxdP (Sn≤ X) = Z

x=−∞

eitxdΦ(x) + n−1/2 Z

−∞

eitxdR1(x) + . . . = (5).

We would thus like to try to find such Rk. By repeating integration by parts j times we find that

e−t2/2= Z

x=−∞

eitxdΦ(x) = (−it)−1 Z

x=−∞

eitx(1)(x) = . . . = (−it)−j

Z x=−∞

eitx(j)(x)

where Φ(k)(x) = dxdkkΦ(x) = (dxd)kΦ(x) = DkΦ(x). Hence R

−∞eitxd((−D)kΦ(x)) = (it)ke−t2/2. Interpreting rk(−D) as a polynomial in D, making rk(−D) a differential operator, we thus have that R

−∞eitxd(rk(−D)Φ(x)) = rk(it)e−t2/2. Thus Rk(x) = rk(−D)Φ(x).

By differentiating Φ(x) we find that, for k ≥ 1, (−D)kΦ(x) = −Hk−1(x)φ(x) where φ(x) is the density function for the standard normal distribution and Hk are the Hermite polynomials:

H0(x) = 1 H1(x) = x, H2(x) = x2− 1 H3(x) = x(x2− 3), H4(x) = x4− 6x2+ 3,

H5(x) = x(x4− 10x2+ 15), . . .

We concluded above that r1(x) = 16κ3x3 and r2(x) = 241κ4x4+721κ23x6 and since Rk(x) = rk(−D)Φ(x) we thus have that

R1(x) = 1

6κ3(−D)3Φ(x) = −1

6κ3H2(x)φ(x) = −1

6(x2− 1)φ(x) and

R2(x) = − 1

24κ4H3(x)φ(x) − 1

72κ23H5(x)φ(x) =

− x1

24κ4(x2− 3) + 1

72κ23(x4− 10x2+ 15) φ(x).

(16)

Thus

P (Sn≤ x) = Φ(x) + n−1/2R1(x) + . . . + n−j/2Rj(x) + . . . =

Φ(x) + n−1/2p1(x)φ(x) + n−1p2(x)φ(x) + . . . + n−j/2pj(x)φ(x) + . . . (6) with

p1(x) = −1

6κ23(x2− 1) and p2(x) = −x1

24κ4(x2− 3) + 1

72κ23(x4− 10x2+ 15) . κ3 is the skewness of X and κ4 the kurtosis.

It can be shown (see Section 2.4 of [22]) that the inversion of (5) leading to (6) is valid when X is nonsingular and E(|X|j+2) < ∞ if we limit the series to j terms:

P (Sn≤ x) = Φ(x) + n−1/2R1(x) + . . . + n−j/2Rj(x) + o(n−j/2) =

Φ(x) + n−1/2p1(x)φ(x) + n−1p2(x)φ(x) + . . . + n−j/2pj(x)φ(x) + o(n−j/2). (7) This completes the proof of half of Corollary 1.

If X ∼ N (µ, σ2) then Sn ∼ N (0, 1) so we’d expect pj to be 0 for all j. This is indeed the case, since κj = 0 for j ≥ 3 for the standard normal distribution. Thus the

”expansion” still holds when P (Sn≤ x) = Φ(x).

It is of interest to note that both the skewness and the kurtosis are scale and trans- lation invariant, so that the third and fourth cumulants for X and Y coincide (we use Y in the calculations above because standardized variables are easier to handle). It can be shown, by looking at characteristic functions or by straightforward calculation (as is done in Appendix A), that the skewness of ¯X is n−1/2 times the skewness of X.

Similarly the kurtosis of ¯X is n−1 times the kurtosis of X and in general, for j ≥ 2 the j:th cumulant of ¯X will be the j:th cumulant of X times n−(j−2)/2. Thus we can view the factors n−(j−2)/2in the Edgeworth expansion for Sn as coming from the cumulants of ¯X.

2.2.2 Edgeworth expansions for more general statistics

The procedure for finding the Edgeworth expansion for more general statistics An is essentially the same as that in the previous section. We briefly mention the result. Let Anbe either AS= n1/2θ − θ)/σ or AT = n1/2θ − θ)/ˆσ, where θ is some unknown scalar parameter, ˆθ is an asymtotically unbiased estimator of θ, σ2 is the asymptotic variance of n1/2θ and ˆˆ σ2 is some consistent estimator of σ2.

Denote by κj,n the j:th cumulant of An. Under the regularity conditions stated in Theorem 2, for j ≥ 1, we can expand κj,n as

κj,n= n−(j−2)/2(kj,1+ n−1kj,2+ n−2kj,3+ . . .).

(17)

for some kj,i, where κ1,1 = 0 and κ2,1 = 1. It can be shown, through calculations that are completely analogous to the Sn case, where the cumulants of An are replaced by their expansions, that the first two polynomials in the Edgeworth expansion for Anare

a1(x) = −



k1,2+1

6k3,1H2(x)



= −



k1,2+1

6k3,1(x2− 1)

(8) and

a2(x) = −1

2(k2,2+ k1,22 )H1(x) + 1

24(k4,1+ 4k1,2k3,1)H3(x) + 1

72k3,12 H5(x)

=

− x1

2(k2,2+ k1,22 ) + 1

24(k4,1+ 4k1,2k3,1)(x2− 3) + 1

72k23,1(x4− 10x2+ 15)

 (9)

As before the inversion is valid, giving the truncated expansion

P (An≤ x) = Φ(x) + n−1/2a1(x)φ(x) + . . . + n−j/2aj(x)φ(x) + o(n−j/2) when E|X|j+2< ∞ and X satisfies Cram´er’s condition.

Thus the problem of finding the Edgeworth expansion for An on the form n1/2θ − θ)/σ or n1/2θ − θ)/ˆσ amounts to finding the terms kj,i in the expansion

κj,n = n−(j−2)/2(kj,1+ n−1kj,2+ n−2kj,3+ . . .) of the cumulants of An.

As we discussed in the previous section, if An= Sn is known then κj,n = n−(j−2)/2κj

for j ≥ 2, where κj is the j:th cumulant for Y = (X − µ)/σ. Thus kj,1= κj and kj,i = 0 for i ≥ 2. This reduces the expressions for a1 and a2 above to those for p1 and p2 in (6).

2.2.3 Edgeworth expansion for Tn

Finally, consider the statistic Tn= n1/2( ¯X − µ)/ˆσ where ˆσ2 = n1P(Xi− ¯X)2. It can be shown that the kj,i in the expansion of the cumulants of Tn are

k1,2= −1 2γ, k2,2= 1

4(7γ2+ 12), k3,1= −2γ and k4,1= 12γ2− 2κ + 6.

Inserting these into (8) and (9) we get q1(x) = a1(x) = −(−1

2γ −2

6γ(x2− 1)) = 1

6γ(2x2+ 1)

(18)

and

q2(x) = a2(x) = −1 2(1

4(7γ2+ 12) + (−1 2γ)2)+

1

24(12γ2− 2κ + 6 + 4(−1

2γ)(−2γ)(x2− 3) + 1

72(−2γ)2(x4− 10x2+ 15)

= x

 1

12κ(x2− 3) − 1

18γ2(x4+ 2x2− 3) − 1

4(x2+ 3)

 .

We require that EX4< ∞ and that X satisifes Cram´er’s condition for the expansion P (Tn≤ x) = Φ(x) + n−1/2q1(x)φ(x) + n−1q2(x)φ(x) + o(n−1)

to hold.

2.2.4 Some remarks; skewness correction

We’ve seen above that for both Sn and Tn the first polynomial a1 depends on the skewness κ3 = γ = E(X − µ)33 and that the second polynomial a2 depends on γ2 and the kurtosis κ4 = κ = E(X − µ)44− 3. This is also true for many more general statistics An. In such cases, a1 is said to describe the primary effect of skewness while a2 is said to describe the primary effect of kurtosis and the secondary effect of skewness.

A skewness corrected statistic is thus a statistic that has been modified in some way so that a1 = 0.

2.3 Cornish-Fisher expansions for quantiles

An interesting use of the Edgeworth expansion is asymptotic expansions of the quantiles of An, obtained by what is essentially an inversion of the Edgeworth expansion. Such expansions are called Cornish-Fisher expansions and first appeared in [6] and [17].

Let Anbe a statistic with the Edgeworth expansion

P (An≤ x) = Φ(x) + n−1/2a1(x)φ(x) + . . . + n−j/2aj(x)φ(x) + . . . (10) and let vα be the α-quantile of An, so that P (An ≤ vα) = α. Furthermore, let λα be the α-quantile of the N (0, 1)-distribution, i.e. let Φ(λα) = α. Then there exists an expansion of vα in terms of λα:

vα = λα+ n−1/2s1α) + n−1s2α) + . . . + n−j/2sjα) + . . . (11) (11) is called the Cornish-Fisher expansion of vα.

The functions sk are polynomials of degree at most k + 1, odd for even k and even for odd k, that depend on cumulants of order at most k + 2. They are determined by the polynomials ak in (10).

[22] contains a short introduction to the Cornish-Fisher expansion, where it is shown that s1(x) = −a1(x) and s2(x) = a1(x)a01(x) − 12a1(x)2− a2(x).

(19)

3 Methods for skewness correction

We discuss the need of skewness correction. Some methods for obtaining second order correct confidence intervals are described. We assume throughout the section that X is absolutely continuous.

3.1 Coverages of confidence intervals

Definition 2. Let Iθ(α) be a confidence interval for an unknown parameter θ with approximate confidence level α. We call α the nominal coverage of Iθ(α). Furthermore we call α0 = P (θ ∈ Iθ(α)) the actual coverage of the confidence interval and define the coverage error as the difference between the actual coverage and the nominal coverage, α0 − α.

If Iθ(α) has been reasonably constructed then this difference will converge to zero as the sample size n increases. We will illustrate how the Edgeworth expansion can be used to estimate the order of coverage errors.

Let X1, . . . , Xn be an i.i.d. sample from a distribution, with mean µ and variance σ2, such that the Edgeworth expansion for Sn and Tn exists. Consider the two sided confidence interval Jµ(α) = ( ¯X − n−1/2σzα, ¯X + n−1/2σzα) where P (|Z| ≤ zα) = α when Z ∼ N (0, 1). We find that

P (µ ∈ Jµ(α)) = P (Sn> −zα) − P (Sn> zα) = P (Sn≤ zα) − P (Sn≤ −zα) =

Φ(zα) − Φ(−zα) + n−1/2(p1(zα)φ(zα) − p1(−zα)φ(−zα)) + n−1(p2(zα)φ(zα) − p2(−zα)φ(−zα))

+ n−3/2(p3(zα)φ(zα) − p3(−zα)φ(−zα)) + o(n−3/2) = α + 2n−1p2(zα)φ(zα) + o(n−3/2).

The last equality follows since p2 is odd and p1, p3and φ are even. The same result holds if we have ˆσ instead of σ, with q2 instead of p2. Thus the coverage error for two-sided normal approximation confidence intervals is of order n−1. In some sense we can think of the two-sided confidence intervals as containing an implicit skewness correction.

The situation is not as good for one-sided confidence intervals. Consider the one- sided normal approximation confidence interval Iµ(α) = (−∞, ¯X + n−1/2σλα), where Φ(λα) = α. The coverage of Iµ(α) is

P (µ ∈ Iµ(α)) = P (µ ≤ ¯X + n−1/2σλˆ α) = P (Sn≥ −λα) = 1 − (Φ(−λα) + n−1/2p1(−λα)φ(−λα) + o(n−1/2)) = α − n−1/2p1α)φ(λα) + o(n−1/2).

(20)

The coverage for the interval Iµ0(α) = (−∞, ¯X + n−1/2ˆσλα) is analogously found to be α − n−1/2q1α)φ(λα) + o(n−1/2)

Thus, for one-sided confidence intervals, normal approximation gives a coverage error of order n−1/2.

The polynomials p1 and q1 both contain the skewness γ. Thus we see that the skewness of X affects the actual coverage of the normal approximation confidence intervals. In particular, when the skewness is zero the n−1/2 term of the coverage error disappears and when the skewness is large the coverage error might be large. When X is skew it is possible to obtain confidence intervals with a better coverage by correcting for skewness.

Some methods for this are presented next.

We will assume that the variance σ2 and the skewness γ are known. It might seem like a bit of a contradiction that the second and third central moments are known, but not the mean. An example where such a situation could occur is when a measuring instrument that has been used sufficiently much, so that the variance and skewness of it’s measures are known, is used to measure something that has not been measured before. One could of course argue that in that case the distribution, or at least the quantiles, of the measurement errors might be known as well and that a parametric confidence interval would make more sense. Let us however assume that the quantiles are unknown and that the density function is unkown or too complicated to work with for such procedures to be fruitful.

3.2 Edgeworth correction

In many cases we wish to derive a confidence interval using some statistic An. In cases where the Edgeworth expansion for Anis known, we can obtain confidence intervals with a coverage error of a smaller order than that of the normal approximation interval. In particular, we can make an explicit correction for skewness using the following theorem, various versions of which were proved in [27], [19], [29] and [1].

Theorem 3. Let An be either Sn= n1/2( ¯X − µ)/σ or Tn= n1/2( ¯X − µ)/ˆσ and assume that An admits the Edgeworth expansion

P (An≤ x) = Φ(x) + n−1/2a1(x)φ(x) + o(n−1/2).

Then

P

An≤ x − n−1/2a1(x)

= Φ(x) + o(n−1/2) (12)

and

P



An≤ x − n−1/2ˆa1(x)



= Φ(x) + o(n−1/2)

where ˆa1 is the polynomial a1 with population moments replaced by sample moments.

References

Related documents

In regard to the severe overconfidence observed with interval production, the previous explanations such as the anchoring-and- adjustment heuristic (Tversky &amp; Kahneman,

The article presents, in a brief conceptual introduction, the option for simulation, not only as economical and statistical alternative but also as conceptual and technical

46 Konkreta exempel skulle kunna vara främjandeinsatser för affärsänglar/affärsängelnätverk, skapa arenor där aktörer från utbuds- och efterfrågesidan kan mötas eller

The increasing availability of data and attention to services has increased the understanding of the contribution of services to innovation and productivity in

Av tabellen framgår att det behövs utförlig information om de projekt som genomförs vid instituten. Då Tillväxtanalys ska föreslå en metod som kan visa hur institutens verksamhet

Regioner med en omfattande varuproduktion hade också en tydlig tendens att ha den starkaste nedgången i bruttoregionproduktionen (BRP) under krisåret 2009. De

Generella styrmedel kan ha varit mindre verksamma än man har trott De generella styrmedlen, till skillnad från de specifika styrmedlen, har kommit att användas i större

Based on feedback information from the error correction process in the receiver, the sender adjusts the amount of redundancy using a so called extremum-seeking controller, which do