• Aucun résultat trouvé

Zero bias transformation and asymptotic expansions

N/A
N/A
Protected

Academic year: 2022

Partager "Zero bias transformation and asymptotic expansions"

Copied!
24
0
0

Texte intégral

(1)

www.imstat.org/aihp 2012, Vol. 48, No. 1, 258–281

DOI:10.1214/10-AIHP384

© Association des Publications de l’Institut Henri Poincaré, 2012

Zero bias transformation and asymptotic expansions

Ying Jiao

Laboratoire de probabilités et modèles aléatoires, Université Paris VII, Paris, France. E-mail:jiao@math.jussieu.fr Received 10 May 2010; revised 17 June 2010; accepted 15 July 2010

Abstract. LetWbe a sum of independent random variables. We apply the zero bias transformation to deduce recursive asymptotic expansions forE[h(W )]in terms of normal expectations, or of Poisson expectations for integer-valued random variables. We also discuss the estimates of remaining errors.

Résumé. SoitWune somme de variables aléatoires indépendants. On applique la transformation zéro biais pour obtenir de façon recursive des développements asymptotiques deE[h(W )]en terme d’espérances par rapport à la loi normale, ou à la loi de Poisson si les variables aléatoires sont à valeurs entières. On discute aussi les bornes des termes d’erreur.

MSC:60G50; 60F05

Keywords:Normal and Poisson approximations; Zero bias transformation; Stein’s method; Reverse Taylor formula; Concentration inequality

1. Introduction

The zero bias transformation has been introduced by Goldstein and Reinert [15] in the framework of Stein’s method.

By the fundamental works of Stein [24,25], we know that a random variable (r.v.) Z with mean zero follows the normal distributionN (0, σ2)if and only ifE[Zf (Z)] =σ2E[f(Z)]for any Borel functionf such that both sides of the equality are well defined. More generally, for any r.v.Xwith mean zero and finite varianceσ2>0, a r.v.Xis said to have the zero biased distribution ofXif the equality

E Xf (X)

=σ2E f

X

(1) holds for any differentiable functionf such that (1) is well defined. ByStein’s equation

xf (x)σ2f(x)=h(x)Φσ(h), (2)

wherehis a given function andΦσ(h)denotes the expectation ofhunder the distributionN (0, σ2), E

h(X)

Φσ(h)=E

Xfh(X)σ2fh(X)

=σ2E fh

X

fh(X)

, (3)

wherefhis the solution of Stein’s equation and is given by fh= 1

σ2φσ(x)

x

h(t)Φσ(h)

φσ(t)dt. (4)

An important remark is thatXneeds not be independent ofX([15], see also [13]). LetW=X1+ · · · +XnwhereXi (i=1, . . . , n)are independent random variables with mean zero and varianceσi>0,σW2 :=Var[W] =σ12+ · · · +σn2. Goldstein and Reinert have proposed the constructionW=W(I )+XI whereW(i):=WXi,Xiis independent

(2)

ofW(i), andI is a random index taking values in{1, . . . , n}and independent of(X1, . . . , Xn, X1, . . . , Xn)such that P(I=i)=σi2W2. This construction of the zero bias transformation is similar to Lindeberg’s method except that we here consider an average of punctual substitutions ofXi byXi, while in Lindeberg’s method,Xi’s are progressively substituted by a central normal r.v. with the same variance.

The asymptotic expansion of E[h(W )] is a classical topic related to the central limit theorems. Using Stein’s method, Barbour [2,3] has obtained a full expansion ofE[h(W )]for sufficiently regular functions. Compared to the classical Edgeworth expansion (see [21], Chapter V, also [22]), the results of [2] do not require that the distributions of Xi’s be smooth. However, as a price paid, we need some regularity conditions on the functionh. The result of [2] can also be compared to those in [17,18] using Fourier transform. The key point of Barbour’s method is a Taylor type formula with cumulant coefficients, which allows to write the differenceE[Wf (W )] −σW2E[f(W )]as a series containing the cumulants ofW. TheW-expectations are then replaced by the normal expectations in an iterative way until the desired order. It has been pointed out in [22] that the key formula of Barbour can also be obtained by Fourier transform.

The zero bias transformation has been used in [11] to obtain a first-order correction term for the normal approxi- mation ofE[h(W )], where the motivation was to find a rapid numerical method for large-sized credit derivatives. The function of interest is the so-calledcall functionin finance:h(x)=(xk)+wherekis a real number. Since suchh is only absolutely continuous, the functionfh is not regular enough to have a third-order derivative. To achieve the estimates, we used a conditional expectation technique, together with a concentration inequality due to Chen and Shao [8,9].

The difficulty for generalizing the result in [11] to obtain a full expansion ofE[h(W )]is thatW andWWare not independent. If we consider the Taylor expansion offh(W)atW and then take the expectation, the terms of the formE[fh(l)(W )(WW )k]appear, wheref(l) denotes thelth derivative off. For the first-order expansion in [11], a conditional expectation argument allowed us to replaceE[fh(W )(WW )]byE[fh(W )]E[WW]and to put the covariance in the error term. However, in the higher-order expansions, the error term can no longer contain such covariances. An alternative way is to consider the Taylor expansion ofE[fh(W)fh(W )]atW(i). SinceXi is independent ofW(i), there is no crossing term. However, the expectations of the formE[fh(l)(W(i))]appear, which makes it difficult to apply the recurrence procedure. To overcome this difficulty, we shall propose a so-called reverse Taylor formula which will enable us to replaceE[fh(l)(W(i))]byE[fh(l)(W )]up to an error term.

The main result of this paper is an expansion formulaE[h(W )] =CN(h)+eN(h),CN(h)being the principal term of theNth-order expansion andeN(h) being the remaining error. We shall precise the conditions under which the expansion formula is well defined. In particular, in order to include some irregular functions, we shall consider the continuous and the jump parts ofhseparately. Moreover, we also need to show thateN(h) is “small” enough as an error term. In our results,eN(h)is given in a recursive way so that it is actually a linear combination of the remaining terms of Taylor and reverse Taylor expansions and can be thus estimated. A key ingredient for the estimates is a concentration inequality which provides an upper bound for P(aWb) involving some power(ba)α of the interval length(ba)with 0< α≤1. This allows to obtain estimates under relatively mild moment conditions on Xi’s. It is interesting to point out that the same method can be used in a discrete setting to obtain a Poisson asymptotic expansion for integer-valued summand random variables.

The rest of the paper is organized as follows. In Section2, we give the normal expansion by using the reverse Taylor formula and we discuss the conditions onhand onXi’s. Section3is devoted to error estimates. We present the Poisson case in Section4. Finally, some technical proofs are left inAppendix.

2. Expansions in normal case

2.1. Reverse Taylor formula

Let N be a positive integer,X andY be two independent random variables such that Y has up toNth moments, andf be anN times differentiable function such thatf(k)(X)andf(k)(X+Y )are integrable for anyk=0, . . . , N.

(3)

We use the notationm(k)Y :=E[Yk]/k!. Denote byδN(f, X, Y )the error term in theNth-order Taylor expansion of E[f (X+Y )]centered atX, i.e.

δN(f, X, Y ):=E

f (X+Y )

N k=0

m(k)Y E

f(k)(X)

. (5)

Recall that for anyN≥1, δN(f, X, Y )= 1

(N−1)! 1

0

(1t )N1E

f(N )(X+t Y )f(N )(X) YN

dt (6)

provided that the right-hand side is well defined.

The so-called reverse Taylor formula gives an expansion ofE[f (X)]using the expectations of functions ofX+Y and the moments ofY. We specify some notation and conventions. First of all, N:=N\ {0} denotes the set of strictly positive integers. For any integerd≥1 and anyJ=(jl)dl=1∈Nd,|J|is defined asj1+ · · · +jd, andm(J)Y :=

m(jY1)· · ·m(jYd). By convention,N0denotes the set{∅}of the empty vector,|∅| =0 andm(Y)=1. We note that in the right–hand side of the expansion formula (7), the variablesX+Y andY are not independent.

Proposition 2.1 (Reverse Taylor formula). With the above notation,the equality

E f (X)

=

d0

(−1)d

J∈Nd,|J|≤N

m(J)Y E

f(|J|)(X+Y )

+εN(f, X, Y ) (7)

holds,whereεN(f, X, Y )is defined as

εN(f, X, Y )= −

d0

(−1)d

J∈Nd,|J|≤N

m(J)Y δN−|J|

f(|J|), X, Y

. (8)

Proof. We replaceE[f(|J|)(X+Y )]on the right-hand side of (7) by

N−|J| k=0

m(k)Y E

f(|J|+k)(X)

+δN−|J|

f(|J|), X, Y

and observe that the sum of terms containingδvanishes withεN(f, X, Y ). Hence the right-hand side of (7) equals

d0

(−1)d

J∈Nd,|J|≤N

m(J)Y

N−|J| k=0

m(k)Y E

f(|J|+k)(X) .

If we split the terms wherek=0 and where 1≤kN− |J|, the above formula can be written as

d0

(−1)d

J∈Nd,|J|≤N

m(J)Y E

f(|J|)(X) +

d0

(−1)d

J∈Nd,|J|≤N

m(J)Y

N−|J| k=1

m(k)Y E

f(|J|+k)(X)

. (9)

We make the index changesJ=(J, k)andu=d+1 in the second part, and observe that it is just

u1

(−1)u1

J∈Nu,|J|≤N

m(JY)E

f(|J|)(X) .

Thus, the terms in the first and the second parts of (9) cancel out except the one whered=0 in the first part. So (9)

equalsE[f (X)], which proves the proposition.

(4)

2.2. Admissible function space

In this subsection, we precise the functions for which we can make theNth-order expansion. We need conditions on both the regularity and the increasing speed at infinity of the function.

Recall ([20], Chapter VI) that any functiongonRwhich is locally of finite variation can be uniquely decomposed into the sum of a function of pure jump and a continuous function locally of finite variation and vanishing at the origin.

That is,g=gc+gd wheregcis called thecontinuous partofgandgdis thepurely discontinuous part.

Letα(0,1]andp≥0 be two real numbers. For any functionf onR, the following quantity has been defined by Barbour in [2]:

fα,p:=sup

x =y

|f (x)f (y)|

|xy|α(1+ |x|p+ |y|p). (10) The finiteness of fα,p implies that the functionf is locally α-Lipschitz, and that the increasing speed off at infinity is at most of order|x|α+p. All functionsf such thatfα,p<+∞form a vector space overR, and · α,p

is a norm on it. We list below several properties of · α,pand we leave the proofs in AppendixA.

Lemma 2.2. Letf be a function onR,α, β(0,1]andp, q≥0.

(1) Ifpq,thenfα,p<+∞impliesfα,q<+∞. (2) Ifαβ,thenfβ,p<+∞impliesfα,p+βα<+∞.

(3) IfP is a polynomial of degreed,thenfα,p<+∞impliesPfα,p+d<+∞.

(4) Assume thatF is a primitive function off,thenfα,p<+∞impliesF1,p+α<+∞. (HenceFα,p+1<

+∞by2.)

Inspired by [2], we introduce the following function space.

Definition 2.3. LetN≥0be an integer,andα(0,1],p≥0be two real numbers.Denote byHNα,pthe vector space of all Borel functionshonRverifying the following conditions:

(a) hhasNth derivative which is locally of finite variation and has finitely many jumps, (b) the continuous part ofh(N )satisfiesh(N )c α,p<+∞.

Condition (a) implies that the pure jump part of h(N ) is bounded. Condition (b) implies that h(N )c has at most polynomial increasing speed at infinity, and so doesh. These conditions allow us to include some irregular functions such as the indicator functions. Letkbe a real number andIk(x)=1{xk}. ThenIkα,pis clearly not finite. However, Ik,cα,p=0, which means that for anyα(0,1]and anyp≥0,Ik(x)H0α,p. Note that any functionhinHα,p0

can be decomposed as h=hc+hd, wherehc satisfieshcα,p<+∞andhd is a linear combination of indicator functions of the form1{xk}plus a constant (so thathc(0)=0).

Proposition 2.4. LetN≥0be an integer,α, β(0,1]andp, q≥0be real numbers.Then the following assertions hold:

(1) whenN≥1,hHNα,pif and only ifhHNα,p1;

(2) ifpq,thenHNα,pHNα,q;ifαβ,thenHβ,pNHNα,p+βα; (3) whenN≥1,HNα,pHN1,α+1pHα,pN+11;

(4) ifhHNα,pand ifP is a polynomial of degreed,thenP hHNα,p+d.

Proof. Assertion (1) results from the definition. Assertions (2)–(4) are consequences of Lemma2.2.

The following result is important and its proof which is rather technical is postponed to AppendixB.

Proposition 2.5. IfhHNα,p,then the solution of Stein’s equation satisfiesfhHNα,p+1.

(5)

2.3. Normal expansion

We consider a family of independent random variablesXi(i=1, . . . , n)with mean zero and finite varianceσi2>0.

LetW=X1+· · ·+XnandσW2 =Var(W ). Denote byXi a random variable which follows the zero-biased distribution ofXiand is independent ofW(i):=WXi.

Theorem 2.6. LetN≥0be an integer,α(0,1]andp≥0.Assume thathHNα,p.LetX1, . . . , Xnbe independent random variables with mean zero and up to(N+max(α+p,2))th moments.ThenE[h(W )]can be written as the sum of two termsCN(h)andeN(h)whereC0(h)=ΦσW(h)ande0(h)=E[h(W )] −ΦσW(h),and recursively forN≥1,

CN(h)=C0(h)+ n

i=1

σi2

d1

(−1)d1

J∈Nd,|J|≤N

m(JX)

i

m(JX)

im(JX)

i

CN−|J|

fh(|J|+1)

, (11)

eN(h)= n i=1

σi2

d1

(−1)d1

J∈Nd,|J|≤N

m(JX)

i

m(JX)

im(JX)

i

eN−|J|

fh(|J|+1) +

N k=0

εNk

fh(k+1), W(i), Xi m(k)X

i +δN

fh, W(i), Xi

, (12)

where for any integerd≥1,and anyJ∈Nd, J∈Ndenotes the last coordinate ofJ,andJdenotes the vector in Nd1obtained fromJby omitting the last coordinate.

Remark 2.7. The moments ofXi in(11)and (12)can be written as moments of Xi.By definition of the zero bias transformation,for anyk∈Nsuch thatXihaskth moment,one has

E Xik

= E[Xki+2] σi2(k+1). Hencem(JX)

i =σi2(J+2)m(JX+2)

i whereJis the strictly positive integer figuring in the last coordinate ofJ.

Proof. We first prove that all terms in (11) and (12) are well defined. WhenN=0,hH0α,pandh(x)=O(|x|α+p).

Hence E[h(W )] and ΦσW(h) are well defined. Assume that we have proved for 0, . . . , N −1. Let hHNα,p. Then by Proposition2.4, h(x)HNα,pHα,pN+11⊂ · · · ⊂H0α,p+N, so h(x)=O(|x|α+p+N). By Proposition 2.5, fhHNα,p+1and by Proposition2.4(1), for any|J| =1, . . . , N,fh(|J|+1)HNα,p−|J|. So the induction hypothesis implies thatCN−|J|(fh(|J|+1))andeN−|J|(fh(|J|+1))are well defined. Furthermore, sincefh(k+1)(x)=O(|x|α+p+Nk)for any k=0, . . . , N, the termsεNkandδN in (12) are well defined. Finally, all moments figuring in (11) and (12) exist.

We then prove the equalityE[h(W )] =CN(h)+eN(h). In the caseN =0, the equality holds by definition. In the following, we assume that the equalityE[h(W )] =Ck(h)+ek(h)has been verified for anyk=0, . . . , N−1. By Stein’s equation (3),E[h(W )] −C0(h)is equal to

σW2E fh

W

fh(W )

= n i=1

σi2 E

fh

W(i)+Xi

−E

fh(W )

. (13)

Consider the following Taylor expansion E

fh

W(i)+Xi

= N k=0

m(k)X

iE fh(k+1)

W(i) +δN

fh, W(i), Xi .

(6)

By replacing E[fh(k+1)(W(i))] in the above formula by its (Nk)th reverse Taylor expansion, we obtain that E[fh(W(i)+Xi)]equals

N k=0

m(k)X

i

d0

(−1)d

J∈Nd|J|≤Nk

m(J)X

iE

fh(|J|+k+1)(W )

+εNk

fh(k+1), W(i), Xi +δN

fh, W(i), Xi .

Note that the term with indexesk=d=0 in the sum inside the bracket isE[fh(W )]. ThereforeE[fh(W(i)+Xi)] − E[fh(W )]can be written as the sum of the following three parts

n k=1

m(k)X

i

d0

(−1)d

J∈Nd,|J|≤Nk

m(J)X

iE

fh(|J|+k+1)(W )

, (14)

d1

(−1)d

J∈Nd,|J|≤N

m(J)X

iE

fh(|J|+1)(W )

, (15)

N k=0

m(k)X

i

εNk

fh(k+1), W(i), Xi +δN

fh, W(i), Xi

. (16)

By interchanging summations and then making the index changesK=(J, k)andu=d+1, we obtain (14)=

u1

(−1)u1

K∈Nu,|K|≤N

m(KX)

i m(KX)

i E

fh(|K|+1)(W ) .

As the equalitym(J)X

i =m(JX)

i m(JX)

i holds for anyJ, (14)+(15) equals

d1

(−1)d1

J∈Nd,|J|≤N

m(JX)

i

m(JX)

im(JX)

i

E

fh(|J|+1)(W ) .

By the hypothesis of induction, we have E[fh(|J|+1)(W )] =CN−|J|(fh(|J|+1))+eN−|J|(fh(|J|+1)). So the equality

E[h(W )] =CN(h)+eN(h)follows.

Denote byμ(k)i thekth moment ofXi. We give the first few explicit expansion formulas below:

C1(h)=ΦσW(h)+C0

fh(2)n

i=1

μ(3)i 2 ; C2(h)=ΦσW(h)+C1

fh(2)n

i=1

μ(3)i 2 +C0

fh(3)n

i=1

μ(4)i

6 −(2)i )2 2

; C3(h)=ΦσW(h)+C2

fh(2)n

i=1

μ(3)i 2 +C1

fh(3)n

i=1

μ(4)i

6 −(2)i )2 2

+C0

fh(4)n

i=1

μ(5)i

24 −5μ(2)i μ(3)i 12

.

Note that inCN(h), there appear normal expectations of the operators which are of the formhfh(l). As pointed out by Barbour [2], p. 294, such expectations can be expressed as those ofhmultiplied by a Hermite polynomial, see also remark below.

Remark 2.8. Let us compare our result with Barbour’s.In the main theorem of[2],the condition on the functionhfor theNth-order expansion is:hisN times differentiable andh(N )α,p<∞.The conditionhHα,pN in Theorem2.6 is weaker since we separate the continuous and the jump parts ofh.An example inH1α,p which does not satisfy the

(7)

condition of[2]ish(x)=(xk)+sincehα,pis infinite.Concerning the summand variables,up to(N+max(α+ p,2))th moments ofXi’s are needed for theNth expansion and up to(N+α+p+2)th moments,which are similar to[2],are required to achieve the error estimates(see Section3.2for details).We consider now the expansion formula.

Letr≥1be an integer.By Remark2.7and the fact thatXi’s are with mean zero,the moment term figuring in(11) satisfies

σi2

d1

(−1)d1

J∈Nd,|J|=r

m(JX)

i

m(JX)

im(JX)

i

=

d1

(−1)d1

J∈Nd,|J|=r

m(JX)

i

J+2 m(JX+2)

i −2m(JX)

i m(2)X

i

=

d1

(−1)d1

J∈Nd,|J|=r

J+2 m(JX)

i m(JX+2)

i

d1

(−1)d1

J∈Nd,|J|=r

2m(JXi)m(JX)

i m(2)X

i

=

d1

(−1)d1

K∈Nd,|K|=r+2 K3

Km(K)X

i +

d1

(−1)d1

K∈Nd,|K|=r+2 K=2

Km(K)X

i

=

K

d≥1Nd

|K|=r+2

(−1) (K)1Km(K)X

i ,

where in the third equality,we use the index changesK=(J,J+2)for the first term andK=(J,2),d=d+1for the second term,and in the last equality, (K)denotes the length of the vectorK.Denote byκi(r)therth cumulant of Xi.By the recurrence formula which relates cumulants and moments(e.g., [23]):

κi(r)=μ(r)i

r1

j=1

r−1 j

μ(j )i κi(rj ), (17)

we obtain

σi2

d1

(−1)d1

J∈Nd,|J|=r

m(JX)

i

m(JX)

im(JX)

i

= κi(r+2)

(r+1)!. (18)

In fact,the formula(17)can also be written as κi(r)

(r−1)!=rm(r)X

i

r1

j=1

m(j )X

i

κi(rj ) (rj−1)!. Denote byηi(r)the sum

|K|=r(−1) (K)1Km(K)X

i .A direct verification showsηi(1)=κi(1)=0and for anyr≥2, rm(r)X

i

r1

j=1

m(j )X

iηi(rj )=rm(r)X

i

r1

j=1

m(j )X

i

|K|=rj

(−1) (K)1Km(K)X

i

=rm(r)X

i

r1

j=1

|K|=rj

(−1) (K)1Km(j,K)X

i

=rm(r)X

i +

|J|=r, (J)2

(−1) (J)1Jm(J)X

i =η(r)i ,

(8)

whereJ=(j,K)in the third equality.By induction,we obtainη(r)i =κi(r)/(r−1)!for anyr∈Nand hence(18).So the expansion formula(11)becomes

CN(h)=C0(h)+ N r=1

CNr

fh(r+1)κ(r+2)(W )

(r+1)! , (19)

whereκ(r)(W )denotes therth cumulant ofW.We thus recover the result of Barbour. 3. Estimates of remaining errors

3.1. Concentration inequalities

We shall prove a concentration inequality which gives an upper bound for P(aWb),a andb being two real numbers. The special point here is to consider a parameterα≤1 and to give a bound depending sub-linearly onba.

The case whereα=1 has been studied in [8,9,11,14]. The following result, measuring the nearness betweenXand X, is useful to prove the concentration inequality.

Lemma 3.1. Letα(0,1]andXbe a r.v.with mean zero,finite varianceσ2>0and up to(α+2)th moments.Let Xhave the zero biased distribution ofXand be independent ofX.Then,for anyε >0,

PX−X> ε

≤ 1

α+1)σ2EXsα+2 ,

whereXs=XXandXis an independent copy ofX.

Proof. Similarly to the Markov inequality, the following inequality holds:

PX−X> ε

≤ 1

εαEX−Xα .

For any locally integrable even functionf, by definition of the zero bias transformation and the fact thatXandX are independent,

E f

XX

= 1 2σ2E

XsF Xs

, whereF (x)=x

0 f (t )dt. In particular, EX−Xα

= 1

2(α+1)σ2EXsα+2 ,

which implies the lemma.

Lemma 3.2. We keep the notation of Theorem2.6and suppose thatf:R→Randg:R2→Rare two functions such that the variance off (W )andg(Xi, Xi)exist fori=1, . . . , n.Then

Cov

f (W ), g

XI, XI≤Var[f (W )]1/2 σW2

n

i=1

σi4Var g

Xi, Xi1/2

. (20)

In particular,for anyε≥0, Cov[1{aWb},1{|XIX

I|≤ε}]≤1 4

n

i=1

σi4 σW4

1/2

. (21)

Références

Documents relatifs

Abstract. In this paper, motivated by some applications in branching random walks, we improve and extend his theorems in the sense that: 1) we prove that the central limit theorem

We study positivity cases, estimates and asymptotic expansions of condenser p-capacities of points and segments in a given bounded domain, having in mind application fields, such

Singular perturbation, combined asymptotic expansion, turning point, canard solution.. 2002 Southwest Texas

Au 31 décembre 2017, 495 conventions collectives de branche – hors branches agricoles – couvrent 15,8 millions de salariés selon les déclarations sociales nominatives (DSN) et

U-Pb analytical data of zircon and rutile from eclogite sample IZ94-74 (Strona-Ceneri Zone, northern Italy)... This unradiogenic Pb isotopic composition does not allow for

A host choice experiment was carried out to determine the host range of some potential biological control agents of yellow starthistle (Centaurea solsti-.. tialis), a

After exclusion of 33 lesions arising in unspecified, multiple or contiguous sites, and 25 melanoma whose site was insufficiently detailed [seven trunk not otherwise specified (NOS),

L’ETT nous a permis de suspecter un LCP, dans le cas n°14, devant une infiltration tumorale du myocarde des cavités cardiaques droites et même du VG donnant un