• Aucun résultat trouvé

Comparison between two types of large sample covariance matrices

N/A
N/A
Protected

Academic year: 2022

Partager "Comparison between two types of large sample covariance matrices"

Copied!
23
0
0

Texte intégral

(1)

www.imstat.org/aihp 2014, Vol. 50, No. 2, 655–677

DOI:10.1214/12-AIHP506

© Association des Publications de l’Institut Henri Poincaré, 2014

Comparison between two types of large sample covariance matrices

Guangming Pan

Division of Mathematical Sciences, School of Physical and Mathematical Sciences, Nanyang Technological University, Singapore 637371.

E-mail:gmpan@ntu.edu.sg

Received 16 March 2011; revised 18 June 2012; accepted 21 June 2012

Abstract. Let {Xij},i, j = · · ·,be a double array of independent and identically distributed (i.i.d.) real random variables with EX11=μ, E|X11μ|2=1 andE|X11|4<∞. Consider sample covariance matrices (with/without empirical centering)S=

1 n

n

j=1(sj− ¯s)(sj− ¯s)T andS= n1n

j=1sjsTj, where¯s= 1nn

j=1sj andsj=T1/2n (X1j, . . . , Xpj)T with(T1/2n )2=Tn, non-random symmetric non-negative definite matrix. It is proved that central limit theorems of eigenvalue statistics ofSandS are different asn→ ∞withp/napproaching a positive constant. Moreover, it is also proved that such a different behavior is not observed in the average behavior of eigenvectors.

Résumé. Soit{Xij},i, j=1,2, . . . ,un tableau à double entrées, lesXij étant des variables aléatoires réelles indépendantes et identiquement distribuées (i.i.d.) et oùEX11=μ,E|X11μ|2=1 etE|X11|4<∞. Considérons les matrices de covariances empiriques suivantes (avec/sans centrage empirique):S=n1n

j=1(sj− ¯s)(sj− ¯s)T etS=1nn

j=1sjsTj, avecs¯= 1nn j=1sjet sj=T1/2n (X1j, . . . , Xpj)T, où(T1/2n )2=Tnest une matrice déterministe définie positive. Nous démontrons que, sous le régime asymptotiquen→ ∞etp/nconverge vers une constante positive, le théorème central limite pour la statistiqueSest différent de celui concernant la statistiqueS. En outre, nous montrons que cette différence de comportement n’est pas observée pour le comportement moyen des vecteurs propres.

MSC:15A52; 60F15; 62E20; 60F17

Keywords:Central limit theorems; Eigenvectors and eigenvalues; Sample covariance matrix; Stieltjes transform; Strong convergence

1. Introduction

Let {Xij},i, j = · · ·,be a double array of independent and identically distributed (i.i.d.) real random variables with EX11=μ, E|X11μ|2=1 and E|X11|4<∞. Write xj =(X1j, . . . , Xpj)T,sj=T1/2n xj andXn=(s1, . . . ,sn) where(T1/2n )2=Tn, non-random symmetric non-negative definite matrix. It is well known that the sample covariance matrix is defined by

S=1 n

n j=1

(sj− ¯s)(sj− ¯s)T,

(2)

wheres¯= 1nn

j=1sj =T1/2n x¯ with x¯= 1nn

j=1xj. It lies at the roots of many methods in multivariate analysis (see [1]). However, the commonly used sample covariance matrix in random matrix theory is

S=1 n

n j=1

sjsTj.

Since the matrixShas been well studied in the literature (see [5]), a natural question is that whether the asymptotic results for eigenvalues and (or) eigenvectors ofSapply for the matrixS as well. This paper attempts to address it.

First, observe that

S=S− ¯s¯sT (1.1)

and thus by the rank inequality (see Theorem A.44 in [5]) there is no difference when the limiting empirical spectral distribution (ESD) of eigenvalues is our concern only. Here the ESD of a matrixBis defined by

FB(x)= 1 p

p i=1

I (μix), (1.2)

whereμ1, . . . , μpare the eigenvalues ofB. WhenFTnconverges weakly to a non-random distributionHandp/nc >0, Marcenko and Pastur [14], Yin [23] and Silverstein [19] proved that, with probability one, F(1/n)XTnXn(x) converges in distribution to a nonrandom distribution functionFc,H(x)whose Stieltjes transformm(z)=mFc,H(z) is, for eachzC+= {zC: z >0}, the unique solution to the equation

m= −

zc

tdH (t ) 1+t m

1

. (1.3)

Here the Stieltjes transformmF(z)for any probability distribution functionF (x)is defined by mF(z)=

1

xzdF (x), zC+. (1.4)

Noting the relationship between the spectra ofSandn1XTnXn, we have m(z)= −1−c

z +cm(z), (1.5)

wherem(z)is the Stieltjes transform ofFc,H(x), the limit ofFS.

Particularly, when the population matrixTnis the identity matrix the limiting ESD ofS is Marcenko and Pastur’s lawFc(x)(see [14] and [12]), which has a density function

pc(x)=

(2πcx)1

(bx)(xa), axb,

0, otherwise

and has point mass 1−c1at the origin ifc >1, wherea=(1−√

c)2andb=(1+√ c)2.

Secondly, whenTn=I, it was reported in [11] that the maximum eigenvalues ofS andS have the same limit, while, [22] announced that the minimum eigenvalues ofSandSalso share the identical limit. That is,

λmax(S)−→a.s. (1+√

c)2 (1.6)

and whenc <1

λmin(S)−→a.s. (1−√

c)2, (1.7)

(3)

whereλmax(S)andλmin(S)denote, respectively, the maximum and minimum eigenvalues ofS. Moreover, it was proved in [10] that asymptotic distribution of the maximum eigenvalues ofSafter centering and scaling is the Tracy–

Widom law of order 1 whenShas a Wishart distributionWp(n,I). In this case, note thatnShas the same distribution asn1

j=1xjxTj and hence it is not difficult to prove that the maximum eigenvalues ofSandSshare the same central limit theorem.

Thirdly, remarkable central limit theorems for the eigevalues ofSwere established in [3]. Moreover, the constraint that EX114 =3, imposed in [3], was removed in [16]. Since the ESDs ofSandS have the same limit it seems that they might have the same central limit theorem for eigenvalues. Surprisingly, a detailed investigation shows they have different central limit theorems. The main reason is thats¯involved inS contributes to the central limit theorem as well. Formally, let

Gn(x)=p

FS(x)Fcn,Hn(x) ,

wherecn=pn andFcn,Hn(x)denotes the function by substitutingcnforcandHnforHinFc,H(x).

Theorem 1. Suppose that

(1) For each n Xij =Xnij, i, j=1,2, . . . ,are i.i.d.real random variables withEX11=μ, E|X11μ|2=1 and E|X11|4<∞.

(2) limn→∞p

n=c(0,).

(3) g1, . . . , gkare functions onRanalytic on an open regionDof the complex plane,which contains the real interval

lim inf

n λmin(Tn)I(0,1)(c)(1−√

c)2,lim sup

n

λmax(Tn)(1+√ c)2

.

(4) LetTnbe ap×pnon-random symmetric non-negative definite matrix with spectral norm bounded above by a positive constant such thatHn=FTnconverges weakly to a non-random distributionH.

(5) Letei be the vector of sizep×1with theith element1and others0.The matrixTnalso satisfies 1

n n i=1

eTi T1/2n

m(z1)Tn+I 1T1/2n eieTi T1/2n

m(z2)Tn+I 1T1/2n eih1(z1, z2)

and 1 n

n i=1

eTi T1/2n

m(z)Tn+I 1T1/2n eieTi T1/2n

m(z)Tn+I2T1/2n eih2(z).

Then(

g1(x)dGn(x), . . . ,

gk(x)dGn(x))converges weakly to a Gaussian vector(Xg1, . . . , Xgk),with mean

EXg= − c 2πi

g(z)

m3(z)t2dH (t) (1+t m(z))3

(1c m2(z)t2dH (t) (1+t m(z))2 )2

dz−c(E(X11μ)4−3) 2πi

g(z) m3(z)h2(z) 1−c m2(z)t2dH (t)

(1+t m(z))2

dz

+ c 2πi

g(z)

m(z) tdH (t)

(1+t m(z))2

z(1c (m(z))2t2dH (t)

(1+t m(z))2 )

dz (1.8)

and covariance function Cov(Xg1, Xg2)= − 1

2

g1(z1)g2(z2) (m(z1)m(z2))2

d dz1

m(z1) d dz2

m(z2)dz1dz2

c(E(X11μ)4−3)

2 g1(z1)g2(z2) d2 dz1dz2

m(z1)m(z2)h1(z1, z2)

dz1dz2. (1.9)

(4)

The contours in(1.8)and(1.9)are both contained in the analytic region for the functionsg1, . . . , gkand both enclose the support ofFcn,Hn(x)for largen.Moreover,the contours in(1.9)are disjoint.

Remark 1. From Theorem1.1of[3],Theorem1.4of[16]and Theorem1we see that functionals of eigenvalues ofS andS have the same asymptotic variances but they have different asymptotic means.Indeed,the additional term in the asymptotic mean is the last term of(1.8),which can be further written as

c 2πi

g(z)

m(z) tdH (t)

(1+t m(z))2

z(1c (m(z))2t2dH (t)

(1+t m(z))2 )

dz= −1 π

b

a

g(x)arg xm(x)

dx, (1.10)

wherem(x)=limzxm(z)for0 =x∈R.Here one should note that that Theorem1.1of[3]did not handle the case when the common mean of the underlying random variables is not equal to zero.

Remark 2. When developing central limit theorems,it is necessary to study the limits of the products of(E(X11μ)4−3)and the diagonal elements of (SzI)1.For general Tn,the limits of such diagonal elements are not necessary the same.That is why Assumption5has to be imposed.Roughly speaking,Assumption5is some kind of rotation invariance.For example, it is satisfied ifTn is the inverse of another sample covariance matrix with the population matrix being the identity.However,whenE(X11μ)4=3,Assumption5can be removed as well because such diagonal elements disappear in this case thanks to the special fourth moment and one may see[3].Also,when Tnis a diagonal matrix Assumption5in Theorem1is automatically satisfied,as pointed out in[16],and in this case

h1(z1, z2)=

t2dH (t )

((m(z1)t+1))((m(z2)t+1)), h2(z)=

t2dH (t ) (m(z)t+1)3. Particulary,whenTn=Iwe have

EXg= 1 2π

b

a

g(x)arg

1− cm2(x) (1+m(x))2

dx− 1

π b

a

g(x)arg xm(x)

dx

c(E(X11μ)4−3) π

b a

g(x)

m3(x)

(m(x)+1)[(m(x)+1)2cm2(x)]

dx, (1.11)

where

m(x)=−(x+1−c)+√

(xa)(bx)i

2x ,

(see[3]).For some simplified formulas for covariance function,refer to(1.24)in[3]and(1.24)in[16].

Finally, let us take a look at the eigenvectors ofSandS. The matrix of orthonormal eigenvectors ofShas a Haar measure on the orthogonal matrices whenX11 is normally distributed and Tn=I. It is conjectured that for large n and for generalX11 the matrix of orthonormal eigenvectors of S is close to being Haar distributed. Silverstein (1981) created an approach to address it. Further work along this direction can be found in Silverstein [18], [20], Bai, Miao and Pan [4] and Ledoit and Peche [13]. Here we would also like to point out that there are similar interests in eigenvectors of large Wigner matrix and one may see [6,7,9] and [21]. To consider the matrixS, let UnΛnUnT denote the spectral decomposition ofS, whereΛn=diag(λ1, λ2, . . . , λp), Un=(uij)is an orthogonal matrix con- sisting of the orthonormal eigenvectors ofS. Assume thatxn∈Rp,xn =1,is an arbitrary non-random unit vector andy=(y1, y2, . . . , yp)T =UnTxn.Following their approach, we define an empirical distribution function based on eigenvectors and eigenvalues as

F1S(x)= p i=1

|yi|2I (λix), (1.12)

(5)

whereλisare eigenvalues ofS. Based on the above empirical spectral distribution function, we may further investigate central limit theorems of functions of eigenvalues and eigenvectors.

As will be seen below, the asymptotic properties of the eigenvectors matrix ofSare the same as those ofSin terms of this property (note: EX11=0 required forS). However, the advantage ofS overSis that the common mean of underlying random variables is not necessarily zero to keep these types of properties. Unfortunately, this is not the case forS. For example, considereT1Se1whenXij are i.i.d. withEX11=μand Var(X11)=1, andTn=I. Then it is a simple matter to show that

eT1Se1

−→a.s. 1+μ2

which is dependent onμ, different from Corollary 1 in Bai, Miao and Pan [4] whenμ =0. Indeed, when dealing with central limit theorems of functions of eigenvalues and eigenvectors, [18] also proved thatEX11=0 is a necessary condition to keep this type of property for the matrixSwithTn=I.

Formally, let Gn1(x)=√

n

F1S(x)Fcn,Hn(x) .

Theorem 2. In addition to Conditions(1), (2)and(4)in Theorem1suppose thatxn∈Rp1= {x∈Rp,x =1}and xTn(TnzI)1xnmH(z)wheremH(z)denotes the Stieltjes transform ofH (x).Then,it holds that

F1S(x)Fc,H(x) a.s.

By Theorem2, we obtain

Corollary 1. Let(Sm)ii, m=1,2, . . . ,denote theith diagonal elements of matrices Sm.Under the conditions of Theorem2forxn=ei,then for any fixedm,

nlim→∞

Sm ii

xmdFc,H(x)

→0, a.s.

Theorem 3. In addition to the assumptions in Theorem1and Theorem2suppose that asn→ ∞ sup

z

n xTn

mF

c,H(z)Tn+I 1xn

dHn(t) mF

c,H(z)t+1

→0. (1.13)

Then the following conclusions hold.

(a) The random vector

g1(x)dGn1(x), . . . ,

gk(x)dGn1(x)

forms a tight sequence.

(b) IfE(X11μ)4=3, the above random vector converges weakly to a Gaussian vectorXg1, . . . , Xgk,with zero means and covariance function

Cov(Xg1, Xg2)= − 1

2 g1(z1)g2(z2) (z2m(z2)z1m(z1))2

c2z1z2(z2z1)(m(z2)m(z1))dz1dz2, (1.14) where the contours in(1.14)are the same as in those in(1.9).

(c) Instead ofE(X11μ)4=3,assuming that maxi

eiT1/2n

zm(z)Tn+zI 1xn→0, (1.15)

the assertions in(b)still holds.

(6)

Remark 3. Assumptions(1.13)and(1.15)are imposed for the same reason as given in Remark2.Besides,whenTn

reduces to the identity matrix, (1.15)becomes

maxi |xni| →0. (1.16)

As can be seen from(1.14)the asymptotic covariance does not depend on the fourth moment ofX11 and hence we have to impose the condition like(1.16)to ensure that the term involving the product of the fourth moment ofX11and the diagonal elements of(SzI)1converges to zero in probability.One may see[16]for more details.

Remark 4. As pointed out in[4],whenTn=I condition(1.13)holds and formula(1.14)becomes Cov(Xg1, Xg2)=2

c

g1(x)g2(x)dFc(x)

g1(x1)dFc(x1)

g2(x2)dFc(x2)

. (1.17)

Theorem1essentially boils down to the study of the Stieltjes transform ofFS(x)while Theorems2and3boils down to the study of the Stieltjes transform ofF1S(x). In view of this we conjecture the same phenomenon holds for other objects involvingFS(x)orF1S(x). For example we believe central limit theorems of the normalized empirical spectral distribution functions ofSandSare different ([17] considers central limit theorems of the smoothed empirical spectral distribution functions).

We conclude this section by stating the structure of this work. Section 2 gives the proof of Theorem1. Sections 3 and 4 pick up the proofs of Theorems2and3, respectively. Throughout the paperMandC denote constants which may change from line to line, and all matrices and vectors are denoted by boldface letters.

2. Proof of Theorem1and (1.11)

The proof of Theorem1is essentially based on the Stieltjes transform following [3]. First, by analyticity of functions it is enough to consider the Stieltjes transforms of the empirical functions of sample covariance matrices. Moreover, note that the Stieltjes transform of the empirical function ofS can be decomposed as the sum of the Stieltjes transform of the empirical function ofS and some random variable involving sample mean andS. Thus, our main aim is to prove that the extra random variable converges in probability in theCspace to some nonrandom variable. Once this is finished, Theorem1follows from Slutsky’s theorem, the continuous mapping theorem and the results in [3] and [16].

Before proceeding we observe the following useful fact. By the structure ofS, without loss of generality, we can assumeEX11=0, EX211=1 in the course of establishing Theorems1–3 (but the fourth moment will beE(X11μ)4).

2.1. Truncation of underlying random variables and random processes

We begin the proof with the replacement of the underlying random variablesXij with truncated and centralized variables. To this end, write

S=1

n(XnB)

XTnBT ,

whereB=(s,¯ s, . . . ,¯ ¯s)p×n.Since the argument for (1.8) (and two lines below (1.8)) in [3] can be carried over directly to the present case we can choose a positive sequenceεnsuch that

εn→0, εn4E|X11|4I

|X11| ≥εn

n →0, εnn1/4→ ∞. (2.18)

LetSˆ=(1/n)(Xˆn− ˆB)(XˆTn − ˆBT)whereXˆn andBˆ are respectively obtained fromXn andBwith the entriesXij

replaced byXijI (|Xij|< εn

n). We then obtain P (Sˆ =S)P

ip,jn

|Xij| ≥εn

npnP

|X11| ≥εnn

n4E|X11|4I

|X11| ≥εn

n →0, (2.19)

(7)

where and in what followsMdenotes a constant which may change from line to line.

DefineS˜=(1/n)(X˜n− ˜B)(X˜Tn − ˜BT)whereX˜nandB˜ are respectively obtained fromXn andBwith the entries Xij replaced by(XˆijEXˆij)/σnwithXˆij=XijI (|Xij|< εn

n)andσn2=E| ˆXijEXˆij|2. Denote byG˜n(x)and Gˆn(x)the analogues ofGn(x)with the matrixS replaced bySˆandS˜respectively. As in [3] (one may also refer to the proof of Corollary A.42 of [5]) we have for anyg(x)∈ {g1(x), . . . , gk(x)},

gj(x)dG˜n(x)

gj(x)dGˆn(x)

M n k=1

λSk˜λSkˆ

M

n tr

(Xˆn− ˆB)(X˜n− ˜B) (Xˆn− ˜Xn)(Bˆ− ˜B)T1/2

tr(S˜+ ˆS)1/2

=n

1− 1

σn

[trS˜]1/2

tr(S˜+ ˆS)1/2

M

σn2−1 max(S˜)1/2

max(S˜)+max(Sˆ)1/2

, where the third step uses the fact that by the structure of(Xˆn− ˆB),

1 n

(Xˆn− ˆB)(X˜n− ˜B) (Xˆn− ˜Xn)(Bˆ− ˜B)T =σn2

1− 1 σn

2

S˜.

From [23] andTnM we obtain lim supnλmax(S˜)is bounded byM(1+√

c)2with probability one. Similarly, lim supnλmax(Sˆ)is bounded byM(1+√

c)2with probability one by the structure ofSˆand estimate (2.20) below.

Moreover it is not difficult to prove that σn2−1=o

ε2nn1 . (2.20)

Summarizing the above we obtain

g(x)dGˆn(x)

g(x)dG˜n(x)−→i.p. 0.

This, together with (2.19), implies that

g(x)dGn(x)

g(x)dG˜n(x)−→i.p. 0.

Therefore, in what follows, we may assume that

|Xij| ≤√

n, EXij=0, E|Xij|2=1 (2.21)

and for simplicity we shall suppress all subscripts or superscripts on the variables.

As in [3], by Cauchy’s formula, (1.6) and (1.7), with probability one for allnlarge,

g(x)dGn(x)= − 1 2πi

g(z)

tr(SzI)1nmcn(z) dz, (2.22)

wheremcn(z)is obtained fromm(z)withcreplaced bycnand the complex integral is overC. The contourCis speci- fied below. Letv0>0 be arbitrary and setCu= {u+iv0, u∈ [ul, ur]},whereur>lim supnλmax(Tn)(1+√

c)2and 0< ul<lim infnλmin(Tn)I(0,1)(c)(1−√

c)2orulis any negative number if lim infnλmin(Tn)I(0,1)(c)(1−√ c)2=0.

Then define C+=

ul+iv: v∈ [0, v0]

Cu

ur+iv: v∈ [0, v0]

(8)

and letCbe the symmetric part ofC+about the real axis. Then setC=C+C. Moreover, since tr(SzI)1=trA1(z)+ s¯TA2(z)s¯

1− ¯sTA1(z)¯s, we have

tr(SzI)1nmcn(z) =

trA1(z)nmcn(z)

+ s¯TA2(z)s¯

1− ¯sTA1(z)¯s, (2.23)

whereA1(z)=(SzI )1. The first term on the right-hand side above was investigated in [3].

We next consider the second term on the right-hand side in (2.23). To this end, introduce a truncation version of

¯

sTA2(z)s. Define¯ Cr = {ur+iv: v∈ [n1ρn, v0]}, Cl= ul+iv:v∈ [n1ρn, v0

iful>0, ul+iv:v∈ [0, v0]

iful<0, where

ρn↓0, ρnnα, for someα(0,1). (2.24)

LetCn+=ClCuCr andCn denote the symmetric part of Cn+ with respect to the real axis. We then define the truncated process¯sTA2(z)¯sof the processs¯TA2(z)s¯forz=u+ivby

¯ sTA2(z)s¯=

⎧⎪

⎪⎩

¯

sTA2(z)s¯ forzCn=Cn+Cn,

nv+ρn

n s¯TA2(zr1)s¯+ρnnnvs¯TA2(zr2)s¯ foru=ur, v

n1ρn, n1ρn ,

nv+ρn

n s¯TA2(zl1)s¯+ρnnnvs¯TA2(zl2)s¯ foru=ul>0, v∈

n1ρn, n1ρn ,

wherezr1 =ur−in1ρn, zr2 =ur+in1ρn, zl1=ul−in1ρnandzl2=ul+in1ρn. Similarly one may define the truncation version¯sTA1(z)¯sof¯sTA1(z)¯s.

From Theorem 2 in [15] we have

¯sT¯sTnx¯Tx¯≤Mx¯Tx¯−→i.p. Mc. (2.25)

Note that

¯

sT(SzI )1s¯= ¯sTA1(z)¯s+ (s¯TA1(z)s)¯ 2 1− ¯sTA1(z)¯s, where we use the identity

rT

D+arrT 1= rTD1

1+arTD1r, (2.26)

whereDandD+arrT are both invertible,r∈Rpanda∈R. This implies that 1

1− ¯sTA1(z)s¯=1+ ¯sT(SzI )1s.¯ (2.27)

We then conclude that 1

1− ¯sTA1(z)¯s ≤1+

M

v0+ M

|λmax(S)ur|+ M

|λmin(S)ul|

¯

sTs.¯ (2.28)

(9)

This, together with (2.25), (1.6) and (1.7), ensures that

g(z)

¯sTA2(z)s¯ 1− ¯sTA1(z)¯s

s¯TA2(z)¯s 1− ¯sTA1(z)¯s

dz

g(z)

(¯sTA2(z)¯s− ¯sTA2(z)¯s) 1− ¯sTA1(z)s¯

dz

+

g(z)

s¯TA2(z)s(¯ ¯sTA1(z)¯s− ¯sTA1(z)¯s) (1− ¯sTA1(z)s)(1¯ − ¯sTA1(z)s)¯

dz

≤[(¯sT¯s)2+(¯sTs)¯ 4]n n

1

|λmax(S)ur|+ 1

|λmin(S)ul|

1

|λmax(S)ur|+ 1

|λmin(S)ul|

, (2.29) converging to zero in probability.

The aim of subsequent Sections 2.2–2.4 is to prove convergence ofs¯TAj(z)¯s, j =1,2,in probability to non- random variables in the space of continuous functions forzC. It is well known that convergence in probability is equivalent to convergence in distribution when the limiting variable is nonrandom (for example see page 25 of [8]).

Thus, instead, we shall prove convergence ofs¯TAj(z)¯s, j=1,2,in distribution to nonrandom variables in the space of continuous functions forzC.

2.2. Convergence of finite-dimensional distributions of¯sTAj(z)s¯−E(¯sTAj(z)¯s),j=1,2

ConsiderzCuand¯sTA2(z)¯sE(s¯TA2(z)s)¯ first. Define theσ-fieldFj=σ (s1, . . . ,sj), and letEj(·)=E(·|Fj) andE0(·)be the unconditional expectation. Introduces¯j= ¯sn1sj,

Aj1(z)=

Sn1sjsTjzI 1, γj(z)=1

nsTjAj1(z)sjE1

ntrAj1(z)Tn,

βj(z)= 1

1+(1/n)sTjAj1(z)sj

, b1(z)= 1

1+(1/n)Et rA11(z)Tn

.

Write

¯

sTA2(z)¯sE

¯

sTA2(z)¯s = n j=1

Ej

s¯TA2(z)s¯ −Ej1

¯sTA2(z)¯s

= n j=1

(EjEj1)

¯

sTA2(z)¯s− ¯sTjAj2(z)¯sj

= n j=1

(EjEj1)[dn1+dn2+dn3],

where

dn1=(¯s− ¯sj)TA2(z)s,¯ dn2= ¯sTj

A2(z)Aj2(z)s,¯ and

dn3= ¯sTjAj2(z)(¯s− ¯sj).

(10)

Considerdn3first. It is proved in (3.12) of [15] that ExT1B¯x1k=O

n(k/2)2εkn4 fork≥4, (2.30)

wherex1andx¯j= ¯xn1xjare independent ofBwith a bounded spectral norm. This implies that EsT1Aj2(z)¯s1k=ExT1T1/2n Aj2(z)T1/2n x¯1k=O

n(k/2)2εnk4 fork≥4, (2.31)

where yields n j=1

(EjEj1)dn3−→i.p. 0.

Furthermore, write

dn1=dn1(1)+dn1(2)+dn1(3)+dn1(4), where

dn1(1)=1

nsTjAj2(z)¯sj, dn1(2)= 1

n2sTjAj2(z)sj

and

dn1(3)=1 nsTj

A2(z)Aj2(z) s¯j, dn1(4)= 1 n2sTj

A2(z)Aj2(z) sj. From (2.31) we have

n j=1

(EjEj1)dn1(1)−→i.p. 0.

By Lemma 2.7 in [3] and (2.21) nkEsT1B(z)s1−trBTk=O

ε2kn 4n1 , k≥2. (2.32)

This gives E

n j=1

(EjEj1)dn1(2)

2

= 1 n4

n j=1

E(EjEj1)

sTjAj2(z)sj−trTnAj2(z) 2=O n2 .

SinceA2(z) ≤1/v2we obtain E

n j=1

(EjEj1)dn1(4)

2

M n4v4

n j=1

Esj4M n. Note that

A1(z)Aj1(z)= −1

nAj1(z)sjsTjAj1(z)βj(z). (2.33)

Applying it we may write dn1(3)= −1

n2sTjAj1(z)sjsTjAj2(z)s¯jβj(z)+ 1

n3sTjAj1(z)sjsTjAj2(z)sjsTjAj1(z)¯sjβj2(z)

− 1

n2sTjAj2(z)sjsTjAj1(z)¯sjβj(z).

Références

Documents relatifs

Sample covariance matrices, Spiked population model, limit theorems, Largest eigenvalue, Extreme eigenvalues.. The research of this Zhidong Bai was supported by CNSF grant 10571020

Based on these evidentiary support and motivated by the fact that the eigenmatrix of Wishart matrix is Haar (uniformly) distributed, we believe that the eigenmatrix of a

In Section 2 we state the assumptions and main results of the article: Proposition 2.1 and Theorem 2.2 are devoted to the limiting behaviour and fluctuations of λ max (S N );

Under a mild dependence condition that is easily verifiable in many situations, we derive that the limiting spectral distribution of the associated sample covariance matrix

The recent references [BFG + 15] and [FBD15] provide asymptotic results for maximum likelihood approaches, and require that the smallest eigenvalue of the covariance matrix of

Various techniques have been developed toward this goal: moment method for polynomial functions ϕ (see [2]), Stieltj`es transform for analytical functions ϕ (see [4]) and

The contribution to the expectation (12) of paths with no 1-edges has been evaluated in Proposition 2.2. This section is devoted to the estimation of the contribution from edge paths

We establish a large deviation principle for the empirical spec- tral measure of a sample covariance matrix with sub-Gaussian entries, which extends Bordenave and Caputo’s result