J EAN -L OUIS P HILOCHE
A new intrinsic construction of the gaussian measure in R
d; with application
Annales scientifiques de l’Université de Clermont-Ferrand 2, tome 98, sérieMathéma- tiques, no28 (1992), p. 167-175
<http://www.numdam.org/item?id=ASCFM_1992__98_28_167_0>
© Université de Clermont-Ferrand 2, 1992, tous droits réservés.
L’accès aux archives de la revue « Annales scientifiques de l’Université de Clermont- Ferrand 2 » implique l’accord avec les conditions générales d’utilisation (http://www.
numdam.org/conditions). Toute utilisation commerciale ou impression systématique est constitutive d’une infraction pénale. Toute copie ou impression de ce fichier doit contenir la présente mention de copyright.
Article numérisé dans le cadre du programme Numérisation de documents anciens mathématiques
http://www.numdam.org/
A NEW INTRINSIC CONSTRUCTION OF THE GAUSSIAN MEASURE IN
Rd;
WITH APPLICATION
1. INTRODUCTION. We
give
in this paper a new intrinsicconstruction of
GQ ,
the(centered) (1) gaussian probability
measure of
covariance Q
in a d- dimensional space.Although
ourapproach
is linked to thetype
of constructionstarting
withthe definition of the
density
whenQ
isinvertible,
we still manage to constructGQ
with a convenientdensity, whether Q
is invertible or not.Then,
using
the tools and results introducedbefore,
anapplication
to statistics isgiven :
a new intrinsic and (in asense)
natural, proof
of the well-known theorem which states that theasymptotic
distribution of theempirical chi-square
isx2 (d-1).
For each of these two
parts
wegive
a short survey of the literature on thesubjet
as well as referencesillustrating
thedifferent
types
of construction andproof
which are essential toour paper.
(1) The presence of a mean in the
(probability) law
of agaussian
random vector X introduces
only
a translation term easy to handle.In the
beginning
of this paper we shall thus, without loss ofgenerality , limit
ourselves to the case where the mean is zero.After the introduction of a few notations we
shall,
via aproposition,
introduce a very useful innerproduct,
denoted(.,.)Q
on ImQ, a
subspace
of the euclidianspace X
in whicheverything happens.
Then we shall find out a suitable
Lebesgue measure (1)
which will enable us to construct thegaussian
distributionGQ’
for anypositive operator.
Let X
denote an innerproduct
space of dimensiond ;
the innerproduct
is(.,.) . let)
is the set of linearaperators
on X. For every T Elet), T*
denotes theadjoint
of T. We shallgive special
interest to
positive operators,
that isQ
esatisfying :
Q=Q*
and(Qx,x) > 0 ,
x .(2.1) Proposition
1. Let Q be apositive operator
in X : : one defines, onImQ, an inner
product,
denoted(.,.)Q , by
the relation :=
(Qx,y),
x,y E X . .(2.2)
In case
rkQ
=d, .,.)Q
is defined on the whole Xby :
= x,y .
(2.3)
For reasons
given
in theproof,
we shall find convenient to say that(.,.)Q
is thereproducing
innerproduct
associated toQ.
Proof. This
proposition
is aparticular
case, but in an intrinsincversion,
of thetheory
ofreproducing
kernels introducedby Aronzsajn
113. This tool has beenpopularized
in statistics andProbability theory by
E. Parzen[23,
so we couldskip
thedemonstration.
Nevertheless,
it is worthgiving
aquick proof
inthe finite dimensional case :
Let Q be a
positive operator in X ;
with thesymetry
of Q, it is firsteasily
checked that(Qx,y) only depends
on Qx andQy.
Bilinearity
andpositivity
of(.,.)Q
are obvious. The fact that(x, x) Q
is0,
on ImQ , iff x =0 ,
resultsdirectly
fromKerQ =
lx;x EX, (Qx,x) =
0)(for
theelementary
results in linearalgebra
admitted assuch,
an excellent reference is still P.Halmos
C3J).
The invertible case followsdirectly
from the firstpart
of theproposition.
0In the
following
section it will be useful to note theequality
(1)
Implicitely,
the Borel «-field willalways
be chosen on innerproduct
spaces.Following
the way indicated in the introduction of thissection,
we have now in hand the suitable innerproduct
on ImQ. We shallproceed by choosing
theconvenient (1) Lebesgue
measure onImQ.
Now, let e =
(el, e2, ... , eq)
be a basis ofImQ ;
it isconvenient to use the same notation e to also denote the linear
application
of~~
into ImQ, which associates toLet L g
denote the usualLebesgue
measure ofRq.
Definition 1. The
Lebesgue
measure on the innerproduct
space ImQ,(.,.)Q
is defined as theimage
ofXq by
theapplication
e(defined above),
where e is a basis of ImQ, orthonormalfor (., .) .
Of course this definition is
easily
seen to be correct, ecause it isindependent
of the choice of the orthonormal basis. Thespecific Lebesgue
measure chosen in this definition will bedenoted
xQ,
....h. t.. d f... f th We are now in
position
togive
the intrinsic definition of thegaussian
random vectors (r.v) we arelooking
for : ,Definition 2. Let Q be a
positive operator in I, rkQ=q.
A r, v x taki values in X is said to begaussian
with Q as a covarianceoperator, if its
(probability)
lawPx
is :The notation
GQ
=Px
will be used.: Now that we have found a law, which is a "candidate" to achieve the construction we aim at, we must check up that
(2.6) coincides,
when 9G =
Rd,
with the similar distribution obtainedby
anotherclassical route.
’
A few remarks must be done befare : Remarks .
1)
Let e be a basis of ImQ , orthonormal for( . , . ~ ~ ;
theapplication
e defined in(2.5)
isclearly
anisometry
ofRq
onImQ.
Then we observe that
G~
is theimage by
e ofGq ,
the standardgaussian
distribution onRq :
(1)
On a euclidian space theLebesgue
measure,being
invariantby
translations,
isunique
up to a constant ; thepoint
isprecisely
to
make anunambiguous
choice.for sake of
clarity (. , . ) q
denotes the natural innerproduct
ofRq,
2)
The Fouriertransform(l) Gq
ofGq
isusing again
theisometry
eeasily gives
a firstexpression
ofthe Fourier transform of
Go :
for any v eImQ,
This formula
gives
the characteristic function of X , insideImQ ;
therefore we have notyet
reached ourgoal
which is toidentify Eexp[i(u,X)3
for every u E~d
5 t ;[here
(., . )d
a(., .) ].
.3)
It should be noted that the lawPX
concernedby (2.~)
and(2.9)
is
always "spherical"
as soon as agood
innerproduct (.,.)Q
isused in the space ImQ in which X
really
takes its values.We shall end our
construction, by proving that,
in the casex =
Rd , G is
identical withG (d)
thegaussian
law withcovariance matrix Q defined
by
any other method in the literature.For this we shall use characteristic functions :
Proposition
2. Let X be a r.v in suppose that itsprobability
law
PX
isGQ ,
then ~ -Proof. As X E
ImQ (a. s) ,
andwith (2.4) ,
using (2.9)
and(2.2) .
aFor further need we shall
give
two usefulpropositions :
( 1 )
InProbability theory
it is classical to define the Fourier transform of aprobability u
inRq
Rq)
. If, asoften,
the law of some r,v X , Theright
hand side of this lastequality
iscalled the characteristic function
(c.fl
of X.Proposition
3.Suppose
theprobability
distribution of a r.v X isGQ,
with rkQ = q , then : X E ImQ a.s, andthe law of
llXil2
isx2 ~q~ (1).
The
proof
is obvious when one uses theisometry
e in(2.5)
andif one relies on the
following
convenient definition ofx2(q) :
Definition 3. Let X be a
gaussian
vector inRq ,
withI (the
, , , , 2 (d) .. ,
,q
identity)
for covariancematrix,
thenx2(d)
isby
definition theprobability
law ofIIXI12
Proposition
4. (Central limit theorem inRd).
Let be i . i . d random vectors in
Rd .
m (meRd)
and Q are the mean an covariance matrix common to allXk .
n
E S
Sn-
nmLet
S n
=E Xk . Then,
when n - oo, the lawof Sn - nm
converges tok=l
in
~
GQ .
A
proof
of thisproposition
can befound,
forexample
in [43.3.SHORT SURVEY OF OTHER AVAILABLE CONSTRUCTIONS. It is well-known that, in the literature one finds many other constructions of the
gaussian probability
inRd . By
"construction" we mean both adefinition and the basic
properties (in
a certainorder).
Thereare three main
types
of constructions :The
density approach
at least dates back to P.Levy [53.
Inthis construction one starts
defining
in~d
thedensity
ofwhich is
obvious,
or thedensity
ofG ~Q~ ,
whenQ
is nondegenerate
and
positive (in
this case the constant in front of theexponential
has to beidentified).
Excellentexemples
of thisapproach
aregiven
in W. Feller163,
A.Renyi [73,
L. Breiman[43,
and
(in french)
M. M6tivier L 8 ) .(1) is a
distribution,
famous instatistics,
where it isextensively
used. As often this distribution can be defined in many ways theinteger
n counts what is called thedegrees
offreedom. x ( n ~ has a
density
onR+, explicitely :
·Or course the fact that such a function is a c.f is
part
of theproblem.
Two excellent illustrations of this
approach
aregiven
in H.Cramer [9] and A.N.
Shiryayev [103.
The definition
by linearity :
a r.v X =(Xk)lCk4d
withvalues
in
~d
is said to begaussian,
if any linear form hasa
gaussian
distribution in ~.We shall not comment here this
apparently
non-constructive definition.Obviously
such a definition is veryappropriate
togaussian
processes(it
also worksperfectly
well for randomvectors).
Thisapproach
is illustrated in four excellent books(two of them in
french) :
C.R. Rao
1113,
J. Neveu[123,
A. Monfort [133 and P. Br6maud [143.4. CONVERGENCE TO
x2 .
In this section we shall use both thereproducing
innerproduct (.,.)
on ImQ and the central limit theorem(proposition 4),
in order togive
a newproof
of thewell-known result
stating
that under some conditions the distribution of(the
so-called"empirical X2")
converges tox~(d-1).
The whole result will be denoted :(Of
course we shall define everyingredient
of this formula in propertime~ .
Beforehand a few
preliminaires :
Let I be a finite set (d = card
I)
and p = aprobability
on I, such asPi - 0,
for every i E I. Letlei,
i E I)be the natural basis of
RI,
and X a r.v with values inRI,
the lawof which is defined
by
PCX =
si3
=pi,
i E I .(4.2)
Of course,
EX =
(Pi;
I e I) =p e RI ; (4.3)
whereas the
covariance Q
of X satisfiesOij = Pi 8ij - Pi Pj , ,
, 1,(4.4)
Qij = 1J pipj,
I(4.4)
where 8 is Knonecker.
We
point
out two facts :As
usual ,
X 6 EX + ImQ(a. s~ ; (4.5)
Q is the covariance matrix of the n-multinomial law in thespecial
cas n = 1.
Now we shall
proceed
andidentify
thereproducing
innerproduct
associated to the covariance Q of the above r.v X. Q can also be
seen as the matrix with
typical
elementgiven by (4.4).
Proposition
5.1 (i)
Thesubspace
ImQ isexactly
constitutedby
the ysatisfying
1(it)
The innerproduct
Proof. For x e
RI
and with(4.4),we
see that(Qx)i=pixi-piX pjxj.
So it is clear that = 0 . On the other way
let 3
be ani
y
element
satisfying E yi
= 0 . Such a ybelongs
to ImQ, since theequations yi
=(Qx)i (we
want to solve inx)
admit a solutionYi
x =
(xi)
~ withxi -
~Yi (4.7)
Pi
Hence ImQ,
exactly
identified, has d-1 for rank.The
proof
of(ii)
isstraightforward :
Let y = Qx and
y’
=Qxl
be in ImQ,(2.2)
and(4.6) give successively :
(y,y’)Q
=(x,Qxl) ,
andNow we are in a
position
to ive aprecise
version of theproposition
of the convergence to x(d-1) -
Proposition
6. (Theorem of convergence tox2~.
Let be an i . i . d sequence of random vectors in
d) ,
the common distribution is the law of X defined inThen, when n - +o
d
the distribution of
Ed npi converges to x2(d-1). (4.8)
I-i
npi
Proof. The fulfills the
hypothesis
ofproposition
4.n
’
Therefore, denoting Nn
=2: xk , as
n - +m, the distribution of k=1Nn-np
Nn-np
converges toGQ ,
where Q is the covariance matrix (4.4).4n
the of
proposition 3)
that,the distribution of converges to
(1) .
n n Q
, ,
By using part (ii)
ofproposition
5, we canexplicit
the square of the norm above and obtain :this
gives
the desired result. a5.SHORT SURVEY OF OTHER AVAILABLE PROOFS. In the literature one can find different
proofs
of thisx2
theorem.They
areessentially
of two
types :
The characteristic fonction
approach
starts withtaking
the r.vNn =
which has been defined inproposition
6. Thedistribution of
Nn
ismultinomial,
hence his c.f is well-known.One deduces from this the c.f of the r.v (Nn (i) -npi) Vnpi As One deduces from this the c.f of the r.v
14i4d
. As
- Vnpi
n + +oo , one obtains
then, essentially
withcalculus,
the limit of this c.f which appears to be the c.f of asingular gaussian
distribution.. The end of the
proof
is close. In their book H.Cramer
[93
and M. Fisz[153 exactly
follow this line ofproof.
The central limit theorem
approach
is a short way to find thegaussian limit
of the r.v . Then thequestion
is to find thegaussian
limit of the r.v 201320132013 . Then thequestion
is to find theVn
image
of this distributionby
the continuousapplication :
This can be done
differently according
to thestyle
of the author.This line of
presentation is, nowadays,
often chosen ; see, forexample C153
and [16](this
reference is infrench).
(1)
We haveimplicitely
used thecontinuity
of the functionx ~
Ilxll2.
References
1. N.
Aronszajn, "Theory
ofreproducing kernels",
Trans. Am.Math. Soc., 68 337-404
(1950).
2. E. Parzen, "Time series
analysis papers",
477 and ff, Holdenday (1967).
3. P.
Halmos,
"Finite-dimensional vectorspaces",
Van Nostrand(1958).
4. L.
Breiman, "Probability",
AddisonWesley (1968).
5. P.
Levy,
"Processusstochastiques
et mouvementbrownien", Gauthier-Villars,
Paris(1948).
6. W.
Feller,
"An introduction toProbability Theory
and itsApplications",
Vol 2Wiley (1966).
7. A.
Renyi,
"Foundations ofprobability", Holden-day (1970).
8. M.
Métivier,
"Notions fondamentales de la théorie desprobabilités", Dunod,
Paris(1972).
9. H. Cramer, "Mathematical Methods of
Statistics",
PrincetonUniversity
Press(1961),
Ninthprinting.
10. A.N.
Shiryayev, "Probability", Springer-Verlag
N.Y.(1984).
11. C.R. Rao, "Linear Statistical Inference and its
Applications", Wiley (1965).
12. J. Neveu, "Processus Aléatoires
Gaussiens",
Presses de l’Université de Montréal(1968).
13. A.
Monfort,
"Cours deProbabilités", Economica,
Paris(1980).
14. P.
Brémaud,
"Introduction toprobabilistic modeling", Springer-Verlag
New-York(1988).
15. M.
Fisz, "Probability theory
and MathematicalStatistics", Wiley (1963).
16. A.
Monfort,
"Cours destatistique mathématique", Economica,
Paris(1982).
Jean-Louis PHILOCHE PG 75000 - PARIS ManuscrM requ le 18 juin 1992