HAL Id: hal-00894091
https://hal.archives-ouvertes.fr/hal-00894091
Submitted on 1 Jan 1995
HAL is a multi-disciplinary open access
archive for the deposit and dissemination of
sci-entific research documents, whether they are
pub-lished or not. The documents may come from
teaching and research institutions in France or
abroad, or from public or private research centers.
L’archive ouverte pluridisciplinaire HAL, est
destinée au dépôt et à la diffusion de documents
scientifiques de niveau recherche, publiés ou non,
émanant des établissements d’enseignement et de
recherche français ou étrangers, des laboratoires
publics ou privés.
Heterogeneous variances in Gaussian linear mixed
models
Jl Foulley, Rl Quaas
To cite this version:
Original
article
Heterogeneous
variances
in
Gaussian
linear mixed models
JL
Foulley
RL
Quaas
1
Institut
national de la rechercheagronomique,
station degénétique
quantitativeet
appliquée,
78352Jouy-en-Josas,
Fra!ece;2
Department
of
AnimalScience,
CornellUniversity, Ithaca,
NY14853,
USA(Received
28February 1994; accepted
29 November1994)
Summary - This paper reviews some
problems
encountered inestimating heterogeneous
variances in Gaussian linear mixed models. The one-way and
multiple
classificationcases are considered. EM-REML
algorithms
andBayesian procedures
are derived. Astructural mixed linear model on
log-variance
components is alsopresented,
which allows identification ofmeaningful
sources of variation ofheterogeneous
residual andgenetic
components of variance and assessment of their
magnitude
and mode of action.heteroskedasticity
/
mixed linearmodel / restricted
maximum likelihood/
Bayesian statisticsRésumé - Variances hétérogènes en modèle linéaire mixte gaussien. Cet article
fait
le point sur un certain nombre de
problèmes
qui surviennent lors de l’estimation de varianceshétérogènes
dans des modèles linéaires mixtesgaussiens.
On considère le casd’un ou
plusieurs facteurs
d’hétéroscédasticité. Ondéveloppe
desalgorithmes
EM-REMLet
bayésiens.
On proposeégalement
un modèle linéaire mixte structurel deslogarithmes
des variances qui permet de mettre en évidence des sources
significatives
de variation des variances résiduelles etgénétiques
etd’appréhender
leur importance et leur mode d’action.hétéroscédasticité
/
modèle linéaire mixte/
maximum de vraisemblance résiduelle/
statistique bayésienneINTRODUCTION
Genetic evaluation
procedures
in animalbreeding rely mainly
on best linearunbi-ased
prediction
(BLUP)
and restricted maximum likelihood(REML)
estimation ofparameters
of Gaussian linear mixed models(Henderson,
1984).
Although
BLUPmixed-model
methodology postulate
homogeneity
of variancecomponents
acrosssubclasses involved in the stratification of data.
However,
there is now agreat
dealof
experimental
evidence ofheterogeneity
of variances forimportant production
traits of livestock
(eg,
milkyield
andgrowth
incattle)
both at thegenetic
and envi-ronmental levels(see,
forexample,
the reviews of Garrick etal, 1989,
and Visscher etal,
1991).
As shown
by
Hill(1984),
ignoring heterogeneity
of variance decreases theef-ficiency
ofgenetic
evaluationprocedures
andconsequently
response toselection,
the
importance
of thisphenomenon
depending
onassumptions
made about sourcesand
magnitude
ofheteroskedasticity
(Garrick
and VanVleck, 1987;
Visscher andHill,
1992).
Thus,
making
correct inferences about heteroskedastic variances is crit-ical. To thatend, appropriate
estimation andtesting
procedures
forheterogeneous
variances are needed. The purpose of this paper is anattempt
to describe such pro-cedures and theirprinciples.
Forpedagogical
reasons, thepresentation
is dividedinto 2
parts
according
to whetherheteroskedasticity
is related to asingle
or to amultiple
classification of factors.THE ONE-WAY CLASSIFICATION
Statistical model
The
population
is assumed to be stratified into severalsubpopulations
(eg,
herds,
regions,
etc)
indexedby
i =1, 2, ... , I, representing
apotential
source ofhetero-geneity
of variances. For the sake ofsimplicity,
we first consider a one-way randommodel for variances such as
where yi is the
(n
2
x 1)
data vector forsubpopulation
i,
13
is the(p x 1)
vector offixed effects with incidence matrix
X
i
,
is *u a(q
x1)
vector of standardized random effects with incidence matrixZ
i
and ei is the(n
i
x 1)
vector of residuals.The usual
assumptions
ofnormality
andindependence
are made for thedistri-butions of the random variances u* and ei, ie u* !
N(0, A) (A
positive
definitematrix of coefficients of
relationship)
and eiNNID(O,
er! 1!;)
andCov(e
i
,
u*!)
= 0 so that y2NN(X
il
3,
a u 2i Z’AZI i +0,2 ei 1,
whereor
2
ei
and
o, ui 2
are the residual andu-components
of variancepertaining
tosubpopulation
i. Asimple example
for[1]
is a2-way
additive mixel model Yij =p,+hi+as!8!
+ e
zjk
with fixed herd(hi)
and random sire(<7..;,!)
effects. Notice that model[1]
includes the case of fixed effectsnested within
subpopulations
as observed in manyapplications.
EM REML estimation of
heterogeneous
variancecomponents
To be consistent with common
practice
for estimation of variancecomponents,
we chose REML(Patterson
andThompson,
1971;
Harville,
1977)
as the basicestimation
procedure
forheterogeneous
variancecomponents
(Foulley
etal,
1990).
A convenientalgorithm
tocompute
such REML estimates is thebe based on the
general
definition of EM(see
pages 5 and 6 and formula 2.17 inDempster
etal,
1977)
which can beexplained
as follows. L tt’(
1 1 1 1)1
2{ 2}
l 2{ 2}
d 2( 2’
Lettln
g
y = y 1 , Y 2, ... , Y i, ... , Y I ,
(y
e
2
=
1
0
,
2
ei 1, U2
,
Ui
2
=
o
f
u
and
U2
=
(
g2
&dquo; e
9
u 2’),
the derivation of the EMalgorithm
for REML stems from acomplete
data set definedby
the vector x= (y’,
*/
I
)
,
U
1
13
and thecorresponding
likelihood functionL(
6
2
;x)
=Inp(xlc¡
2
).
In thispresentation,
the vector(3
is treated in aBayesian
manner as a vector of random effects with variance fixed at
infinity
(Dempster
etal,
1977; Foulley,
1993).
Afrequentist
interpretation
of thisalgorithm
based on errorcontrasts can be found in De Stefano
(1994).
A similar derivation wasgiven
for thehomoskedastic case
by
Cantet(1990).
Asusual,
the EMalgorithm
is an iterative oneconsisting
of an’expectation’
(E)
and of a ’maximization’(M)
step.
Given thecurrent estimate
c¡
2
=c¡
2[t]
at iteration[t],
the Estep
consists ofcomputing
the conditionalexpectation
ofL( c¡
2
; x),
iegiven
the data vector y and()&dquo;2
= ()&dquo;2[t].
The M
step
consists ofchoosing
the next value()&dquo;2[
t+l]
ofU
2
by
maximizing
Q()&dquo;
2
1()&dquo;
2[t]
)
withrespect
toU
2
Since
In p(xl(T2)
= ln p (y ! (3, u* ,
(T2)+ln p(l3, u* 1(T2) with In p(l3, u
*
I (
T
2
)
providing
no information about
o-
2
,
Q( (
T
2
1
(T2[
t]
)
can bereplaced by
Under model
!1!,
theexpression
forQ
*
(c
r
21U
2
[l]
)
reduces towhere
E!t!(.)
indicates a conditionalexpectation
taken withrespect
to thedistribu-tion of
[3,
U
*
I y,
6
2
=(J’
2[t]
.
Thisposterior
distribution is multivariate normal withmean
E(l3ly,
6
2
)
= BLUE(best
linear unbiasedestimate)
ofj3,
E(u!y,
(
7
’)
= BLUPof u, and
Var(l3, uly,
(J’2)
= inverse of the mixed-model coefficient matrix.The
system
ofequations
åQ
*
( (J’21
o’!)/9o’!
= 0 can be written as follows: Withrespect
to theu-component,
we haveFor the residual
component,
Since
E!t] (e!ei)
is a function of the unknown Quionly,
equation
[5]
depends
only
on that unknown whereas
equation
[6]
depends
of both variancecomponents.
Wethen solve
[5]
first withrespect
to Ju, , and then solve[6]
secondsubstituting
the solutiona!t+1!
to o,,,, back intoE!t](e!ei)
of(6!,
ie withHence
It is worth
noticing
that formula[7]
gives
theexpression
of the standard deviation of theu-component,
and has the form of aregression
coefficient estimator.Actually
Ju
, is the coefficient of
regression
of any element of yi on thecorresponding
elementof Zju
*
.
Let the
system
of mixed-modelequations
be written asand
C
= [
C
,
3,3
C
,
3
.
J =
g inverse of the coefficient matrix.L
C.,3
CUU
I
=
-The elements of
[7]
and[8]
can beexpressed
as functions of y,(3,
u and blocks ofC as follows
For readers interested in
applying
the aboveformulae,
a smallexample
is theis worth
noticing
that formulae[7]
and[8]
can also beapplied
to the homoskedasticcase
by considering
that there isjust
onesubpopulation
(I = 1).
Theresulting
algorithm
looks like aregression
in contrast to the conventional EM whose formula(a![t+1]
=E
l
t]
(u’A-
1
u)/q)
where u is not standardized(u
=cr!u*)
is in terms of a variance. Notonly
do the formulae lookquite different,
butthey
alsoperform
quite
differently
in terms of rounds to convergence. The conventional EM tends todo
quite
poorly
ifor »
o, and
(or)
with littleinformation,
whereas the scaled EM is at its best in these situations. This can be demonstratedby examining
abalanced
paternal
half-sibdesign
(q
families with progeny group size neach).
This is convenient because in this case the EMalgorithms
can be written in terms ofthe between- and within-sire sums of squares and convergence
performance
checkedfor a
variety
of situations withoutsimulating
individual records. For thissimple
situation
performance
wasfairly
wellpredicted
by
the criterionR
2
=n/(n
+a),
where a =a2/0,2 .
Figure
1 is aplot
of rounds toconvergence for the scaled and usual EM
algorithms
for anarbitrary
set of values of n and a. As notedby
Thompson
and
Meyer
(1986),
the usual EMperforms
verypoorly
at lowR
2
,
eg, n = 5 andh
2
=4/(a
+1)
= 0.25 or n = 33 andh
2
=0.04,
ieR
2
=0.25,
butvery well at the other end of the
spectrum: n
= 285 andh
2
= 0.25 or n = 1881 andh
2
=0.04,
ieR
2
= 0.95. Theperformance
of the scaled version is the exactopposite.
Interestingly,
both EMalgorithms
perform
similarly
forR
2
values
typical
of many animalbreeding
data sets(n
= 30and h
2
=0.25,
ieR
2
=2/3).
Moreover,
solutionsgiven
by
the EMalgorithm
in[7]
and[8]
turn out to be within theparameter
space in the homoskedastic case(see
proof
in theAppendix)
Bayesian approach
When there is little information per
subpopulation
(eg,
herd or herdxmanagement
unit),
REML estimation ofQ
e
i
andQ
u
z
can be unreliable. This led Hill(1984)
and Gianola
(1986)
tosuggest
estimates
shrunken towards some common meannatural extension of the EM-REML
technique
describedpreviously.
Theparameters
ol2 ei
ando,
U
,
2
are assumed to beindependently
andidentically
distributed random variables with scaled invertedchi-square density functions,
theparameters
of whichare
s2,,
q
,
e
andsu,
r!!respectively.
Theparameters
se and
s! are
locationparameters
of the
prior
distributions of variancecomponents,
and TIeand 7
7
,,
(degrees
ofbelief)
are
quantities
related to thesquared
coefficient of variation(cv)
of true variancesby
qe =(2/cve )
+ 4and
q
u
=(2/cufl)
+ 4respectively:
Moreover,
let us assume as in Searle et al(1992,
page99),
that thepriors
for residual andu-components
are assumedindependent
so thatp (,72i, U2i) = p(,71i)p(0,2i).
The
2
Q
t
[
])
1 O’
2
( 0’
@
function to maximize in order toproduce
theposterior
modeof
o-
2
is now(Dempster
etal,
1977,
page6):
with
Equations
based on first derivatives set to zero are:Using
!l2ab!,
one can use thefollowing
iterativealgorithm
Comparing
[13b]
and[14]
with the EM-REML formulae[7]
and[8]
shows howprior
information modifies data information(see
also tables I andII).
Inparticular
whenTJe
(TJ
u
)
= 0(absence
ofknowledge
onprior
variances),
formulae[13b]
and[14]
are very similar to the EM-REML formulae.
They
would have beenexactly
thesame if we had considered the
posterior
mode oflog-variances
instead ofvariances,
!7e and !7.!
replacing
17, + 2 and !7! + 2respectively
in!11!,
and,
consequently
also in the denominator of[13b]
and!14!.
Incontrast,
if!7e(!/u)
---> 00(no
variation amongvariances),
estimates tend to the locationparameters
s!(s!).
Extension to several
u-components
The EM-REML
equations
caneasily
be extended to the case of a linear mixedmodel
including
severalindependent
u-components
(uj; j
=1, 2, ... ,
J),
ieIn that case, it can be shown that formula
[7]
isreplaced by
the linearsystem
The formula in
[8]
for the residualcomponents
of variance remains the same.This
algorithm
can be extended tonon-independent
u-factors. As in asire,
maternal
grand
siremodel,
it is assumed that correlatedfactors j and
k are suchthat
Var(u*)
=Var(u*)
=A,
andCov(uj, u!/)
=p
jk
A
withdim(u!)
= m for allj.
Leta
2
=(or2&dquo; or2&dquo;
p’)
with p =vech(S2),
S2being
a(m
xm)
correlation matrix with pjk as elementjk.
TheQ#(êT2IêT
2[t]
)
function to maximize can be written here aswhere
The first term
Q7 (u! ] 8&dquo;!°! )
=ErJ[lnp(yll3,u*,(J’!)]
has the same form as withthe case of
independence
except
that theexpectation
must taken withrespect
to the distribution of(3,
u
*
Iy,
Õ
’
2
=Õ’
2[t]
.
The second termQ!(plÕ’2[t])
=E
l
c
’
]
) [In p(u
*
1
&2
)]
can beexpressed
aswhere D =
{uj’ A
-l
uk}
is a(J
xJ)
symmetric
matrix.The maximization of
Q#(¡
2
1(¡
2[t]
)
withrespect
to6
2
can be carried out in 2stages:
i)
maximization ofQr(¡
2
1(¡2
[t]
)
withrespect
to the vector!2 of
variancecomponents
which can be solved asabove;
andii)
maximization ofQ#(p 1&211,)
withrespect
to the vector of correlation coefficients p which can beperformed
viaTHE MULTIPLE-WAY CLASSIFICATION
The structural model on
log-variances
Let us assume as above that the
a2s
(u
and etypes)
are apriori
independently
distributed as inverted
chi-square
random variables withparameters
5!
(location)
and riz
(degrees
ofbelief),
such that thedensity
function can be written as:where
r(
x
)
is the gamma function.From
!19),
one canalternatively
consider thedensity
of thelog-variance
1n
Q2
,
I or moreinterestingly
thatof v
z
=ln(a2/s2).
Inaddition,
it can be assumed that 71i = ! for all
i,
and thatlns2
can bedecomposed
as a linear combinationp’S
ofsome vector 5 or
explanatory
variables(p’
being
a row vector ofincidence),
such thatwith
For v
i
-->0,
the kernel of the distribution in[21]
tends towardsexp( -r¡v’f /4),
thusleading
to thefollowing
normalapproximation
where the variance a
priori
(!)
of true variances isinversely proportional
to
q
(!
= 2/?!), !
alsobeing
interpretable
as the square coefficient of variation oflog-variances.
Thisapproximation
turns out to be excellent for most situations encountered inpractice
(cv ! 0.50).
Formulae
[20]
and[21]
can benaturally
extended to severalindependent
classi-fications in v =(v!, v2, ... ,
vj, ... ,
v!)’
such thatwith
where
K
j
=dim(v
j
)
and J1 =(t,’, v’)’
is the vector ofdispersion
parameters
andC’
=(p!, q
’)
is thecorresponding
row vector of incidence.This
presentation
allows us to mimick a mixed linear model structure with fixed 5 and random v effects onlog-variances,
similar to what is done on cell means(vt
i
=x!13
+z’u
=t!0),
and thusjustifies
the choice of thelog
as the link function(Leonard,
1975; Denis,
1983;
Aitkin, 1987;
Nair andPregibon,
1988)
to use for thisgeneralized
linear mixed-modelapproach.
where y! =
in 0,2i 1,
ye =in or’ 1;
P
u
,
P
e
are incidence matricespertaining
to fixed effects
5
u
,
be
respectively; Q
u
,
Qg
are incidence matricespertaining
to random effects vu =(V!&dquo;V!2&dquo;&dquo;,V!j&dquo;&dquo;)’
and Ve =(V!&dquo;V!2&dquo;&dquo;,V!jl)’
withv! -Nid(0,!I!.)
J J UJand V
el
J-NID(0,!,I! ,)
J ej’respectively.
Estimation
Let A =
(À!, A
’)’
and
(ç!,
g[I’ where
g
u
=
{ç
uJ
and
Ç
e
=!ei, 1.
Inferenceu e
e
>about 71 is of an
empirical Bayes type
and is based on the mode a of theposterior
density
p(Àly, E,
=i;) given I = I
itsmarginal
maximum likelihoodestimator,
ieMaximization in
[26ab]
can be carried outaccording
to theprocedure
describedby
Foulley
et al(1992)
and San Cristobal et al(1993).
Thealgorithm
forcomputing
X
can be written as(from
iteration t to t +1)
where
z =
(z’
u
,
z!)
areworking
variablesupdated
at each iteration and such thatw =
W
-- Wue J is
a (2I, 21)
matrix ofweights
described inFoulley
et alW
euWee
(1990, 1992)
for the environmental variancepart,
and in San Cristobal et al(1993)
for thegeneral
case.where
9tl ,
9t/
are
solutions of[26]
for ! =
![nl
andH[n]
H!n!e!&dquo;
are blocks ofwere uj ’
ej’
are so utlonS 0 lor L, =L, , an VjVj’ ej,ej&dquo; are oc SO
the inverse of the coefficient matrix of
[27]
pertaining
tov
U
j’
Vej’
respectively.
Testing procedure
The
procedure
describedpreviously
reduces to REML estimation of J1 when flatpriors
are assumed for v! and v,, orequivalently
when the structural model forlog-variances
is apurely
fixed model. Thisproperty
allows derivation oftesting
procedures
toidentify significant
sources ofheteroskedasticity.
This can be achievedvia the likelihood ratio test
(LRT)
asproposed
by
Foulley
et al(1990, 1992)
and SanCristobal et al
(1993)
but inusing
the residual(marginal)
log-likelihood
functionL(7!;
y)
instead of the fulllog-likelihood
L(71,
[3;
y).
Let
o
:À
H
c!lo
be the nullhypothesis
and:À
H
1
E Il -A
o
itsalternative,
whereA refers to the
complete
parameter
space and!lo,
a subset ofit, pertains
toH
o
.
Under
H
o
,
the statistichas an
asymptotic chi-square
distribution with rdegrees
offreedom;
r is thedifference in the number of estimable
parameters
involved in A andIl
o
,
eg,r =
rank(K)
whenH
o
is definedby
K’71 =0,
K’being
a full row rank matrix.Fortunately,
for a Gaussian linearmodel,
such as the one considered herey 2
N
N(Xil3, a!iZ!AZi
+U2i
j&dquo;
i
),
L(À; y),
can beeasily expressed
as1
0
0
where N =
! ni,
p = rank(X),
Ti
=(
x
i,or.
i
zi),
E- -[0 0]
and where N= L
ni,
P =
rank(X),
Ti
=(Xi,
aUiZi),
!- =10
AJ
and<=1
! !,
e =
(13, û
*
’)
is a solution of thesystem
of mixed-modelequations
in[9],
ieI
1 I[t a;:2T!Ti + !-] ê = t a;:2T!Yi’
L
i=1
&dquo; i_
i=l
&dquo; iyiCONCLUSION
This paper is an
attempt
tosynthesize
the current state of research in the field of statisticalanalysis
ofheterogeneous
variancesarising
in mixed-modelmethodology
and in itsapplication
to animalbreeding.
Forpedagogical
reasons, the papersuccessively
addressed the cases of one-way andmultiple-way
classification. Inany case, the estimation
procedures
chosen were REML and its naturalBayesian
extension,
theposterior
modeusing
inverted gammaprior
distributions. For asingle classification, simple
formulae forcomputing
REML andBayes
estimationsFor
multiple classifications,
thekey
ideaunderlying
the structuralapproach
is to render models asparsimonious
aspossible,
which isespecially
critical with thelarge
data sets used ingenetic
evaluationhaving
alarge
number of subclasses.Consequently,
it was shown howheterogeneity
oflog-residual
andu-components
of variance can be described with a mixed linear model structure. This makes thecorresponding
estimationprocedures
a natural extension of what has been done fordecades
by
BLUPtechniques
on subclass means.An
important
feature of thisprocedure
is itsability
to assess via likelihood ratiotests the effects of
potential
sources ofheterogeneity
considered eithermarginally
or
jointly. Although computationally demanding,
theseprocedures
havebegun
to beapplied.
San Cristobal et al(1993)
applied
theseprocedures
to theanalysis
of muscledevelopment
scores atweaning
of 8 575 progeny of 142 sires in the MaineAnjou
cattle breed whereheroskedasticity
was found both for the sire and residualcomponents.
Weigel
et al(1993)
and DeStefano(1994)
also used the structuralapproach
for sire and residual variances to assess sources ofheterogeneity
in within-herd variances of milk and fat records in Holstein. Herd size and within-within-herd means were associated withsignificant
increases in residual variances as well as variousmanagement
factors(eg,
milking
system).
Approximations
for the to estimation of withinregion-herd-year-parity
phenotypic
variances were alsoproposed
fordairy
cattle evaluationby Wiggans
and Van Raden(1991)
andWeigel
and Gianola(1993).
Thesetechniques
also open newprospects
of research in different fields ofquantitative genetics,
eg,analyses
ofgenotype
x environment interactions(Foulley
etal,
1994;
Robert etal,
1994),
testing procedures
forgenetic
parameters
(Robert
etal,
1995ab),
crossbreeding
experiments
andQTL
detection.However,
research isstill needed to
improve
themethodology
andefficiency
ofalgorithms.
ACKNOWLEDGMENTS
This paper draws from material
presented
at the 5th WorldCongress
of GeneticsApplied
to LivestockProduction, Guelph, August 7-12,
1994.Special
thanks areexpressed
to ALDeStefano,
VDucrocq,
RLFernando,
DGianola,
DH6bert,
CRHenderson,
SIm,
CRobert,
MGaudy-San
Cristobal and KWeigel
for their valuable contribution to the discussion and to thepublication
of papers on thissubject.
The assistance of C Robert inthe
computations
of the numericalexample
is alsogreatly acknowledged.
REFERENCES
Aitkin M
(1987)
Modelling
varianceheterogeneity
in normalregression using
GLIM.Appl
Stat36,
332-339Cantet RJC
(1990)
Estimation andprediction problems
in mixed linear models for maternalgenetic
effects. PhDthesis, University
ofIllinois, Urbana, IL,
USADempster AP,
LairdNM,
Rubin DB(1977)
Maximum likelihood fromincomplete
datavia the EM
algorithm.
J R Stat Soc B 39, 1-38Denis JB
(1983)
Extension du mod6le additifd’analyse
de variance par mod6lisationmultiplicative
des variances. Biometrics39,
849-856DeStefano AL
(1994)
Identifying
andquantifying
sources ofheterogeneous
residual andFoulley
JL(1993)
Asimple
argumentshowing
how to derive restricted maximum likeli-hood. JDairy
Sci76,
2320-2324Foulley JL,
GianolaD,
San CristobalM,
Im S(1990)
A method forassessing
extent andsources of
heterogeneity
of residual variances in mixed linear models. JDairy
Sci73,
1612-1624Foulley JL,
San CristobalM,
GianolaD,
Im S(1992)
Marginal
likelihood andBayesian
approaches
to theanalysis
ofheterogeneous
residual variances in mixed linear Gaussian models.Comput
Stat Data Anal 13, 291-305Foulley JL,
H6bertD, Quaas
RL(1994)
Inferences onhomogeneity
ofbetween-family
components of variance and covariance among environments in balanced cross-classified
designs.
Genet SelEvol 26,
117-136Garrick
DJ,
Van Vleck LD(1987)
Aspects
of selection forperformance
in severalenviron-ments with
heterogeneous
variances. J Anim Sci65,
409-421Garrick
DJ,
PollakEJ, Quaas RL,
Van Vleck LD(1989)
Varianceheterogeneity
in direct and maternalweight
traitsby
sex and percentpurebred
for Simmental sired calves.J Anim Sci
67,
2515-2528Gianola D
(1986)
On selection criteria and estimation of parameters when the variance isheterogeneous.
TheorAppl
Genet72,
671-677Gianola
D, Foulley JL,
FernandoRL,
HendersonCR, Weigel
KA(1992)
Estimation ofheterogeneous
variancesusing empirical
Bayes
methods: theoretical considerations. JDairy
Sci 75, 2805-2823Graybill
FA(1969)
Introduction to Matrices withApplications
to Statistics. WadsworthPublishing Co, Bemont, CA,
USAHarville DA
(1977)
Maximum likelihoodapproaches
to variance component estimationand related
problems.
J Am Stat Assoc 72, 320-340Henderson CR
(1984)
Applications of
Linear Models in AnimalBreeding. University
ofGuelph, Guelph, Ontario,
Canada.Hill WG
(1984)
On selection among groups withheterogeneous
variance. Anim Prod 39, 473-477Johnson
DL, Thompson
R(1994)
Restricted maximum likelihood estimation of variancecomponent for univariate animal models
using
sparse matrixtechniques
and aquasi-Newton
procedure.
In: 5th WorldCongress
on GeneticsApplied
to Livestock Production(C
Smith etal,
eds),
University
ofGuelph, Guelph, ON, Canada,
vol18,
410-413 Leonard T(1975)
ABayesian approach
to the linear model withunequal
variances.Technometrics
17,
95-102Liu
C,
Rubin DB(1994)
Application
of the ECMEalgorithm
and Gibbssampler
togeneral
linear mixed model. In: Proc XVllth International Biometric
Conference,
McMasterUniversity, Hamilton, ON, Canada,
vol 1, 97-107Meyer
K(1989)
Restricted maximum likelihood to estimate variance components for animal models with several random effectsusing
a derivative-freealgorithm.
GenetSel
Evol 21,
317-340Nair VN,
Pregibon
D(1988)
Analysing dispersion
effects fromreplicated
factorialexper-iments. Technometrics
30,
247-257Patterson
HD, Thompson
R(1971)
Recovery
of interblock information when block sizesare
unequal.
BioTn,etrika 58, 545-554Robert
C, Foulley JL, Ducrocq
V(1994)
Estimation andtesting
of a constantgenetic
correlation among environments. In: 5th WorldCongress
on GeneticsApplied
toLivestock Production
(C
Smith etal,
eds),
University
ofGuelph, Guelph, ON, Canada,
vol
18,
402-405Robert
C, Foulley JL, Ducrocq
V(1995a)
Genetic variation of traits measured inRobert
C, Foulley
JL,Ducrocq
V(1995b)
Genetic variation of traits measured inseveral environments. II. Inference on between-environments
homogeneity
of intra-class correlations. Genet SelEvol 27,
125-134B
San Cristobal M,
Foulley JL,
Manfredi E(1993)
Inference aboutmultiplicative
het-eroskedastic components of variance in a mixed linear Gaussian model with anap-plication
to beef cattlebreeding.
Genet Sel Evol25,
3-30Searle
SR,
CasellaG,
McCulloch CE(1992)
VarianceComponents.
JWiley
andSons,
New
York,
USAThompson R, Meyer
K(1986)
Estimation of variance components: what ismissing
in theEM
algorithm?
J StatComput
Simul24,
215-230Visscher PM, Hill WG
(1992)
Heterogeneity
of variance anddairy
cattlebreeding.
Anim Prod 55, 321-329Visscher
PM, Thompson R,
Hill WG(1991)
Estimation ofgenetic
and environmental variances for fatyield
in individual herds and aninvestigation
intoheterogeneity
ofvariance between herds. Livest Prod Sci 28, 273-290
Weigel KA,
Gianola D(1993)
Acomputationally simple Bayesian
method for estimationof
heterogeneous
within-herdphenotypic
variances. JDairy
Sci76,
1455-1465Weigel KA,
GianolaD,
YandelBS,
Keown JF(1993)
Identification of factorscausing
heterogeneous
within-herd variance componentsusing
a structural model for variances.J
Dairy
Sci76,
1466-1478Wiggans GR,
VanRaden PM(1991)
Method and effect ofadjustment
forheterogeneous
variance. JDairy
Sci74,
4350-4357APPENDIX
A
proof
thatequation
[7]
isnon-negative
in the homoskedastic case is as follows.The numerator of
[7]
can be written as:where e =
y -
XI3,
Ê= y -X(3
=VPy,
and u =-
(y -
1
a
y
Z’
u
X!)
=u
,,ZPy
with V =
Var(y),
P =y-1- Y-1X(X’y-1X)-X’Y-1,
andK’y,
anN-rank(X)
vector of linear contrasts. Now
so that
Moreover
V-’ -
P =V-’X(X’V-’X)-X’V-
1
isthen
nnd
matrices,
then
tr(AB) >
0(Graybill, 1969).
COMMENT
Robin
Thompson
Roslin
Institute, Roslin,
Midlothian EH259PS,
UKThis paper is a
synthesis
of recent work of the authors and their co-workers in the area ofheterogeneous
variances. I think it is a valuable reviewgiving
alogical
presentation
showing
howheterogeneous
variancemodelling
can be carried out.The
parallels
between estimation of linearparameters
and varianceparameters
ishighlighted. My
comments relate totransformation,
convergence,simplicity
andutility.
Transformation
Emphasis
is on thegenetic
standard deviation as aparameter
rather than thegenetic
variance. Such aparameterization
has been used to allow estimation ofbinary
variancecomponents
(Anderson
andAitkin,
1985).
Using
scaled variablescan also allow reduction of the dimension of search in derivative-free methods of multivariate estimation
(Thompson
etal,
1994).
It is also aspecial
case of the Choleskitransformation,
which has beensuggested
tospeed
up estimation andkeep
parameters
within bounds forrepeated
measures data(Lindstrom
andBates,
1988)
and multivariategenetic
data(Groeneveld, 1994).
Reverter et al
(1994)
haverecently
suggested
regression
type
methods for estimation of varianceparameters.
As the authorspoint
out,
there is a naturalregression interpretation
to the similarequations
[7]
and[8].
However[7]
and[8]
include trace terms thatessentially
correct for attenuation oruncertainty
inknowing
the fixed or random effects.Convergence
In the discussion of
Dempster
et al(1977)
Ipointed
out that the rate of convergence for a balanced one-wayanalysis
is(in
the authors’notation)
approximately
1-R
2
,
which,
Ithink, explains
one of thegraphs
infigure
1. In the timeavailable,
I havenot been able to derive the rate of convergence for the scheme based on
[7]
and[8],
but ifQ
e
is known and[7]
is used to estimate(T2i,
which should be agood
approximation,
then the rate of convergence is 1 -2R
2
(1 -
R
Z
).
This is ingood
agreement
withfigure
1,
suggesting
symmetry
aboutR
2
= 0.5 andequality
of thespeed
of convergencewhen R
2
=2/3.
As someone who has never understood the EM
algorithm
or itspopularity,
I would havethought
schemes based on some form of second differentials would bemore
useful,
especially
as some of the authors’ schemes allownegative
standarddeviations.
Simplicity
highlight
the need for aheterogeneous analysis,
both forfitting
the data andmeasuring
thepossible
loss of response.Utility
I can see the use of these methods at a
phenotypic level,
but I am less clear if it isrealistic to
attempt
to detect differences at agenetic
level(Visscher, 1992).
At thesimplest
level if the residuals areheterogeneous
can onerealistically
discriminatebetween models with
homogeneous
genetic
variances orhomogeneous
heritabilities ?Additional references
Anderson
DA,
Aitkin MA(1985)
Variance component models withbinary
response:interviewer
variability.
J R Stat Soc B47,
203-210Groeneveld E
(1994)
Areparameterization
toimprove
numericaloptimization
inmulti-variate REML
(co)variance
component estimation. Genet SelEvol 26,
537-545Lindstrom
MJ,
Bates DM(1988)
Newton-Raphson
and EMalgorithms
for linear mixed-effects models forrepeated
measures data. J Am Stat Assoc 83, 1014-1022Thompson R, Crump RE, Juga J,
Visscher PM(1995)
Estimating
variances andcovari-ances for bivariate animal models
using
scaling
and transformation. Genet SelEvol 27,
33-42
Reverter
A,
GoldenBL,
BourdonRM,
Brinks JS(1994)
Method R variance componentsprocedure: application
on thesimple breeding
value model. J Anim Sci72,
2247-2253Visscher PM
(1992)
Power of likelihood ratio tests forheterogeneity
of intraclasscorrela-tion and variance in balanced half-sib
designs.
JDairy
Sci75,
1320-1330COMMENT
Daniel Gianola
Department
of
Meat and AnimalScience, University
of
Wisconsin-Madison,
USA
This paper reviews and extends
previous
work doneby
the senior author and collaborators in the area of inferences aboutheterogeneous
variancecomponents
in animal
breeding.
The authorsimpose
a model on the variancestructure,
andsuggest
estimation andtesting procedures,
so as to arrive toexplanatory
constructs that have as fewparameters
aspossible.
Theirapproach
issystematic,
in contrast toother
suggestions
available in the literature(some
of which are referencedby Foulley
and
G!uaas),
where a model on the variances isadopted
implicity
without reference to the amount ofsupport
that isprovided by
the data athand,
and often on thebasis of mechanistic or ad hoc considerations. In this
front,
theirdevelopments
arewelcome.
They
adopt
either a likelihood or aBayesian viewpoint,
and restrict themselvesto
conducting
a search of theappropriate
maximizers(REML
orposterior modes,
respectively)
via the EMalgorithm,
which isimplemented
in aclear, straightforward
way. In so
doing,
they
arrive at a formulation(’scaled’
EM)
which exhibits adifferent numerical behavior from that of the ’standard’ EM in a
simple
model.at
estimating equations
that are similar to those of Gianola et al(1992),
citedby
the authors. It is now well known thatsetting
alldegree
of beliefparameters
tozero leads to an
improper posterior distribution;
surprisingly,
the authors(as
well as Gianola etal,
1992)
do not alert the readers about thispitfall.
The authors do not
provide
measures of the curvature of the likelihood or of thepertinent posterior
distribution in theneighborhood
of the maximum. It is usefulto recover second derivatives either to
implement
fully
the maximum likelihoodanalysis,
or toapproximate
thejoint
(or marginal)
posterior
with a Gaussian distribution. Gianola et al(1992)
gaveexpressions
for second differentials for somesimple
heterosckedasticmodels,
and an extension of these would have been aninteresting
contribution.Their
testing procedure
relies onasymptotic properties
of the(marginal)
like-lihood ratio test. I wonder how accurate this test is in situations where a
het-eroskedastic model of
high
dimensionality
may be warranted. In thissituation,
theasymptotic
distribution of the test criterion may differdrastically
from the exactsampling
distribution. It is somewhatsurprising
that the authors do not discussBayesian
test and associated methods forassessing uncertainty;
some readers maydevelop
the falseimpression
that there is a theoretical vacuum in this domain(see,
for
example, Robert,
1992).
There is a lot more information in a
posterior
(or
normalizedlikelihood)
thanthat contained in first and second differentials. In this
respect,
animplementation
based on Monte-Carlo Markov chain
(MCMC)
methods such as the Gibbssampler
(eg,
Tanner,
1993)
can be used to estimate the whole set ofposterior
distributions in the presence ofheterogeneous
variances. In somesimple
heteroskedastic linearmodels it can be shown that the random walk involves
simple
chains of normaland inverted
chi-square
distributions.Further,
it ispossible
to arrive at the exact(within
the limits of the Monte-Carloerror)
posterior
distributions of linear and nonlinear functions of fixed and random effects. In thesampling
theory
frameworkone encounters
immediately
a Behrens-Fisherproblem,
even in asimple
contrast between ’treatments’. TheBayesian
approach
via MCMC would allow an exactanalysis of,
forexample,
breedcomparison
experiments,
when the sources ofgermplasm
involved haveheterogeneous
and unknowndispersion.
I have been
intrigued
for some time about thepossible
consequences of hetero-geneous covariance matrices in animal evaluation in amultiple-trait
analysis.
Ifthere is
heterogeneous
variance there must beheterogeneity
in covariances as well !Perhaps
the consequences on animalranking
are even more subtle than in theuni-variate case. It is not obvious how the structural model
approach
can begeneralized
here,
and this is achallenge
for future research.Additional References