HAL Id: hal-00894156
https://hal.archives-ouvertes.fr/hal-00894156
Submitted on 1 Jan 1997
HAL is a multi-disciplinary open access
archive for the deposit and dissemination of
sci-entific research documents, whether they are
pub-lished or not. The documents may come from
teaching and research institutions in France or
abroad, or from public or private research centers.
L’archive ouverte pluridisciplinaire HAL, est
destinée au dépôt et à la diffusion de documents
scientifiques de niveau recherche, publiés ou non,
émanant des établissements d’enseignement et de
recherche français ou étrangers, des laboratoires
publics ou privés.
Gibbs sampling, adaptive rejection sampling and
robustness to prior specification for a mixed linear model
Cm Theobald, Mz Firat, R Thompson
To cite this version:
Original
article
Gibbs
sampling, adaptive rejection
sampling
and robustness
to
prior
specification
for
a
mixed
linear
model
CM
Theobald,
MZ Firat
R
Thompson
Department of
Mathematics andStatistics, University of Edinburgh,
King’s Buildings, Ma
y
f
ield
Road, Edinburgh
EH93JZ,
UK(Received
4 October1995; accepted
13January
1997)
Summary - Markov chain Monte-Carlo methods are
increasingly being applied
to make inferences about themarginal posterior
distributions of parameters inquantitative genetic
models. This paper considers theapplication
of one suchmethod,
Gibbssampling,
toBayesian
inferences about parameters in a normal mixed linear model when a restriction isimposed
on the relative values of the variance components. Twoprior
distributionsare
proposed
whichincorporate
this restriction. For one ofthem,
theadaptive rejection
sampling technique
is used tosample
from the conditionalposterior
distribution ofa variance ratio. Simulated data from a balanced sire model are used to compare
different
implementations
of the Gibbssampler
and also inferences based on the twoprior
distributions.
Marginal posterior
distributions of the parameters are illustrated. Numericalresults suggest that it is not necessary to discard any
iterates,
and that similar inferencesare made
using
the twoprior specifications.
adaptive rejection sampling
/
Bayesianinference /
Gibbs sampling/
mixed linearmodel / sire
modelRésumé -
Échantillonnage
de Gibbs, échantillonnage de rejet adapté et robustesse relative à l’a priori spécifié pour un modèle linéaire mixte. On utilise deplus
enplus
des méthodes de Monte-Carlo basées sur des chaînes de Markovpour faire
desinférences
surdes distributions
marginales
a posteriori desparamètres
de modèlesgénétiques quantitatifs.
Cet article concernel’application
d’une telleméthode, l’échantillonnage
deGibbs,
à desinférences bayésiennes
sur desparamètres
d’un modèle linéaire mixtenormal, quand
on
impose
une contrainte sur les valeurs relatives des composantes de variance. Deux distributions a prioriincorporant
cette contrainte sontproposées.
Pour l’uned’elles,
latechnique
del’échantillonnage
derejet adapté
estemployée
pour échantillonner dans unedistribution conditionnéllé a
posteriori
d’un rapport de variance. Des données simulées*
Present address:
Department
of AnimalScience, University
ofpukurova, Faculty
ofAgriculture,
01330Balcah, Adana, Turkey.
tp
dans un modèle
père équilibré
sont utilisées pour comparerdifférentes
mises en oeuvre del’échantillonnage
de Gibbs et aussi lesinférences
basées sur les deux distributions a priori. Les distributionsmarginales
aposteriori
desparamètres
sont données commeillustrations. Les
exemples numériques suggèrent qu’il
n’est nécessaire d’éliminer aucunrésultat d’itération et que des
inférences
similaires sontfaites
en utilisant l’un ou l’autre des deux a priori.échantillonnage de rejet adapté
/
inférence bayésienne/
échantillonnage de Gibbs/
modèle linéaire mixte
/
modèle pèreINTRODUCTION
Parameters such as
heritability,
andpredictors
of individualbreeding values,
de-pend
on the unknown variancecomponents
inquantitative
genetic
models. The widearray of methods available for
estimating
variancecomponents
includesANOVA,
likelihood-based methods and
Bayesian
methods. Adifficulty
with ANOVA meth-ods which has concerned many authors is that ANOVA estimates may lie outsidethe
parameter
space, forexample
estimates of heritabilities which arenegative
orgreater
than one. The use of such estimates forgenetic
parameters
can lead to veryinefficient selection decisions.
Owing
torapid
advances incomputing
technology,
methods based on likelihood have become common in animalbreeding;
inparticular
REML(Patterson
andThompson,
1971)
has beenwidely
used. REML estimators areby
definitionalways
within theparameter
space, andthey
areconsistent,
asymptotically
normal andefficient
(Harville,
1977).
However,
interval estimates for variances based on theasymptotic
distribution of REML estimators can includenegative
values(Gianola
andFoulley,
1990).
In recent years,
Bayesian
methods have beenapplied
to variancecomponent
estimation in animalbreeding
(Harville,
1977;
Gianola andFernando,
1986;
Gianolaet
al, 1986;
Foulley
etal, 1987;
Cantet etal, 1992; Wang
etal, 1993,
1994).
For normalmodels,
thejoint posterior
distribution of thedispersion
parameters
can bederived,
but numericalintegration techniques
arerequired
to obtain themarginal
distributions of functions of interest.
Gibbs
sampling
is a Markov chain Monte-Carlo method whichoperates
by
generating samples
from a sequence of conditionaldistributions;
under suitableconditions,
itsequilibrium
distribution is thejoint
posterior
distribution. Gilkset al
(1996)
provide
athorough
introduction to the Gibbssampler.
The methodis
applied
to variancecomponent
estimation in ageneral
univariate mixed linearmodel
by Wang
et al(1993).
Also,
Wang
et al(1994)
obtainmarginal
inferences about fixed and randomeffects,
variancecomponents
and functions of themthrough
a scalar version of the Gibbs
sampling
algorithm.
The
objective
of this paper is to broaden theapplication
of Gibbssampling
inan animal
breeding
contextby considering
prior
distributions whichincorporate
a constraint on the relative values of variancecomponents:
such a constraintmight
arise fromrestricting
heritability
to values less than one. The next section definessampling
can beapplied
to inferences about modelparameters
under the twospecifications.
Simulated data from a balanced univariate one-way sire model are used to demonstrate thesemethods,
to compare threeimplementations
of Gibbssampling
and to examine the robustness of the inferences to the choice ofprior.
THE MODEL AND ALTERNATIVE PRIOR DISTRIBUTIONS Model
We consider inferences about model
parameters
for a univariate mixed linear modelof the form
where y and e are n-vectors
representing
observations and residualerrors, 0
is ap-vector
of ’fixedeffects’,
u is aq-vector
of ’randomeffects’,
X is a known n x pmatrix of full column rank and Z is a known n x
q matrix.
Here e is assumed tohave the distribution
N
n
(0,
Q
e
I),
independently of fl
and u. Also u is taken to beN
9
(0, Q
u
G)
with G aknown positive
definite matrix.Restrictions on variance ratios
In animal
breeding
applications,
interest may be inmaking
inferences about ratios of variances or functionsthereof,
rather than about individual variancecomponents
themselves. Forexample,
if u and yrepresent
vectors of sire effects and ofphenotypic
values recorded on the sires’offspring
in unrelatedpaternal
half-sibfamilies,
and ’Y denotes the variance ratioor 2/
0
,2,
then theheritability,
h
2
,
of the trait is anincreasing
functionof -y
given
by
h!
=4/(1
+-y-
1
).
The constraintthat the
heritability
is between 0 and 1 restricts ’Y to lie in the interval(0, ! ) .
Amethod of inference which
ignores
such restrictions may lead to ridiculous estimates ofheritability.
Moregenerally,
wemight
considerrestricting -
Y
to an interval(0, g),
corresponding
to the constraint that 0 <Q!
<
g
Q
e,
where g is a known constant.Bayesian
formulation with varianceparameters
Q
e
and
Q!
u2In the above
model,
theassumption
made about the distribution of u must besupplemented by
prior
distributions forfl,
Q
e
and or u 2.
Inspecifying
theirform,
wemight
aim to reflect the animal breeder’sprior
knowledge
whilepermitting
convenient calculations on theresulting
posterior
distributions.The
prior
distributionof fl
is taken here to beuniform, corresponding
to littlebeing
known about its valueinitially. Wang
et al(1993,
1994)
make the sameprior
assumption
about P
(in
a model with further randomeffects),
and considertaking
ae and a
to
beindependent
with scaledinverse-x
2
distributions,
e
2
(v
-
X
(v
e
k
e
)-
1
)
andX
-
2
(v
u
,
);
1
)-
u
k
u
(v
ke
and
k§ !
can
then beinterpreted
asprior expectations
ofor-2
and
a
:
;¡
2
,
respectively,
and v. and vu asprecision
parameters
analogous
toThis
prior assumption
may be modified to take account of an upper bound onthe variance
ratio,
so thato,2and Q
u
are
notindependent,
but have ajoint prior
density
of the formThe conditional
density
of ygiven !,
u andafl
for the modelgiven
in[1]
isproportional
to(!)’s&dquo;exp[--<7!(y -
Xfl -
Zu)’(y -
Xfl -
Zu)],
so thejoint
posterior density
isgiven by
It is shown in the
Appendix
that if vu ispositive
and n + ve exceeds p then themarginal
posterior
distributionof q
is proper and hence thisjoint
distribution isalso.
To
implement
the Gibbssampling
algorithm,
werequire
the conditionalposterior
distributions of each of(3,
u,u
2
and
a
given
theremaining
parameters,
theso-called
full
conditionaldistributions,
which are as follows(using
the notation[. 1.]
for a conditionaldistribution)
(truncated
atga!).
Apart
from the use here of properprior
distributions for thevariances,
these are similar to the conditional distributionsgiven
inWang
et al(1993).
Methods forsampling
from these distributions are well known.An alternative
Bayesian
formulation withprior
independence
of
0
’
e 2 and
As
Wang
et al(1993)
show for a model with aslightly simpler
prior
forQ
e
and
o,,2,,
theprior
specification
in terms of these twoparameters
is a convenient one forapplying
the method of Gibbssampling.
It is alsoeasily generalized
when several traits are recorded. It maynot,
however,
be the most usefulspecification
foreliciting
may
prefer
to think in terms of theheritability
of atrait,
equivalent
for apaternal
half-sib model to the variance
ratio
7
=a£ lafl
defined above. A second functionof the two variances must also be considered in order to
specify
theirjoint
prior
distribution,
andor may
be anappropriate
choice because inferences aboutQ
e
2 fromprevious
dataare
likely
to be much moreprecise
than those about0’
u 2.
Toe
specify
ajoint prior
distribution for the sire and residual variancecomponents,
it may therefore be convenient toassign priors
toQ
e
and
the variance ratio y. Wethus consider the alternative
parameter
vector(p,
u,ae,
q),
but make the sameprior assumptions for
and u asbefore, noting
thatou
becomes-
YU2
.
We canincorporate
an upperbound g
on the rangeof q
morenaturally
than with theprevious
formulationby using
thefamily
of Beta distributions on the interval(0, g):
this is afairly
flexiblefamily
withprobability density
of the formIf y is taken to be
independent of P
and ea apriori,
anda
§
assumed to bex-2
(v
e
,
(
Le
k
e
)-
1
)
then thejoint posterior
density
isgiven
by
The full conditional
posterior
distributionsfor fl
and u are the same as[4]
and!5!,
while that for<7!
is
The full conditional
density for y
isproportional
toThis
family
is not a well-known one, so it is notimmediately
clear how tosample
from it.
However,
theadaptive
rejection
sampling
technique
of Gilks and Wild(1992)
provides
areasonably
efficient method forsampling
from the distribution ofln
7
.
Given these two methods of
including
the restriction onŒ!/ Œ;,
thequestion
arises of howposterior
inferences, especially
about q, are affected when the secondprior
specification
is used instead of the first. To answerthis,
we consider alternativespecifications
in which thepriors
for the varianceparameters
are similar: to match these distributions we choose a and b in[8]
togive -y
the same upper and lowerCALCULATIONS ON THE POSTERIOR DISTRIBUTIONS
Gibbs
sampling
Markov chain Monte-Carlo
methods, particularly
Gibbssampling,
are nowwidely
used for
making
inferences fromposterior
distributions ofparameters.
Gelfandand Smith
(1990)
review the Gibbssampler
and other Monte-Carlomethods,
and Gelfand et al(1990)
illustrate the use of the Gibbssampler
on a range of statisticalmodels, including
variancecomponents
models.Wang
et al(1993, 1994)
give
twoversions of the Gibbs
sampler
for a univariate normal mixed linear modelassuming
prior
independence
of the variancecomponents.
The Gibbs
sampler algorithm
generates
samples
iteratively
from the full condi-tional distributions ofdisjoint
subsets of theparameters
given
the current value ofthe
complementary
subset.The
following
version of the Gibbssampler
can beapplied
tosample
from theposterior
distribution defined in[3]
using
the full conditional distributionsgiven
in!4!-!7!.
Itincorporates
the restriction that(}&dquo;!/ ()&dquo;!
cannot exceed an upperbound g
by discarding
pairs
of variances which do notsatisfy
this condition.(i)
Choose anarbitrary
starting
value0(°)
=(p
CO
),
e u
for the vectorof
parameters 0
=(fl, u,
Q
e, or2 U);
(ii)
draw a valuep(
1
)
from the distribution[p
u(°),
e
o
a2(
),
0
l
u
2
(
0
),
y];
(iii)
drawu(
1
)
from[
u
!(1),
o
re
,au ,
y!;
(i
v
)
drawQ
u(
1
)
from[
o
r2 u
P(1),U(1)10,e 2(0),
y]
ignoring
the restriction onor
u/
or
e
;
2
2
(
v
)
draw0’e
from[
0
,2 e
!(1),
u(1),
),
u(
1
Q
y]
ignoring
the restriction onor
u/
u,;
2
2
(vi)
if (}&dquo;!(1) /(}&dquo;!C1) ;;::
g th
e
n
repeat
)
v
(i
and(
v
)
until(}&dquo;!Cl) /(}&dquo;!C1)
< g.Thus,
subsets of theparameters
aresampled
in anarbitrary
order and thiscycle
defines the transition from0(°)
toB(
1
).
Subsequent cycles replace
0(
l
)
togive
furthervectors
B(
2
), 0(!) ,
and so on.Opinions
differ on how the sequence of vectors used insubsequent analyses
shouldbe defined. The
following
three methods ofimplementation
have beenproposed,
and arecompared
on simulated data in the next section.(a)
Asingle long
chain: onegenerates
asingle
run of the chain aspractised
by
Geman and Geman
(1984),
ie,
(vii)
repeat
(ii)-(vi)
m timesusing
updated
values and store all the values.A common modification is to discard vectors from the start of the sequence to allow it to
approach equilibrium.
Theremaining
vectors0(
l
),
... ,,
0Cm
)
ap-proximate
simulated values from theposterior
distribution of0,
but successivevectors may be correlated and the sequence may be
strongly
influencedby
thestarting
values.(b)
Equally spaced
samples:
to reduce serialcorrelations,
one can choose suitableintegers
k and m,perform
asingle
run oflength
km,
and then form asample
from every kthvector,
ie,
useo(
k
),
0(
2k
),
... ,0(!!) .
(c)
Multiple
short chains:by
contrast,
Gelfand and Smith(1990)
and Gelfand et alIn
support
ofimplementation
(c),
Gelman and Rubin(1992)
arguethat,
even ifusing
onelong
run may be moreefficient,
it may still beimportant
to use several differentstarting
points.
It has been common
practice
to discard a substantial number of initialiterations,
and to useimplementations
(b)
and(c)
(especially
(b),
storing
only
everykth,
usually
10th or20th,
iterate).
The resultsof Raftery
and Lewis(1992)
suggest
that in many cases this is ratherprofligate.
For any individual
parameter,
the collection of m values can be viewed as a simulatedsample
from theappropriate
marginal
distribution. Thissample
can beused to calculate
summaries,
such as themarginal
posterior
mean, or to estimatemarginal posterior
distributions ofparameters
or of functions of themusing
kerneldensity
estimation. We use kerneldensity
estimation below with normal kernels and window widthsuggested by
Silverman(1986).
Adaptive rejection sampling
The
adaptive
rejection
sampling
technique
introducedby
Gilks and Wild(1992)
is a method ofrejection
sampling
intended for situations in which evaluation ofa univariate
density
iscomputationally
expensive.
It uses asqueezing
function(Devroye,
1986),
and isadaptive
in the sense that therejection
envelope
andsqueezing
function(which
arepiecewise
exponential)
are revised assampling
proceeds,
andapproach
thedensity
function. In Gibbssampling,
the full conditionaldistribution for each
parameter
changes
as the iterationproceeds,
so thatusually
only
one value isrequired
from aparticular density.
Adaptive
rejection
sampling
has the
advantage
that it does notrequire
anoptimization
step
to locate the mode of thedensity
before asample
can begenerated.
The method
requires
that thedensity
islog-concave,
that is that the derivative of thelogarithm
of thedensity
is monotonedecreasing.
It is shown in theAppendix
that thedensity of y
inexpression
!11!
isguaranteed
to belog-concave only
if a isgreater
than !q 2
+
1,
but that if we transform to 6 =In&dquo;(
then the full conditionaldensity
of 6 islog-concave
if b is at least one. The condition on a is not a reasonableone, since a
prior
with a maximum at zero must have a nogreater
than one, forexample.
However,
a value of b less than oneimplies
that theprior
density
in[8]
has a maximum at its upperbound,
g(which
might
correspond
to aheritability
ofone),
and it is not unreasonable to exclude such values. Hence we useadaptive
rejection
sampling
withln
7
.
APPLICATION TO A ONE-WAY CLASSIFICATION
The methods described above were
applied
to a one-way sire model of the formin which p and q
equal
1 and s, andX,
Z and G becomeI
sd
(a
vector ofIs with
length
sd), diag(l
d
, ... , l
d
)
andIs.
The full conditional distributionsfor
{3,
u,o,2
and
Q
u
defined
inexpressions
[4]-[7]
becomeN !
I
where
y,, Y..
and u. denote the vector offamily
means andthe
overall means of the yg and the ui.Four sets of data were
generated using
this model withparameter
values 0
=0,
a2
=0.975,
er!
= 0.025 and
hence h
2
= 0.1 with 25 families and 20offspring
per
sire. ANOVA estimates are summarized in table I: the data sets are ordered
by
the estimates ofheritability,
which range from -0.073 to 0.306. The first ANOVAestimate of
a
2
is
negative,
making
conventional inference about thisparameter
difficult.
Analysis
using
theprior
forQ
e
and Q
u
u2
Bayesian analyses
of the four data sets aregiven using
theprior
distribution in[2]
with v! = vu =I,
k!
=0.975, ku
= 0.025 andg
1
Theprior
distribution for with v, = vu =1,
k
e
=0.975, k
u
= 0.025 and9
= 3.
Theprior
distribution forQ
e
and au
is
proper and leads to a properjoint posterior
distribution(as
shownin the
Appendix),
but the small values chosen for v, and Vucorrespond
to a verydispersed
distribution;
this is intended to reveal anyproblems
with convergence ofthe Gibbs
sampler
algorithm.
Thesample
dataprovide
little information aboutQ
u
uas there are
only
25 sires. Thealgorithm
is carried out for the four sets of data with thefollowing
implementations,
which eachprovide
1 000 Gibbssamples:
(a)
asingle
run of 1000iterations;
(b)
taking
every(1)
10th,
(2)
20th and(3)
30th value insingle
runs oflength
10 000,
20 000 and30 000,
respectively, using only
onestarting
value for theparameter
vector;
(c)
short runs oflength
(1)
10,
(2)
20 and(3)
30,
storing
the last iterate andThe
parameter
values used togenerate
the data were alsoemployed
asstarting
values for all threeimplementations.
No iterates were discarded since the trueparameter
values were known.The
resulting marginal
posterior
means and standard deviations for{3,
0
&dquo;;, a2 e
q and
h
2
for the four data sets are shown in table II. This table reveals little difference between theimplementations
in themarginal
posterior
means,except
fora
slight tendency
for standard deviations to be lowerusing
(a)
for{3,
a u, 2 -y
andh
2
Figures
1-4 show the estimatedmarginal
posterior
densities for{3,
0
&dquo;;, or2
and
h
2
2using
density
estimators with normal kernels under theprior
specification
of this subsection(with
implementation
(a))
and that considered in the next subsection.Analysis assuming prior independence
of0
&dquo;; e 2and 7
We now compare the above
analyses
with those based on the secondprior
specification,
which assumesprior
independence
ofQ
e
and q
andprior
density
[8]
for!y; we use
adaptive
rejection
sampling
on the full conditionalposterior
distribution of 6 = In q, as defined in theAppendix.
TheAppendix
also describes how the values of a and b in the Betaprior
for -y
are determined from those of v,, vu,k,
andk!,.
At each Gibbs
sampling
iteration,
adaptive rejection sampling
requires
at leasttwo
points
for the construction of therejection
envelope
andsqueezing
function(Gilks
andWild,
1992).
Theparameter
space for 6 is unbounded on theleft,
so Gilks and Wild(1992)
advise that the smallest of these initialpoints
be chosen where the derivative of thelog
posterior
density
ispositive.
The 15th and 85thpercentiles
of thesampling density
from theprevious
iteration were tried as initialpoints,
additionalpoints
being supplied
if the derivatives at these values had the samesign.
Three evaluations of thelog
posterior
density
of 6 wererequired
onaverage at each iteration of
adaptive
rejection
sampling.
Convergence
of the Gibbssampler,
assessedby
monitoring
summary statisticsfor each
parameter
based on every teniterations,
hadclearly
occurred before 1 000 iterations. Themarginal
posterior
summaries shown in table II and thedensity
estimates for{3,
o, 2, afl
andh
2
infigures
1-4 are based on this number of iterations. The results obtainedusing
the twoprior
specifications
agree veryclosely, especially
those for
/3
andce
2
DISCUSSION
Our results from
using
Gibbssampling
on a balanced one-way sire model withthe
prior
density
ofexpression
[2]
for0
&dquo;; and
0
&dquo;; suggest
that it is unnecessary todiscard all but the
10th,
20th or 30thiterate;
doing
so isprobably
quite
wasteful. A muchlarger
proportion
of the iterates should be used unlessstorage
is an issue.Using
1000 successive iterates seemsadequate
to make inferences about theparameters
ofinterest,
although
theslightly
lower standard deviations obtained withimplementation
(a)
may reflect the need to discard ’burn-in’ iterations. We encountered no difficulties with the convergence of the Gibbssampling algorithms
used. With real
data,
we would use REML estimates asstarting
values for theGibbs iterations and monitor convergence
diagnostics
to decide how manyThe
prior
distribution which assumesu2
and
-y to beindependent provides
apotentially
useful alternative to the usualprior
in whichU2
and
Q
u
are
taken to beindependent.
Theadaptive rejection
sampling
algorithm
of Gilks and Wild(1992)
is found to workefficiently
with theresulting
full conditional distribution forIn q.
The
sensitivity
of aBayesian
analysis
to the choice ofprior
hyperparameters
andto different functional forms of
prior
can beinvestigated.
For our simulateddata,
the use of an alternative form of
prior
distribution led to very similarmarginal
posterior
inferences on eachparameter.
The use of the alternative
prior
distribution forQ
e
e 2 and -y
can be extended tomodels with further variance
components,
forexample
if[1]
isgeneralized
towhere v is an r-vector of random
effects,
T is a known n x r matrix and v is taken tobe
N,(0,
Q
v
H)
independently
offl,
u and e with H a known matrix. If!,
!e
and
U2
v are taken to bemutually independent
apriori,
withuniform,
x-
2
(v
e
,
(veke)-1)
andx-2
(
L
v,
(v
v
k
v
)-
1
)
distributions,
respectively,
and q
isindependent
of them with theBeta distribution defined in
[8],
then the full conditional distributions for!,
u, v,!e
eand
Q
v
are,respectively,
where e denotes
a 2/
0
,2,
while thatfor q
is the same as!11!.
ACKNOWLEDGEMENTS
The authors are
grateful
to W Gilks forproviding
a FORTRAN program toperform
adaptive rejection sampling.
MZ Firat wassupported by
the Turkish Government.REFERENCES
Cantet
RJC,
FernandoRL,
Gianola D(1992)
Bayesian
inference aboutdispersion
parame-ters of univariate mixed models with maternal effects: theoretical considerations. Genet Sel Evol
24,
107-135Devroye
L(1986)
Non-Uniform
Random Variate Generation.Springer-Verlag,
New YorkFoulley JL,
ImS,
GianolaD,
Hoschele I(1987)
Empirical Bayes
estimation of parametersGelfand
AE,
Smith AFM(1990)
Sampling-based approaches
tocalculating marginal
densities. J Am Stat Assoc 85, 398-409
Gelfand
AE,
HillsSE,
Racine-PoonA,
Smith AFM(1990)
Illustration ofBayesian
inference in normal data modelsusing
Gibbssampling.
J Am Stat Assoc 85, 972-985Gelman, A,
Rubin DB(1992)
Asingle
series from the Gibbssampler provides
a false senseof
security.
In:Bayesian Statistics
4 (JM
Bernardo,
JOBerger,
APDavid,
AFMSmith,
eds),
ClarendonPress, Oxford,
625-631Geman
S,
Geman D(1984)
Stochasticrelaxation,
Gibbs distributions and theBayesian
restoration ofimages.
IEEE Transactions on PatternAnalysis
and MachineIntelligence
6, 721-741Gianola
D,
Fernando RL(1986)
Bayesian
methods in animalbreeding theory.
J Anim Sci63, 217-244
Gianola
D, Foulley
JL(1990)
Variance estimation fromintegrated
likelihoods(VEIL).
Genet Sel Evol22,
403-417Gianola
D, Foulley JL,
Fernando RL(1986)
Prediction ofbreeding
values when variancesare not known. Genet Sel Evol
18,
485-498Gilks
WR,
RichardsonS, Spiegelhalter
DJ(1996)
Markov Chain Monte Carlo in Practice.Chapman
andHall,
LondonGilks
WR,
Wild P(1992)
Adaptive rejection sampling
for Gibbssampling. A
PP
I
Statist41, 337-348
Harville DA
(1977)
Maximum likelihoodapproaches
to variance component estimation and to relatedproblems.
J Am Stat Assoc72,
320-338Patterson
HD, Thompson
R(1971)
Recovery
of inter-block information when block sizesare
unequal.
Biometrika 58, 545-554Raftery AE,
Lewis SM(1992)
How many iterations in the Gibbssampler?
In:Bayesian
Statistics
4 (JM
Bernardo,
JOBerger,
APDavid,
AFMSmith,
eds),
ClarendonPress,
Oxford,
763-773Silverman BW
(1986)
Density
Estimationfor
Statistics and DataAnalysis. Chapman
andHall,
LondonWang CS, Rutledge JJ, Gianola,
D(1993)
Marginal
inferences about variance components in a mixed linear modelusing
Gibbssampling.
Genet SelEvol 25,
41-62Wang CS, Rutledge JJ,
Gianola D(1994)
Bayesian analysis
of mixed linear models via Gibbssampling
with anapplication
to litter size in Iberianpigs.
Genet Sel Evol26,
91-115APPENDIX
Marginal posterior probability density
functionof -y using
theprior
for92
and
u
2
uIntegrating
theposterior
density
forfl,
u,Q
e
and
0
-;
given
in[3]
withrespect
to fl
and u, theposterior density
ofQ
e
and
0
-;
is
found to beproportional
toThe
joint
density
ofQ
e
and y
is thereforeproportional
toand the
marginal
posterior
density of y
is
proportional
toprovided
that n + LIe + v,, - P ispositive.
Whether or not this is a properdensity
function
depends
on its behaviour near zero. For small !y,I G-
1
+
7
Z’QZI
isapproximately
constant,
while the&dquo;(-
l
Llu
k
u
term is dominant inDb)
solong
asv
u is
positive.
So thedensity of
7
behaves like&dquo;(!(n+ve-p-2)
near
zero, and has afinite
integral
if theexponent
exceeds -1,
that is if n + ve exceeds p.Matching
quartiles of q
in the twoprior
distributionsTo compare inferences under the two
prior
distributions considered in thisarticle,
we determine values of thehyperparameters
a and bdefining
the Beta distributionfor y
in the secondprior
so that the twospecifications
are matched in some sense. The firstspecification
for the varianceparameters
(in
terms ofQ
e
and u u 2)
implies
that&dquo;(
(which
equals
o, 2 /C2 )
hasprobability density
functionproportional
toHence 0 =
-yk,lk,,
has anF(v
e
,
vu)
distribution truncated atgk
e/
k
u
.
We choosea and b to match the Beta distribution used for -y in the second
prior
to thisby
equating
theirquartiles.
If G denotes the distribution function ofF(v
e
,
vu)
then 0
has lower and upperquartiles O
L
,
O
U
which are solutions ofand -y
hasquartiles
(/J
L
k
e/
k
u
andouk,lk,,.
Thus if xL and xU denote the lower and upperquartiles
of the Beta distributionB(a, b),
weequate
XL and xU tocP
L
k
u/
(gk
e
)
and
Ouk.l(gk, ,),
respectively, by
solving
for a and b the twoequations
Example:
We use theprior
specification
foror2and
or
with
ve = l/u =1,
k
e
=0.975,
k
u
= 0.025 andg
= 3,
so that’Y
ke
/
k
u
hasan
F(l, 1)
distributiontruncated at
-ke/ku 1
=13;
G(13)
=0.8278,
soThe
corresponding
values of a and b are 0.4038 and 3.0678.Log-concavity
of the full conditional distributionsof
7
andln
7
A
density
is defined to belog-concave
if the derivative of itslogarithm
is monotonedecreasing.
For thedensity of q
in!11!,
therequired
derivative isThe second and third terms are
decreasing
in q
if b exceeds1,
but the first isnot unless a is
greater
than2
q +1;
log-concavity
cannot beguaranteed
for allafl
2 without this condition.If we transform to 6 =
In q
(and
include the Jacobian of thetransformation),
thelogarithm
of thedensity
is,
apart
from an additiveconstant,
and this has derivative
The first term is constant and the second and third are