HAL Id: hal-00894092
https://hal.archives-ouvertes.fr/hal-00894092
Submitted on 1 Jan 1995
HAL is a multi-disciplinary open access
archive for the deposit and dissemination of
sci-entific research documents, whether they are
pub-lished or not. The documents may come from
teaching and research institutions in France or
abroad, or from public or private research centers.
L’archive ouverte pluridisciplinaire HAL, est
destinée au dépôt et à la diffusion de documents
scientifiques de niveau recherche, publiés ou non,
émanant des établissements d’enseignement et de
recherche français ou étrangers, des laboratoires
publics ou privés.
Bayesian inference in threshold models using Gibbs
sampling
Da Sorensen, S Andersen, D Gianola, I Korsgaard
To cite this version:
Original
article
Bayesian
inference
in
threshold models
using
Gibbs
sampling
DA Sorensen
S Andersen
2
D
Gianola
I
Korsgaard
1
National Institute
of
AnimalScience,
Research CentreFoulum,
PO Box39,
DK-8830Tjele;
2
National Committee
for Pig Breeding,
Health andProdv,ction,
Axeltorv3,
Copenhagen V, Denmark;
3
University of Wisconsin-Madison, Department of
Meat and AnimalSciences,
Madison, WI53706-1284,
USA(Received
17 June1994; accepted
21 December1994)
Summary - A
Bayesian analysis
of a threshold model withmultiple
orderedcategories
is
presented. Marginalizations
are achievedby
means of the Gibbssampler.
It is shownthat use of data
augmentation
leads to conditionalposterior
distributions which are easy tosample
from. The conditionalposterior
distributions of thresholds and liabilities areindependent
uniforms andindependent
truncatednormals, respectively.
Theremaining
parameters of the model have conditional
posterior
distributions which are identical tothose in the Gaussian linear model. The
methodology
is illustratedusing
a siremodel,
with an
analysis
ofhip dysplasia
indogs,
and the results arecompared
with those obtainedin a
previous study,
based onapproximate
maximum likelihood. Twoindependent
Gibbs chains oflength
620 000 each were run, and the Monte-Carlosampling
error of momentsof
posterior
densities were assessedusing
time series methods. Differences between resultsobtained from both chains were within the range of the Monte-Carlo
sampling
error.With the
exception
of the sire variance andheritability, marginal posterior
distributions seemed normal. Hence inferencesusing
the present method were ingood
agreement with those based onapproximate
maximum likelihood. Threshold estimates werestrongly
autocorrelated in the Gibbs sequence, but this can be alleviated
using
an alternativeparameterization.
threshold
model / Bayesian
analysis/
Gibbs sampling/
dogRésumé - Inférence bayésienne dans les modèles à seuil avec échantillonnage de
Gibbs. Une
analyse bayésienne
du modèle à seuil avec descatégories multiples
ordonnéesest
présentée
ici. Lesmarginalisations
nécessaires sont obtenues paréchantillonnage
deGibbs. On montre que l’utilisation de données
augmentées -
la variable continue sous-jacente non observée étant alors considérée comme une inconnue dans le modèle -con-duit à des distributions conditionnelles a
posteriori
faciles
à échantillonner. Celles-ci sonttronquées indépendantes
pour les sensibilités(les
variablessous-jacentes).
Lesparamètres
restants du modèle ont des distributions conditionneLles aposteriori identiques
à cellesqu’on
trouve en modèle linéairegaussien.
Laméthodologie
est illustrée sur un modèlepaternel appliquée
à unedysplasie
de la hanche chez lechien,
et les résultats sontcom-parés
à ceux d’une étudeprécédente
basée sur un maximum de vraisemblanceapproché.
Deuxséquences
de Gibbsindépendantes, longues
chacune de 620 000échantillons,
ont été réalisées. Les erreursd’échantillonnage
de type Monte Carlo des moments des densités aposteriori ont été obtenues par des méthodes de séries
temporelles.
Les résultats obtenusavec les 2
séquences indépendantes
sont dans la limite des erreursd’échantillonnage
de Monte-Carlo.
À
l’exception
de la variancepaternelle
et del’héritabilité,
lesdistribu-tions
marginales
a posteriori semblent normales. De cefait,
lesinférences
basées sur laprésente
méthode sont en bon accord avec celles du maximum de vraisemblanceapproché.
Pour l’estimation des
seuils,
lesséquences
de Gibbs révèlent defortes autocorrélations,
auxquelles
il estcependant possible
de remédier en utilisant un autreparamétrage.
modèle à seuil
/
analyse bayésienne/
échantillonnage de Gibbs/
chienINTRODUCTION
Many
traits in animal andplant
breeding
that arepostulated
to becontinuously
inherited arecategorically
scored,
such as survival and conformation scores,degree
ofcalving difficulty,
number ofpiglets
born dead and resistance to disease. Anappealing
model forgenetic
analysis
ofcategorical
data is based on the thresholdliability
concept,
first usedby
Wright
(1934)
in studies of the number ofdigits
inguinea pigs,
andby
Bliss(1935)
intoxicology
experiments.
In the thresholdmodel,
it is
postulated
that there exists a latent orunderlying
variable(liability)
which has a continuous distribution. A response in agiven
category
isobserved,
if the actual value ofliability
falls between the thresholdsdefining
theappropriate
category.
Theprobability
distribution of responses in agiven population depends
on theposition
of its mean
liability
withrespect
to the fixed thresholds.Applications
of this model in animalbreeding
can be found in Robertson and Lerner(1949),
Dempster
and Lerner(1950)
and Gianola(1982),
and in Falconer(1965),
Morton and McLean(1974)
and Curnow and Smith(1975),
in humangenetics
andsusceptibility
todisease.
Important
issues inquantitative genetics
and animalbreeding
includedrawing
inferences about(i)
genetic
and environmental variances and covariances inpopulations;
(ii)
liability
values of groups of individuals and candidates forgenetic
selection;
and(iii)
prediction
and evaluation of response to selection. Gianola andFoulley
(1983)
usedBayesian
methods to deriveestimating equations
for(ii)
above,
assuming
known variances. Harville and Mee(1984)
proposed
anapproximate
method for variancecomponent
estimation,
andgeneralizations
to severalpolygenic
binary
traitshaving
ajoint
distribution werepresented by Foulley
et al(1987).
In these methods inferences aboutdispersion
parameters
were based on the mode of theirjoint posterior distribution,
afterintegration
of locationparameters.
This involved the use of a normalapproximation which,
seemingly,
does not behave well in sparsecontingency
tables(H6schele
etal,
1987).
These authors found thatcombination of fixed and random levels in the model was smaller than
2,
andsuggested
that this may be causedby inadequacy
of the normalapproximation.
Thisproblem
can render the method less useful for situations where the number of rows in acontingency
table isequal
to the number of individuals. A datastructure such as this often arises in animal
breeding,
and is referred to as the ’animal model’(Quaas
andPollak,
1980).
Anderson and Aitkin(1985)
proposed
amaximum likelihood estimator of variance
component
for abinary
threshold model. In order to construct thelikelihood, integration
of the random effects was achievedusing
univariate Gaussianquadrature.
Thisprocedure
cannot be used when the random effects arecorrelated,
such as ingenetics. Here, multiple
integrals
ofhigh
dimension would need to becalculated,
which is unfeasible even in data sets withonly
50genetically
related individuals. In animalbreeding,
a data set may contain thousands of individuals that are correlated to differentdegrees,
and some of thesemay be inbred.
Recent reviews of statistical issues
arising
in theanalysis
of discrete data inanimal
breeding
can be found inFoulley
et al(1990)
andFoulley
and Manfredi(1991).
Foulley
(1993)
gaveapproximate
formulae forone-generation predictions
of response to selectionby
truncation forbinary
traits based on asimple
threshold model.However,
there are no methods described in the literature fordrawing
inferences aboutgenetic
change
due to selection forcategorical
traits in the contextof threshold models.
Phenotypic
trends due to selection can bereported
in terms ofchanges
in thefrequency
of affected individuals.Unfortunately,
due to the nonlinearrelationship
betweenphenotype
andgenotype,
phenotypic changes
do not translatedirectly
into additivegenetic changes,
or, in otherwords,
to response to selection. Here wepoint
out that inferences about realized selection response forcategorical
traits can be drawnby extending
results for the linear model described in Sorensenet al
(1994).
With the advent of Monte-Carlo methods for numerical
integration
such as Gibbssampling
(Geman
andGeman, 1984;
Gelfand etal,
1990),
analytical
approximations
to
posterior
distributions can beavoided,
and a simulation-basedapproach
toBayesian
inference aboutquantitative genetic
parameters
is nowpossible.
In animalbreeding, Bayesian
methodsusing
the Gibbssampler
wereapplied
in Gaussian modelsby Wang
et al(1993, 1994a)
and Jensen et al(1994)
for(co)variance
component
estimation andby
Sorensen et al(1994)
andWang
et al(1994b)
forassessing
response to selection.Recently,
a Gibbssampler
wasimplemented
forbinary
data(Zeger
andKarim,
1991)
and ananalysis
ofmultiple
threshold models was describedby
Albert and Chib(1993).
Zeger
and Karim(1991)
constructed the Gibbssampler using rejection
sampling techniques
(Ripley, 1987),
while Albert and Chib(1993)
used it inconjunction
with dataaugmentation,
which leads toa
computationally simpler
strategy.
The purpose of this paper is to describe a Gibbssample
for inferences in threshold models in aquantitative
genetic
context.First,
theBayesian
threshold model ispresented,
and all conditionalposterior
distributions needed forrunning
the Gibbssampler
aregiven
in closed form.Secondly,
aquantitative genetic
analysis
ofhip
dysplasia
in Germanshepherds
ispresented
as anillustration,
and 2 differentparameterizations
of the modelleading
MODEL FOR BINARY RESPONSES
At the
phenotypic
level,
a Bernoulli random variableY
i
is observed for each individuali (i
=1, 2, ... , n)
taking
values y
i
=1 or y
2
= 0(eg,
alive ordead).
The variable
Y
is theexpression
of anunderlying
continuous random variableU
i
,
theliability
of individual i. WhenU
Z
exceeds an unknown fixed thresholdt,
thenY
=1, and Y
= 0 otherwise. We assume thatliability
isnormally
distributed,
with the mean value indexedby
aparameter
0,
and,
without loss ofgenerality,
that it has unit variance(Curnow
andSmith,
1975).
Hence:where
0’ =
(b’, a’)
is a vector ofparameters
with p fixed effects(b)
and
q random
additivegenetic
values(a),
andw’
is a row incidence vectorlinking
e to the ith observation.It is
important
to note thatconditionally
on0,
theU
i
areindependent,
so for the vector U ={U
i
}
given 0,
we have asjoint
density:
where
!U(.)
is a normaldensity
withparameters
as indicated in theargument.
In!2!,
put
WO = Xb +Za,
where X and Z are known incidence matrices of order nby
p and nby
q,respectively,
and,
without loss ofgenerality,
X is assumed to have full column rank. Given themodel,
we have:where
<p(.)
is the cumulative distribution function of a standardized normal variate. Without loss ofgenerality,
andprovided
that there is a constant term in themodel,
t can be set to
0,
and[3]
reduces toConditionally
on both 0 andon Y
i
=y 2
,
U
i
follows a truncated normal distribution. Thatis,
for Yi = 1:where
I(X
EA)
is the indicator function that takes the value 1 if the random variable X is contained in the setA,
and 0 otherwise. For Yi =0,
thedensity
is§
u
; (w§ e, I) /V(-w§ e) I (U
i
x 0) .
Invoking
an infinitesimal model(Bulmer, 1971),
the conditional distribution of additivegenetic
valuesgiven
the additivegenetic
variance in theconceptual
basewhere A is a q
by
q matrix of additivegenetic relationships.
Note that a can include animals withoutphenotypic
scores.We discuss next the
Bayesian inputs
of the model. The vector of fixed effects b will be assumed to follow apriori
theimproper
uniform distribution:For a
description
ofuncertainty
about the additivegenetic
variance,
or a 2,
an inverted gamma distribution can beinvoked,
withdensity:
where v and
S’
areparameters.
When v = -2 andS
Z
=0,
[8]
reduces to theimproper
uniformprior
distribution. A proper uniformprior
distribution forQ
a
is:
where k
a
is a constant anda a 2m!’. is
the maximum value whichJ£
can take apriori.
To facilitate thedevelopment
of the Gibbssampler,
the unobservedliability
Uis included as an unknown
parameter
in the model. Thisapproach,
known as dataaugmentation
(Tanner
andWong,
1987;
Gelfand etal, 1992;
Albert andChib, 1993;
Smith and
Roberts,
1993)
leads to identifiable conditionalposterior distributions,
as shown in the next section.Bayes
theoremgives
asjoint posterior
distribution of theparameters:
The last term is the conditional distribution of the data
given
theparameters.
We noticethat,
forY
i
=1,
say, we have
For
Y
i
=
0,
we have:The
joint posterior
distribution[10]
can then be written as:where the
conditioning
onhyperparameters
v andS’
isreplaced
by
0
a
ma
.
when the uniformprior
[9]
for the additivegenetic
variance isemployed.
Conditional
posterior
distributionsIn order to
implement
the Gibbssampler,
all conditionalposterior
distributions of theparameters
of the model are needed. Thestarting point
is the fullposterior
distribution!13!.
Among
the 4 terms in(13!,
the third is theonly
one that is a function of b and we therefore have for the fixed effects:which is
proportional
to!U(Xb+Za, I).
As shown inWang
et al(1994a),
the scalar form of the Gibbssampler
for the ith fixed effect consists ofsampling
from:where xi is the ith column of the matrix
X, and b
i
satisfies:In
!16!,
X-
i
is the matrix X with the column associated withi deleted,
andb_
i
is b with the ith element deleted. The conditionalposterior
distribution of thevector of
breeding
values isproportional
to theproduct
of the second and thirdterms in
!13!:
which has the form
!(0,Acr!)!(u!b,a).
Wang
et al(1994a)
showed that the scalar Gibbssampler
drawssamples
from:where zi is the ith column of
Z,
cii is the element in the ith row and column ofA-1 , /B B =
(Qa)-1,
and a
i
satisfies:In
[19],
ci,-i is the row ofA-
1
corresponding
to the ith individual with the ith element excluded. We notice from[14]
and[17],
thataugmenting
with theFor the variance
component,
we have from!13!:
Assuming
that theprior
foro,2is
the inverted gammagiven
in!8!,
this becomes:and
assuming
the uniformprior
!9!,
it becomes:Expression
!21a!
is in the form of a scaled inverted gammadensity,
and[21b]
inthe form of a truncated scaled inverted gamma
density.
The conditional
posterior
distribution of theunderlying
variableU
i
ispropor-tional to the last 2 terms in
!13!.
This can be seen to be a truncated normaldis-tribution,
on the left ifY
i
= 1 and on theright
otherwise. Thedensity
function ofthis truncated normal distribution is
given
in!5!.
Thus,
depending
on the observedYi,
we have:or
Sampling
from the truncated distribution can be doneby generating
from the untruncated distribution andretaining
those values which fall in the constraintregion.
Alternatively
and moreefficiently,
suppose that U is truncated and defined in the interval!i, j]
only,
wherei and j
are the lower and upperbounds,
respectively.
Let the distribution function of U beF,
and let v be a uniform[0, 1]
variate. Then U =F-1 !F(i)
+v(F(j) —
F(i))!
is adrawing
from the truncated random variable(Devroye,
1986).
Albert and Chib
(1993)
also constructed the mixed model in terms of a hier-archicalmodel,
butproposed
a blocksampling
strategy
for theparameters
in theunderlying
scale,
instead.Essentially, they
suggest
sampling
from the distributions(
Q
a,
a,blU)
as(a!IU)(a,
blU,
J£
),
instead of from the full conditionalposterior
distributions
!15!,
!18!
and!21!,
and
they
assumed a uniformprior
forlog(a2)
in a finite interval. To facilitatesampling
fromp(or2lU),
they
use anapproximation
which consists ofplacing
allprior probabilities
on agrid
ofor2
values,
thusmaking
theprior
and theposterior
discrete. The need for thisapproximation
isquestion-able,
since the full conditionalposterior
distribution of(T has
asimple
form as noted in[21]
above. Inaddition,
in animalbreeding,
the distribution(a, b[U,
a a 2)
is ahigh
dimensional multivariate normal and it would not besimple computationally
MULTIPLE ORDERED CATEGORIES
Suppose
now that the observed random variable Y can take values in one of Cmutually
exclusive orderedcategories
delimitedby
C + 1 thresholds. Letto
=- oo,
t
c
= +oo, with theremaining
thresholdssatisfying
t1 ! t
2
... <
t
C
-
1
-Generalizing
[3]:
Conditionally
onA,
Y
i
=j,
t
j
-1
andt
j
,
theunderlying
variable associated with the ith observation follows a truncated normal distribution withdensity:
Assuming
thato, a, 2
b and t areindependently
distributed apriori,
thejoint
posterior
density
is written as:where
p(Ulb,
a,t)
=p(Ulb, a).
Generalizing
[12],
the last term in[25]
can beexpressed
as(Albert
andChib,
1993):
All the conditional
posterior
distributions needed toimplement
the Gibbssam-pler
can be derived from!25!.
It is clear that the conditionalposterior
distributions ofb
i
,
ai andu2
are the same as for thebinary
response model andgiven
in(15!,
[18]
and!21!.
For theunderlying
variable associated with the ith observation we have from!25!:
This is a truncated
normal,
withdensity
function as in!24!.
The thresholds t =
(t
l
,
t
2
, ... ,
tC-1)
areclearly
dependent
apriori,
since the modelpostulates
that these are distributed as order statistics from a uniform distribution in the interval[t
mini
t
max]
.
However,
the full conditionalposterior
distributions of the thresholds are
independent.
Thatis,
p(t
j
It-
j
, b, a, U, o
l
a
2
,
Y
) =
p(t
where T =
{(h, t2,&dquo;’, tc-dltmin !
/ x t2 ! ... ! t
1
v
-
C
t
max
)
(Mood
eta
l
,
1974).
Note that the thresholds enteronly
indefining
thesupport
ofp(t).
The conditionalposterior
distributionof t
j
isgiven by:
which has the same form as
!26!.
Regarded
as a functionof t,
[26]
showsthat, given
U and y, the upper bound ofthreshold t
j
ismin (U I Y
= j
+1)
and the lower boundis
max(UIY
=j).
The apriori
condition t E T isautomatically
fulfilled,
and the bounds are unaffectedby
knowledge
of theremaining
thresholds. Thust
j
has a uniform distribution in this intervalgiven
by:
This
argument
assumes that there are nocategories
withmissing
observations. To accommodate for thepossibility
ofmissing
observations in 1 or morecategories,
Albert and Chib(1993)
define the upper and lower bounds of thresholdj,
asminfmin(UIY
= j
+1), t
j+d
and asmax{max(U!Y
=j),t
j
-’
11
,
respectively.
In this case, the thresholds are notconditionally independent.
The Gibbssampler
isimplemented
by sampling repeatedly
from!15!, !18!, !21!, [24]
and(28!.
Alternativeparameterization
of themultiple
threshold modelThe
multiple
threshold model can also beparameterized
such that the conditional distribution of theunderlying
variableU, given 0,
has unknown variance0’
instead
of unit variance. Theequivalence
of the 2parameterizations
is shown in theAppendix.
Thisparameterization requires
that records fall in at least 3mutually
exclusive orderedcategories;
for Ccategories,
only
C-3 thresholds are identifiable.In this new
parameterization,
one mustsample
from the conditionalposterior
distributionof o, e 2.
Under thepriors
[8]
or(9!,
the conditionalposterior
distribution ofJfl can
be shown to be in the form of a scaled inverted gamma. Theparameters
of this distributiondepend
on theprior
used forae
2
If this is in the form(8!,
thenwhere,
SSE =(U - Xb - Za)’ (U - Xb - Za),
and v, andS
e
areparameters
of theprior
distribution. If a uniformprior
of the form[9]
is assumed to describe theprior
uncertainty
aboutu . 2,
the conditionalposterior
distribution is a truncated version of[29]
(ie
[21b]),
with v
e
= -2and S,2
= 0. Withexactly
3categories,
the GibbsEXAMPLE
We illustrate the
methodology
with ananalysis
of data onhip
dysplasia
in Germanshepherd dogs.
Results of anearly
analysis
and a fulldescription
of the data can be found in Andersen et al(1988).
Briefly,
the records consisted ofradiographs
of 2 674offspring
from 82 sires. Theseradiographs
had been classifiedaccording
toguidelines approved by
FCI(Federation
Cynologique
Internationale,
1983),
eachoffspring
record was allocated to 1 of 7mutually
exclusive orderedcategories.
The model for theunderlying
variable was:.
where ai is the effect of sire
i (i
=1, 2, ... , 82; j
=1, 2,... , n
i
).
Theprior
distribution of /! was as in
[7]
and sire effects were assumed to follow the normal distribution:The
prior
distribution of the sire variance(
Q
a)
was in the formgiven
in!8!,
with v = 1 andS
2
= 0.05. Theprior
fort
2
, ... , t6
was chosen to be uniform on the ordered subset of[f,
=-1.365,
f
7
=+00!5
for whichti
<t
2
< ... <t
7
,
wheref
l
was the value at whicht
1
wasset,
andf
7
is the value of the 7th threshold. The value forf
l
was obtained from Andersen et al(1988),
in order to facilitatecomparisons
with thepresent
analysis.
Theanalysis
was also carried out under theparameterization
where the conditional distribution of Ugiven
0 has variancea 2
Here,
Q
e
was assumed to follow aprior
of the form of!8!,
with v = 1 andS
Z
= 0.05and
t
6
was set to 0.429. Results of the 2analyses
weresimilar,
soonly
those from the secondparameterization
arepresented
here.Gibbs
sampler
andpost
Gibbsanalysis
The Gibbs
sampler
was run as asingle
chain. Twoindependent
chains oflength
620 000 each were run, and in both cases, the first 20 000samples
were discarded.Thereafter, samples
were saved every 20iterations,
so that the total number ofsamples kept
was 30 000 from each chain. Start values for theparameters
were, for the case of chain1,
o,2=
2.0,
Q
a
=0.5,
t
2
=-0.8, t
3
=
-0.5,
t
4
=-0.2,
t
5
= 0.1. Forchain
2,
estimates from Andersen et al(1988)
wereused,
and these wereJ
fl
=1.0,
or =
0.1, t
2
=-1.05,
t
3
=4
-0.92, t
=-0.62, t
5
= -0.34. In both runs,starting
values for sire effects were set to zero.
Two
important
issues are the assessment of convergence of the Gibbssampler,
and the Monte-Carlo error of estimates of features of
posterior
distributions. Both issues are related to thequestion
of whether thechain,
orchains,
have been runlong enough.
This is an area of active research in which someguidelines
based on theoretical work(Roberts,
1992; Besag
andGreen, 1993;
Smith andRoberts,
1993;
Roberts andPolson,
1994)
and onpractical
considerations(Gelfand
etal,
1990;
Gelman andRubin, 1992; Geweke,
1992;
Raftery
andLewis,
1992)
have beenand Lewis
(1992)
proposed
a method based on 2-state Markov chains to calculate the number of iterations needed to estimateposterior quantiles.
Let
X
l
, ... , X&dquo;,
be elements of asingle
Gibbs chain oflength
m. The msampled
values,
which aregenerally
correlated,
can be used tocompute
features of theposterior
distribution. In ourmodel,
the Xs couldrepresent
samples
from theposterior
distribution of a sirevalue,
or of the sirevariance,
or of functions of these.Invoking
theergodic theorem,
forlarge
m, theposterior
mean can be estimatedby
and the variance of the
posterior
distribution can be estimatedby
(Cox
andMiller, 1965; Geyer,
1992).
Similarly,
amarginal posterior
distributionfunction,
F(Z)
=P(X
<Z),
can be estimatedby
which means that
F(Z)
is estimatedby
theempirical
distribution function. All these estimators aresubject
to Monte-Carlosampling
error, which is reducedby
prolongation
of the chain.Consider the sequence
g(X
1
),...,
g(Xm),
whereg(.)
is a suitablefunction,
eg,g(X)
= X org(X)
=l!X!Z!,
and letE(g(X)) =
p. Thelag-time
auto-covariance of the sequence is estimated as:where
is the
sample
mean for a chain oflength
m, and t is thelag.
The auto-correlation is then estimated asTm(!)/!m(0).
The variance of thesample
mean of the chain isgiven by
which will exceed
y(0)/m
ifq(t)
> 0 for allt,
as is usual. Several estimators of the variance of thesample
mean have beenproposed
(Priestley, 1981),
but we chose oneLet
F
m
(t)
=$
m(2t)
+(2t + 1),
im
f
t =0, 1, ....
The estimator can then be writtenas
where t is chosen such that it is the
largest integer
satisfying
fm (I)
>0,
i =1,
1 ...t.
The
justification
for this choice is thatr(i)
is astrictly positive,
strictly’decreasing
function of i. IfX
l
,
X
2
, ...
X&dquo;,
areindependent,
thenVar(í
1m
)
=q(0) /m.
To obtainan indication of the effect of the correlation on
Var(í
1m
),
an ’effective number’ ofindependent
observations can be assessedas j
m
=5
m
(0)/ Vâr(í
1m
)
’
When theelements of the Gibbs chain are
independent, !m
= m.RESULTS
Estimates of the
empirical
distribution function of variousparameters
of the model for each of the 2 chains of the Gibbssampler
are shown infigures
1-5. Forunder 0.10.
Similarly, figure
3 indicates that there is90%
posterior
probability
thatheritability
in theunderlying
scale(h
2
=4J£ / (J£ + Jfl ) )
lies between 0.24 and0.49,
and the median of theposterior
distribution is 0.35.Although
this distribution isslightly skewed,
the estimate of the median agrees well with the MLtype
estimate ofheritability
of0.35, reported
in Andersen et al(1988).
Figure
4depicts
estimatesof distribution functions for the mean
(
J1
)
and for each of 3 sire effects(a
l
,
a2,a
3
).
Figure
5gives
corresponding
distributions for 2 thresholdparameters
(t
2
,
andt
3
).
Thefigures
fall in 3categories.
The distribution functions obtained from chains 1and 2 coincide for each of the variables
Q
a,
h2,
al, a2 and a3, where the sire effects al
, a2 and a3
pertain
to 3 males with31,
5 and 158offspring, respectively.
A small deviation between chains 1 and 2 is observed forJfl
and
J1, and alarger
deviation is observed for the thresholdparameters
(fig 5).
The Gibbs sequence for the thresholdparameters
showed very slowmixing properties.
Forexample,
for threshold2,
the autocorrelations betweensampled
values were0.785,
0.663 and0.315,
forlags
between5,
10 and 50samples, respectively.
The reason is that thesampled
value for agiven
threshold is boundedby
the values of theneighbouring underlying
variables U. If these are veryclose,
the value of the threshold insubsequent
samples
islikely
tochange
veryslowly.
Under theparameterization
where 1 of the thresholds is substitutedby
the residualvariance,
the autocorrelations associated with thelags above,
betweensamples
from themarginal
posterior
distribution ofe 2,
were0.078,
0.064 and0.032,
respectively.
Another scheme that may acceleratemixing
is tosample
jointly
from the threshold and theliability.
For sireeffects, lag
autocorrelations were close to zero.
A
comparison
betweenmarginal
posterior
means(average
of the 30 000samples)
estimated from the 2 chains is shown in table I. The difference between chains 1and 2 is in all cases within Monte-Carlo
sampling
error,which was
estimated withinchains, using
[34].
The ’effective number’ of observationsu
Jm
for the means ofmarginal
posterior
distributions is close to 30 000 foror a 2,
, la a2 and a3, but isabout 3 000 for
Q
e
and p
and between 200 and 400 fort
2
, ... , ts.
The
marginal posterior
distributions for J1,t
2
, ... , t
5
,
, al a2, a3are wellapproxi-mated
by
normal distributions. Theposterior
means and standard deviations of thesemarginal
distributions can becompared
to estimatesreported
in Andersen etal
(1988),
who used a2-stage
procedure.
The authors first estimated variancesusing
anREML-type
estimator(Harville
andMee,
1984).
Secondly,
assuming
that the estimated variances were the true ones, fixed effects and sire effects were estimated assuggested by
Gianola andFoulley
(1983).
Forexample,
for the 3 sires with158,
31 and 5offspring respectively,
the2-stage
procedure
yielded
estimates of sire effects(approximate
posterior
standarddeviations)
of 0.30(0.09),
0.17(0.17)
and -0.093(0.26),
respectively.
Thepresent
Bayesian
approach
with the Gibbssampler, yielded
estimates ofmarginal
posterior
means and standard deviations for these sires of 0.304(0.088),
0.177(0.166),
and -0.092(0.263),
respectively
(table
I).
DISCUSSION
here likelihood inference with numerical
integration
is acompeting
alternative. For this model anddata, marginal
posterior
distributions of sire effects are wellap-proximated by
normal distributions. On the otherhand,
with little information perrandom
effect,
eg, animalmodels,
thenormality
ofmarginal
posterior
distributions when variances are not known isunlikely
to hold. Astrength
of theBayesian
ap-proach
via Gibbssampling
is that inferences can be made frommarginal
posterior
distributions in smallsample problems,
withoutresorting
toasymptotic
approxi-mations.
Further,
the Gibbssampler
can accommodatemultivariately
distributed randomeffects,
such as is the case with animalmodels,
and this cannot beimple-mented with numerical
integration
techniques.
It seems
important
toinvestigate
threshold modelsfurther,
especially
with sparse data structuresconsisting
of many fixed effects with few observations per subclass.This case was studied
by
Moreno et al(manuscript
submitted forpublication)
in thebinary model,
wherethey
investigated frequency
properties
ofBayesian
point
estimators in a simulationstudy. They
showed thatimproper
uniformprior
distributions for the fixed effects lead to an average of themarginal posterior
mean of
heritability
which waslarger
than the simulated value.They
obtained betteragreement
when fixed effects wereassigned
proper normalprior
distributions. The use of ’non-informative’improper prior
distributions isdiscouraged
on severalgrounds by,
amongothers, Berger
and Bernardo(1992),
as this can lead toimproper
posterior
distributions. It seems that thedisagreement
between simulated and estimated values in Moreno et al is due to lack of information and not due toimpropriety
ofposterior
distributions.Thus,
the biaspersists,
though
smaller,
when allparameters
of the model areassigned
properprior
distributions(Moreno,
contain data
falling
intoonly
one of the 2 dichotomies. There is no information in the data(in
the Fisheriansense)
to estimate these fixed effects and the likelihoodis ill-conditioned. In the case of these sparse data
structures,
the choice of theprior
distribution for the fixed effects may well be the most criticalpart
of theproblem.
This needs to be studied further.Data
augmentation
in the Gibbssampler
led to conditionalposterior
distribu-tions which are easy tosample
from. This facilitatesprogramming.
We have notedthough,
that thresholdparameters
have very slowmixing properties,
and this isprobably
related to the dataaugmentation approach
used in thisstudy
(Liu
etal,
1994).
With ourdata,
theparameterization
in terms of the residual varianceresulted in smaller autocorrelations between
samples
ofQ
e
than betweensamples
of the thresholds. A scheme that islikely
to acceleratemixing
is tosample
jointly
from the threshold and
liability.
Thisstep
may necessitate other Monte-Carlosam-pling techniques
such as aMetropolis
algorithm
(Tanner, 1993),
sincesampling
is from a non-standard distribution. Alternativecomputational strategies
and param-eterizations of the model may be more critical with animal models.Here,
thereis
typically
little information on additivegenetic
effects,
and these are correlated. Theseproperties
slow down convergence of the Gibbs chain.The methods described in this paper can be
adapted easily
to draw inferences aboutgenetic
change
when selection is forcategorical
data. Sorensen et al(1994)
described how to make inferences about response to selection in the context of the Gaussian model. In the thresholdmodel,
theonly
difference is that observed data arereplaced by
the unobservedunderlying
variable U(liability).
In order to make inferences about response toselection,
theparameterization
must be in terms of an animal model.REFERENCES
Albert
JH,
Chib S(1993)
Bayesian analysis
ofbinary
andpolychotomous
response data.J Am Stat Assoc 88, 669-679
Andersen
S,
AndresenE,
Christensen K(1988)
Hip
dysplasia
selection indexexemplified
by
data from Germanshepherd dogs.
J Arcim Breed Genet 105, 112-119Anderson
DA,
Aitkin M(1985)
Variance component models withbinary
response: inter-viewervariability.
J R Stat Soc B47,
203-210Berger JO,
Bernardo JM(1992)
On thedevelopment
of reference priors. In:Bayesian
StatisticsIV (JM
Bernardo,
JOBerger,
APDawid,
AFMSmith,
eds),
OxfordUniver-sity Press,
UK, 35-60Besag J,
Green PJ(1993)
Spatial
statistics andBayesian computation.
J R Stat Soc B55, 25-37
Bliss CI
(1935)
The calculation of thedosage-mortality
curve. AnnAppl
Biol22,
134-167 Bulmer MG(1971)
The effect of selection ongenetic variability.
Am Nat 105, 201-211Cox DR, Miller HD
(1965)
TheTheory of
Stochastic Processes.Chapman
andHall,
London,
UKCurnow
R,
Smith C(1975)
Multifactorial models for familial diseases in man. J R Statist Soc A138,
131-169Dempster ER, ,Lerner
IM(1950)
Heritability
of threshold characters. Genetics 35, 212-236Falconer DS
(1965)
The inheritance ofliability
to certaindiseases,
estimated from the incidence among relatives. Ann Hum Genet29, 51-76
Federation
Cynologique
Internationale(1983)
Hip dysplasia-international
certificate and evaluation ofradiographs
(W
Brass, S Paatsama,eds),
Helsinki,
FinlandFoulley
JL(1993)
Prediction of selection response for threshold dichotomous traits.Genetics 132, 1187-1194
Foulley JL,
Manfredi E(1991)
Approches statistiques
de 1’evaluationg6n6tique
desreproducteurs
pour des caract6res binaires 4 seuils. Genet Sel EvoL23,
309-338Foulley JL,
ImS,
GianolaD,
H6schele I(1987)
Empirical
Bayes
estimation of parametersfor n
polygenic binary
traits. Genet SelEvol 19,
197-224Foulley
JL, GianolaD,
Im S(1990)
Genetic evaluation for discretepolygenic
traits in animalbreeding.
In: Advances in Statistical Methodsfor
GeneticImprovement
of
Livestock
(D
Gianola,
HammondK,
eds)
Springer-Verlag, NY, USA,
361-409Gelfand
AE,
HillsSE,
Racine-PoonA,
Smith AFM(1990)
Illustration ofBayesian
inference in normal data modelsusing
Gibbssampling.
J Am Stat Assoc85,
972-985Gelfand
A,
SmithAFM,
Lee TM(1992)
Bayesian analysis
of constrained parameter and truncated dataproblems using
Gibbssampling.
J Am Stat Assoc87,
523-432German
S,
Geman D(1984)
Stochastic relaxation, Gibbs distributions andBayesian
restoration of
images.
IEEE Trans Pattern Anal MachineIntelligence
6, 721-741 GelmanA,
Rubin DB(1992)
Asingle
series from the Gibbssampler provides
a false senseof
security.
In:Bayesian
Statistics IV(JM
Bernardo,
JOBerger,
APDawid,
AFMSmith,
eds),
OxfordUniversity
Press,UK,
625-631Geweke J
(1992)
Evaluating
the accurracy ofsampling-based approaches
to the calculationof
posterior
moments. In:Bayesian
StatisticsIV (JM
Bernardo,
JOBerger,
APDawid,
AFMSmith,
eds),
OxfordUniversity Press, UK,
169-193Geyer
CJ(1992)
Practical Markov chain Monte Carlo(with discussion).
Stat Sci7, 467-511
lGianola D
(1982)
Theory
andanalysis
of threshold characters. J Anim Sci54,
1079-1096Gianola
D,
Foulley
JL(1983)
Sire evaluation for orderedcategorical
data with a thresholdmodel. Genet Sel
Evol 15,
201-224Harville
DA,
Mee RW(1984)
A mixed-modelprocedure
foranalyzing
orderedcategorical
data. Biometrics40,
393-408H6schele I, Gianola
D, Foulley
JL(1987)
Estimation of variance components withquasi-continuous data
using Bayesian
methods. J Anim Breed Genet 104, 334-349Jensen
J, Wang CS,
SorensenDA,
Gianola D(1994)
Marginal
inferences of variance andcovariance components for traits influenced
by
maternal and directgenetic
effectsusing
the Gibbssampler.
ActaAgric
Scand44,
193-201Liu
JS, Wong WH, Kong
A(1994)
Covariance structure of the Gibbssampler
withapplications
to the comparisons of estimators andaugmentation
schemes. Biometrika81,
27-40Mood
AM, Graybill FA,
Boes DC(1974)
Introduction to theTheory of
Statistics.McGraw-Hill, NY,
USAMorton
NE,
McLean CJ(1974)
Analysis
offamily
resemblance. III.Complex segregation
of
quantitative
traits. Am J Hum Genet26,
489-503Priestley
MB(1981)
Spectral Analysis
and Time Series. AcademicPress, London,
UKQuaas RL,
Pollak EJ(1980)
Mixed-modelmethodology
for farm and ranch beef cattletesting
programs. J Anim Sci51,
1277-1287Raftery AE,
Lewis SM(1992)
How many iterations in the Gibbssampler?
In:Bayesian
StatisticsIV (JM
Bernardo,
JOBerger,
APDawid,
AFMSmith,
eds),
OxfordUniver-sity Press, UK,
763-773Roberts GO
(1992)
Convergence diagnotics
of the Gibbssampler.
In:Bayesian
Statistics IV(JM
Bernardo,
JOBerger,
APDawid,
AFMSmith,
eds),
OxfordUniversity
Press,775-782 <
Roberts
GO,
Polson NG(1994)
On thegeometric
convergence of the Gibbssampler.
J RStatist Soc B 56, 377-384
Robertson
A,
Lerner IM(1949)
Theheritability
of all-or-none traits:viability
ofpoultry.
Genetics
34,
395-411 1Smith
AFM,
Roberts GO(1993)
Bayesian computation
via the Gibbssampler
and related Markov chain Monte-Carlo methods. J R Stat Soc B 55, 3-23Sorensen
DA, Wang CS,
JensenJ,
Gianola D(1994)
Bayesian analysis
ofgenetic change
due to selectionusing
Gibbssampling.
Genet Sel Evol26,
333-360Tanner MA
(1993)
Toolsfor
StatisticalInference. Springer-Verlag,
NY, USATanner
MA, Wong
WH(1987)
The calculation ofposterior
distributionsby
dataaugmen-tation. J Am Stat Assoc 82, 528-540
! Wang
CS, Rutledge JJ,
Gianola D(1993)
Marginal
inferences about variance componentsin a mixed linear model
using
Gibbssampling.
Genet Sel Evol 21, 41-62Wang CS, Rutledge
JJ, Gianola D(1994a)
Bayesian analysis
of mixed linear models viaGibbs
sampling
with anapplication
to litter size in Iberianpigs.
Genet Sel Evol 26,91-115
Wang CS,
GianolaD,
SorensenDA,
JensenJ,
ChristensenA, Rutledge
JJ(1994b)
Response
to selection for litter size in Danish Landracepigs:
aBayesian analysis.
TheorAppl
Genet88,
220-230Wright
S(1934)
Ananalysis
ofvariability
in number ofdigits
in an inbred strain ofguinea
pigs.
Genetics 19, 506-536Zeger SL,
Karim MR(1991)
Generalized linear models with random effects: a Gibbssampling approach.
J Am Stat Assoc 86, 79-86APPENDIX
Here it is shown that there is a one-to-one
relationship
between 2 chosen param-eterizations of the thresholdmodel,
and thatthey
lead to the sameprobability
distribution.As a
starting point
assume that the conditional distribution ofliability
is:where e =
(b
l
, ... , bp,
a l
, ...
aq),
b
l
is anintercept
common to allobservations,
and associated with Ccategories
there are thresholdsti,
i =0, 1, ... , C,
withto
= -oo andt
c
= oo. We assume that the p fixed effects are estimable. In order to make theparameters
identifiable,
in what we call the standardparameterization,
we define:Note that
by
setting
f
l
=0,
thent
l
= 0. In terms of thisparameterization,
theThere are p + q + C — 1 identifiable unknown
parameters
which are:In the alternative
parameterization,
one can defineUi
=a
i
Ji,
such that:where 8
=a,6,a2
=a2U2,g,
=l
,
l
i
andt
l
=f
l
.
The number of identifiableparameters
is of course the same as before. In this paper we chose to settc_
1
=f
2
,
where
f
2
>f
l
is anarbitrary
known constant.The