Gelman and Rubin (1992) proposed a general approach to monitoring
convergence of MCMC output in which \(m > 1\) parallel chains are
updated with initial values that are overdispersed relative to each
target distribution, which must be normally distributed. Convergence
is diagnosed when the chains have `forgotten' their initial values,
and the output from all chains is indistinguishable. The
Gelman.Diagnostic
function makes a comparison of within-chain
and between-chain variances, and is similar to a classical analysis of
variance. A large deviation between these two variances indicates
non-convergence.
This diagnostic is popular as a stopping rule, though it requires
parallel chains. The LaplacesDemon.hpc
function is an
extension of LaplacesDemon
to enable parallel chains.
As an alternative, the popular single-chain stopping rule is based on
MCSE
.
Gelman.Diagnostic(x, confidence=0.95, transform=FALSE)
This required argument accepts an object of class
demonoid.hpc
, or a list of multiple objects of class
demonoid
, where the number of components in the list
is the number of chains.
This is the coverage probability of the confidence interval for the potential scale reduction factor (PSRF).
Logical. If TRUE
, then marginal posterior
distributions in x
may be transformed to improve the
normality of the distribution, which is assumed. A log-transform is
applied to marginal posterior distributions in the interval \((0,
\infty]\), or a logit-transform is applied to marginal
posterior distributions in the interval \((0,1)\).
A list is returned with the following components:
This is a list containing the point-estimates of the
potential scale reduction factor (labelled Point Est.
) and
the associated upper confidence limits (labelled Upper C.I.
).
This is the point-estimate of the multivariate potential scale reduction factor.
To use the Gelman.Diagnostic
function, the user must first have
multiple MCMC chains for the same model, and three chains is usually
sufficient. The easiest way to obtain multiple chains is with the
LaplacesDemon.hpc
function.
Although the LaplacesDemon
function does not
simultaneously update multiple MCMC chains, it is easy enough to
obtain multiple chains, and if the computer has multiple processors
(which is common), then multiple chains may be obtained simultaneously
as follows. The model file may be opened in separate, concurrent R
sessions, and it is recommended that a maximum number of sessions is
equal to the number of processors, minus one. Each session constitutes
its own chain, and the code is identical, except the initial values
should be randomized with the GIV
function so the chains
begin in different places. The resulting object of class
demonoid
for each chain is saved, all objects are read into one
session, put into a list, and passed to the Gelman.Diagnostic
function.
Initial values must be overdispersed with respect to each target
distribution, though these distributions are unknown in the
beginning. Since the Gelman.Diagnostic
function relies heavily
on overdispersion with respect to the target distribution, the user
should consider using MCMC twice, first to estimate the target
distributions, and secondly to overdisperse initial values with
respect to them. This may help identify multimodal target
distributions. If multiple modes are found, it remain possible that
more modes exist. When multiple modes are found, and if chains are
combined with the Combine
function, each mode is
probably not represented in a proportion correct to the distribution.
The `potential scale reduction factor' (PSRF) is an estimated factor
by which the scale of the current distribution for the target
distribution might be reduced if the simulations were continued for
an infinite number of iterations. Each PSRF declines to 1 as the
number of iterations approaches infinity. PSRF is also often
represented as R-hat. PSRF is calculated for each marginal posterior
distribution in x
, together with upper and lower confidence
limits. Approximate convergence is diagnosed when the upper limit is
close to 1. The recommended proximity of each PSRF to 1 varies with
each problem, but a general goal is to achieve PSRF < 1.1. PSRF is an
estimate of how much narrower the posterior might become with an
infinite number of iterations. When PSRF = 1.1, for example, it may be
interpreted as a potential reduction of 10% in posterior interval
width, given infinite iterations. The multivariate form bounds above
the potential scale reduction factor for any linear combination of the
(possibly transformed) variables.
The confidence limits are based on the assumption that the
target distribution is stationary and normally distributed. The
transform
argument may be used to improve the normal
approximation.
A large PSRF indicates that the between-chain variance is
substantially greater than the within-chain variance, so that longer
simulation is needed. If a PSRF is close to 1, then the associated
chains are likely to have converged to one target distribution. A
large PSRF (perhaps generally when a PSRF > 1.2) indicates convergence
failure, and can indicate the presence of a multimodal marginal
posterior distribution in which different chains may have converged
to different local modes (see is.multimodal
), or the
need to update the associated chains longer, because burn-in (see
burnin
) has yet to be completed.
The Gelman.Diagnostic
is essentially the same as the
gelman.diag
function in the coda
package, but here it is
programmed to work with objects of class demonoid
.
There are two ways to estimate the variance of the stationary distribution: the mean of the empirical variance within each chain, \(W\), and the empirical variance from all chains combined, which can be expressed as
$$ \widehat{\sigma}^2 = \frac{(n-1) W}{n} + \frac{B}{n}$$
where \(n\) is the number of iterations and \(B/n\) is the empirical between-chain variance.
If the chains have converged, then both estimates are unbiased. Otherwise the first method will underestimate the variance, since the individual chains have not had time to range all over the stationary distribution, and the second method will overestimate the variance, since the initial values were chosen to be overdispersed (and this assumes the target distribution is known, see above).
This convergence diagnostic is based on the assumption that each
target distribution is normal. A Bayesian probability interval (see
p.interval
) can be constructed using a t-distribution
with mean
$$\widehat{\mu}=\mbox{Sample mean of all chains combined,}$$
variance
$$\widehat{V} = \widehat{\sigma}^2 + \frac{B}{mn},$$
and degrees of freedom estimated by the method of moments
$$d = \frac{2\widehat{V}^2}{\mbox{Var}(\widehat{V})}$$
Use of the t-distribution accounts for the fact that the mean and variance of the posterior distribution are estimated. The convergence diagnostic itself is
$$R=\sqrt{\frac{(d+3) \widehat{V}}{(d+1)W}}$$
Values substantially above 1 indicate lack of convergence. If the chains have not converged, then Bayesian probability intervals based on the t-distribution are too wide, and have the potential to shrink by this factor if the MCMC run is continued.
The multivariate version of Gelman and Rubin's diagnostic was proposed by Brooks and Gelman (1998). Unlike the univariate proportional scale reduction factor, the multivariate version does not include an adjustment for the estimated number of degrees of freedom.
Brooks, S.P. and Gelman, A. (1998). "General Methods for Monitoring Convergence of Iterative Simulations". Journal of Computational and Graphical Statistics, 7, p. 434--455.
Gelman, A. and Rubin, D.B. (1992). "Inference from Iterative Simulation using Multiple Sequences". Statistical Science, 7, p. 457--511.
Combine
,
GIV
,
is.multimodal
,
LaplacesDemon
,
LaplacesDemon.hpc
,
MCSE
, and
p.interval
.
# NOT RUN {
#library(LaplacesDemon)
###After updating multiple chains with LaplacesDemon.hpc, do:
#Gelman.Diagnostic(Fit)
# }
Run the code above in your browser using DataLab