TY - JOUR
T1 - Markov chain monte carlo methods for computing bayes factors
T2 - A comparative review
AU - Han, Cong
AU - Carlin, Brad
PY - 2001/9/1
Y1 - 2001/9/1
N2 - The problem of calculating posterior probabilities for a collection of competing models and associated Bayes factors continues to be a formidable challenge for applied Bayesian statisticians. Current approaches that take advantage of modern Markov chain Monte Carlo computing methods include those that attempt to sample over some form of the joint space created by the model indicators and the parameters for each model, others that sample over the model space alone, and still others that attempt to estimate the marginal likelihood of each model directly (because the collection of these is equivalent to the collection of model probabilities themselves). We review several methods and compare them in the context of three examples: a simple regression example, a more challenging hierarchical longitudinal model, and a binary data latent variable model. We find that the joint model-parameter space search methods perform adequately but can be difficult to program and tune, whereas the marginal likelihood methods often are less troublesome and require less additional coding. Our results suggest that the latter methods may be most appropriate for practitioners working in many standard model choice settings, but the former remain important for comparing models of varying dimension (e.g., multiple changepoint models) or models whose parameters cannot easily be updated in relatively few blocks. We caution, however, that all methods we compared require significant human and computer effort, and this suggests that less formal Bayesian model choice methods may offer a more realistic alternative in many cases.
AB - The problem of calculating posterior probabilities for a collection of competing models and associated Bayes factors continues to be a formidable challenge for applied Bayesian statisticians. Current approaches that take advantage of modern Markov chain Monte Carlo computing methods include those that attempt to sample over some form of the joint space created by the model indicators and the parameters for each model, others that sample over the model space alone, and still others that attempt to estimate the marginal likelihood of each model directly (because the collection of these is equivalent to the collection of model probabilities themselves). We review several methods and compare them in the context of three examples: a simple regression example, a more challenging hierarchical longitudinal model, and a binary data latent variable model. We find that the joint model-parameter space search methods perform adequately but can be difficult to program and tune, whereas the marginal likelihood methods often are less troublesome and require less additional coding. Our results suggest that the latter methods may be most appropriate for practitioners working in many standard model choice settings, but the former remain important for comparing models of varying dimension (e.g., multiple changepoint models) or models whose parameters cannot easily be updated in relatively few blocks. We caution, however, that all methods we compared require significant human and computer effort, and this suggests that less formal Bayesian model choice methods may offer a more realistic alternative in many cases.
KW - Bayesian model choice
KW - Gibbs sampler
KW - Marginal likelihood
KW - Metropolis–hastings algorithm
KW - Reversible jump sampler
UR - http://www.scopus.com/inward/record.url?scp=0442312140&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=0442312140&partnerID=8YFLogxK
U2 - 10.1198/016214501753208780
DO - 10.1198/016214501753208780
M3 - Article
AN - SCOPUS:0442312140
SN - 0162-1459
VL - 96
SP - 1122
EP - 1132
JO - Journal of the American Statistical Association
JF - Journal of the American Statistical Association
IS - 455
ER -