Abstract
We use simulation studies, whose design is realistic for educational and medical\nresearch (as well as other fields of inquiry), to compare Bayesian and likelihood-based\nmethods for fitting variance-components (VC) and random-effects logistic regression\n(RELR) models. The likelihood (and approximate likelihood) approaches we examine are\nbased on the methods most widely used in current applied multilevel (hierarchical)\nanalyses: maximum likelihood (ML) and restricted ML (REML) for Gaussian outcomes, and\nmarginal and penalized quasi-likelihood (MQL and PQL) for Bernoulli outcomes. Our\nBayesian methods use Markov chain Monte Carlo (MCMC) estimation, with adaptive hybrid\nMetropolis-Gibbs sampling for RELR models, and several diffuse prior distributions\n($\\Gamma^{ -1 }( \\epsilon, \\epsilon )$ and $U( 0, \\frac{ 1 }{ \\epsilon } )$ priors for\nvariance components). For evaluation criteria we consider bias of point estimates and\nnominal versus actual coverage of interval estimates in repeated sampling. In two-level\nVC models we find that (a) both likelihood-based and Bayesian approaches can be made to\nproduce approximately unbiased estimates, although the automatic manner in which REML\naccomplishes this is an advantage, but (b) both approaches had difficulty achieving\nnominal coverage in small samples and with small values of the intraclass correlation.\nWith the three-level RELR models we examine we find that (c) quasi-likelihood methods\nfor estimating random-effects variances perform badly with respect to bias and coverage\nin the example we simulated, and (d) Bayesian diffuse-prior methods lead to\nwell-calibrated point and interval RELR estimates. While it is true that the\nlikelihood-based methods we study are considerably faster computationally than MCMC, (i)\nsteady improvements in recent years in both hardware speed and efficiency of Monte Carlo\nalgorithms and (ii) the lack of calibration of likelihood-based methods in some common\nhierarchical settings combine to make MCMC-based Bayesian fitting of multilevel models\nan attractive approach, even with rather large data sets. Other analytic strategies\nbased on less approximate likelihood methods are also possible but would benefit from\nfurther study of the type summarized here.
Keywords
Affiliated Institutions
Related Publications
Improved Estimation Procedures for Multilevel Models with Binary Response: A Case-Study
Summary During recent years, analysts have been relying on approximate methods of inference to estimate multilevel models for binary or count data. In an earlier study of random...
The No-U-Turn Sampler: Adaptively Setting Path Lengths in Hamiltonian Monte Carlo
Hamiltonian Monte Carlo (HMC) is a Markov chain Monte Carlo (MCMC) algorithm that avoids the random walk behavior and sensitivity to correlated parameters that plague many MCMC ...
Bayesian Statistics: An Introduction
Bayesian Statistics is the school of thought that combines prior beliefs with the likelihood of a hypothesis to arrive at posterior beliefs. The first edition of Peter Lees book...
Markov Chain Monte Carlo Methods and the Label Switching Problem in Bayesian Mixture Modeling
In the past ten years there has been a dramatic increase of interest in the Bayesian analysis of finite mixture models. This is primarily because of the emergence of Markov chai...
Inference in Molecular Population Genetics
Summary Full likelihood-based inference for modern population genetics data presents methodological and computational challenges. The problem is of considerable practical import...
Publication Info
- Year
- 2006
- Type
- article
- Volume
- 1
- Issue
- 3
- Citations
- 607
- Access
- Closed
External Links
Social Impact
Social media, news, blog, policy document mentions
Citation Metrics
Cite This
Identifiers
- DOI
- 10.1214/06-ba117