Improved estimation procedures for multilevel models with binary response:a case-study

Citation
G. Rodriguez et N. Goldman, Improved estimation procedures for multilevel models with binary response:a case-study, J ROY STA A, 164, 2001, pp. 339-355
Citations number
35
Categorie Soggetti
Economics
Journal title
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES A-STATISTICS IN SOCIETY
ISSN journal
09641998 → ACNP
Volume
164
Year of publication
2001
Part
2
Pages
339 - 355
Database
ISI
SICI code
0964-1998(2001)164:<339:IEPFMM>2.0.ZU;2-2
Abstract
During recent years, analysts have been relying on approximate methods of i nference to estimate multilevel models for binary or count data. In an earl ier study of random-intercept models for binary outcomes we used simulated data to demonstrate that one such approximation, known as marginal quasi-li kelihood, leads to a substantial attenuation bias in the estimates of both fixed and random effects whenever the random effects are non-trivial. In th is paper, we fit three-level random-intercept models to actual data for two binary outcomes, to assess whether refined approximation procedures, namel y penalized quasi-likelihood and second-order improvements to marginal and penalized quasi-likelihood, also underestimate the underlying parameters. T he extent of the bias is assessed by two standards of comparison: exact max imum likelihood estimates, based on a Gauss-Hermite numerical quadrature pr ocedure, and a set of Bayesian estimates, obtained from Gibbs sampling with diffuse priors. We also examine the effectiveness of a parametric bootstra p procedure for reducing the bias. The results indicate that second-order p enalized quasi-likelihood estimates provide a considerable improvement over the other approximations, but all the methods of approximate inference res ult in a substantial underestimation of the fixed and random effects when t he random effects are sizable. We also find that the parametric bootstrap m ethod can eliminate the bias but is computationally very intensive.