An empirical comparison of methods for meta-analysis of diagnostic accuracy showed hierarchical models are necessary

J Clin Epidemiol. 2008 Nov;61(11):1095-103. doi: 10.1016/j.jclinepi.2007.09.013.

Abstract

Objective: Meta-analysis of studies of the accuracy of diagnostic tests currently uses a variety of methods. Statistically rigorous hierarchical models require expertise and sophisticated software. We assessed whether any of the simpler methods can in practice give adequately accurate and reliable results.

Study design and setting: We reviewed six methods for meta-analysis of diagnostic accuracy: four simple commonly used methods (simple pooling, separate random-effects meta-analyses of sensitivity and specificity, separate meta-analyses of positive and negative likelihood ratios, and the Littenberg-Moses summary receiver operating characteristic [ROC] curve) and two more statistically rigorous approaches using hierarchical models (bivariate random-effects meta-analysis and hierarchical summary ROC curve analysis). We applied the methods to data from a sample of eight systematic reviews chosen to illustrate a variety of patterns of results.

Results: In each meta-analysis, there was substantial heterogeneity between the results of different studies. Simple pooling of results gave misleading summary estimates of sensitivity and specificity in some meta-analyses, and the Littenberg-Moses method produced summary ROC curves that diverged from those produced by more rigorous methods in some situations.

Conclusion: The closely related hierarchical summary ROC curve or bivariate models should be used as the standard method for meta-analysis of diagnostic accuracy.

Publication types

  • Comparative Study
  • Research Support, Non-U.S. Gov't
  • Review

MeSH terms

  • Data Interpretation, Statistical
  • Diagnostic Tests, Routine / standards*
  • Humans
  • Meta-Analysis as Topic*
  • Models, Statistical*
  • ROC Curve
  • Review Literature as Topic