An approach for addressing the multiple testing problem in social policy impact evaluations

Eval Rev. 2009 Dec;33(6):539-67. doi: 10.1177/0193841X09350590.

Abstract

In social policy evaluations, the multiple testing problem occurs due to the many hypothesis tests that are typically conducted across multiple outcomes and subgroups, which can lead to spurious impact findings. This article discusses a framework for addressing this problem that balances Types I and II errors. The framework involves specifying confirmatory and exploratory analyses in study protocols, delineating confirmatory outcome domains, conducting t tests on composite domain outcomes, and applying multiplicity corrections to composites across domains to obtain summative impact evidence. The article presents statistical background and discusses multiplicity issues for subgroup analyses, designs with multiple treatments, and reporting.

Publication types

  • Research Support, U.S. Gov't, Non-P.H.S.

MeSH terms

  • Analysis of Variance
  • Data Interpretation, Statistical*
  • Evaluation Studies as Topic*
  • Humans
  • Public Policy*
  • Regression Analysis
  • Research Design*
  • Selection Bias*