Date of this Version
Journal of the American Statistical Association
An observational or nonrandomized study of treatment effects may be biased by failure to control for some relevant covariate that was not measured. The design of an observational study is known to strongly affect its sensitivity to biases from covariates that were not observed. For instance, the choice of an outcome to study, or the decision to combine several outcomes in a test for coherence, can materially affect the sensitivity to unobserved biases. Decisions that shape the design are, therefore, critically important, but they are also difficult decisions to make in the absence of data. We consider the possibility of randomly splitting the data from an observational study into a smaller planning sample and a larger analysis sample, where the planning sample is used to guide decisions about design. After reviewing the concept of design sensitivity, we evaluate sample splitting in theory, by numerical computation, and by simulation, comparing it to several methods that use all of the data. Sample splitting is remarkably effective, much more so in observational studies than in randomized experiments: splitting 1,000 matched pairs into 100 planning pairs and 900 analysis pairs often materially improves the design sensitivity. An example from genetic toxicology is used to illustrate the method.
This is an Accepted Manuscript of an article published by Taylor & Francis in Journal of the American Statistical Association on 01 Jan 2012, available online: http://wwww.tandfonline.com/10.1198/jasa.2009.tm08338.
coherence, multiple comparisons, permutation test, sensitivity analysis
Heller, R., Rosenbaum, P. R., & Small, D. S. (2009). Split Samples and Design Sensitivity in Observational Studies. Journal of the American Statistical Association, 104 (487), 1090-1101. http://dx.doi.org/10.1198/jasa.2009.tm08338
Date Posted: 27 November 2017