Statistics Papers

Document Type

Journal Article

Date of this Version

2009

Publication Source

Journal of the American Statistical Association

Volume

104

Issue

487

Start Page

1090

Last Page

1101

DOI

10.1198/jasa.2009.tm08338

Abstract

An observational or nonrandomized study of treatment effects may be biased by failure to control for some relevant covariate that was not measured. The design of an observational study is known to strongly affect its sensitivity to biases from covariates that were not observed. For instance, the choice of an outcome to study, or the decision to combine several outcomes in a test for coherence, can materially affect the sensitivity to unobserved biases. Decisions that shape the design are, therefore, critically important, but they are also difficult decisions to make in the absence of data. We consider the possibility of randomly splitting the data from an observational study into a smaller planning sample and a larger analysis sample, where the planning sample is used to guide decisions about design. After reviewing the concept of design sensitivity, we evaluate sample splitting in theory, by numerical computation, and by simulation, comparing it to several methods that use all of the data. Sample splitting is remarkably effective, much more so in observational studies than in randomized experiments: splitting 1,000 matched pairs into 100 planning pairs and 900 analysis pairs often materially improves the design sensitivity. An example from genetic toxicology is used to illustrate the method.

Copyright/Permission Statement

This is an Accepted Manuscript of an article published by Taylor & Francis in Journal of the American Statistical Association on 01 Jan 2012, available online: http://wwww.tandfonline.com/10.1198/jasa.2009.tm08338.

Keywords

coherence, multiple comparisons, permutation test, sensitivity analysis

Share

COinS
 

Date Posted: 27 November 2017