Marketing Papers
Document Type
Journal Article
Date of this Version
April 2007
Abstract
Based on a summary of prior literature, I conclude that tests of statistical significance harm scientific progress. Efforts to find exceptions to this conclusion have, to date, turned up none. Even when done correctly, significance tests are dangerous. I show that summaries of scientific research do not require tests of statistical significance. I illustrate the dangers of significance tests by examining an application to the M3-Competition. Although the authors of that reanalysis conducted a proper series of statistical tests, they suggest that the original M3 was not justified in concluding that combined forecasts reduce errors and that the selection of the best method is dependent upon the selection of a proper error measure. I show that the original conclusions were justified and that they are correct. Authors should try to avoid tests of statistical significance, journals should discourage them, and readers should ignore them. Instead, to analyze and communicate findings from empirical studies, one should use effect sizes, confidence intervals,replications/extensions, and meta-analyses.
Keywords
accuracy measures, combining forecasts, confidence intervals, effect size, M-competition, meta-analysis, null hypothesis, practical significance, replications
Recommended Citation
Armstrong, J. S. (2007). Significance Tests Harm Progress in Forecasting. Retrieved from https://repository.upenn.edu/marketing_papers/99
Date Posted: 15 June 2007
This document has been peer reviewed.
Comments
Postprint version. Published in International Journal of Forecasting, Volume 23, Issue 2, April 2007, pages 321-327.
Publisher URL: http://dx.doi.org/10.1016/j.ijforecast.2007.03.004