Munich Personal RePEc Archive

Significance Tests Harm Progress in Forecasting

Armstrong, J. Scott (2007): Significance Tests Harm Progress in Forecasting. Published in: International Journal of Forecasting No. 23 (2007): pp. 321-336.

[img]
Preview
PDF
MPRA_paper_81664.pdf

Download (1MB) | Preview

Abstract

Based on a summary of prior literature, I conclude that tests of statistical significance harm scientific progress. Efforts to find exceptions to this conclusion have, to date, turned up none. Even when done correctly, significance tests are dangerous. I show that summaries of scientific research do not require tests of statistical significance. I illustrate the dangers of significance tests by examining an application to the M3-Competition. Although the authors of that reanalysis conducted a proper series of statistical tests, they suggest that the original M3 was not justified in concluding that combined forecasts reduce errors and that the selection of the best method is dependent upon the selection of a proper error measure. I show that the original conclusions were justified and that they are correct. Authors should try to avoid tests of statistical significance, journals should discourage them, and readers should ignore them. Instead, to analyze and communicate findings from empirical studies, one should use effect sizes, confidence intervals, replications/extensions, and meta-analyses.

UB_LMU-Logo
MPRA is a RePEc service hosted by
the Munich University Library in Germany.