Geoff Cumming, Fiona Fidler, and colleagues at La Trobe University in Melbourne, Australia, examined whether guidelines set forth in 1999 by the APA Task Force on Statistical Inference (TSFI) had been implemented in psychological research.
The authors analyzed articles published in 10 leading international psychology journals from 1998 to 2006, focusing on three practices central to the statistical reform debate: null hypothesis significance testing, confidence intervals, and figures with error bars. The results demonstrate that psychologists still rely on traditional null hypothesis significance testing but are also using considerably more graphs with error bars to report their research.
“For more than 50 years, statistical significance testing has been psychologists’ main statistical method” Fidler explains, “but there’s evidence it’s widely misunderstood and leads to dreadful research decisions.”
According to the authors, the shift toward using graphs with error bars signals a step forward in data interpretation. “Error bars,” says Fidler, “can give a clear impression of a study’s precision and lead to better conclusions.”
Many academic psychologists and journal editors agree. Results from a survey sent by the authors indicated that statistical reform is necessary in the field. But reform was not regarded as a priority, with a few editors noting resistance to change from some authors.
Cumming and Fidler insist that changes in statistical practices in research are needed if researchers and readers are to have a more accurate understanding of experimental results. They strongly recommend that scientific psychology “change its emphasis from the dichotomous decision making of null hypothesis significance testing to estimation of effect sizes.” Therefore, “researchers need further detailed guidance, examples of good practice, and editorial or institutional leadership.”