This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

peer-reviewed publication

trusted source

written by researcher(s)

proofread

How small differences in data analysis make huge differences in results

How small differences in data analysis make huge differences in results
Violin plot of Box-Cox transformed deviation from meta-analytic mean as a function of categorical peer rating. Gray points for each rating group denote model-estimated marginal mean deviation, and error bars denote 95% CI of the estimate. A Blue tit dataset, B Eucalyptus dataset. Credit: BMC Biology (2023). DOI: 10.32942/X2GG62

Over the past 20 years or so, there has been growing concern that many results published in scientific journals can't be reproduced.

Depending on the field of research, studies have found efforts to redo published studies lead to different results in between 23% and 89% of cases.

To understand how different researchers might arrive at different results, we asked hundreds of ecologists and evolutionary biologists to answer two questions by analyzing given sets of data. They arrived at a huge range of answers.

Our study has been accepted by BMC Biology as a stage 1 registered report and is currently available as a preprint ahead of peer review for stage 2.

Why is reproducibility a problem?

The causes of problems with reproducibility are common across science. They include an over-reliance on simplistic measures of "statistical significance" rather than nuanced evaluations, the fact journals prefer to publish "exciting" findings, and that make articles more exciting at the expense of transparency and increase the rate of false results in the literature.

Much of the research on reproducibility and ways it can be improved (such as "open science" initiatives) has been slow to spread between different fields of science.

Interest in these ideas has been growing among ecologists, but so far there has been little research evaluating replicability in ecology. One reason for this is the difficulty of disentangling environmental differences from the influence of researchers' choices.

One way to get at the replicability of ecological research, separate from , is to focus on what happens after the data is collected.

Birds and siblings, grass and seedlings

We were inspired by work led by Raphael Silberzahn which asked to analyze a dataset to determine whether soccer players' skin tone predicted the number of red cards they received. The study found a wide range of results.

We emulated this approach in ecology and with an open call to help us answer two research questions:

  • "To what extent is the growth of nestling (Cyanistes caeruleus) influenced by competition with siblings?"

  • "How does grass cover influence Eucalyptus spp. seedling recruitment?" ("Eucalyptus spp. seedling recruitment" means how many seedlings of trees from the genus Eucalyptus there are.)

Two hundred and forty-six ecologists and answered our call. Some worked alone and some in teams, producing 137 written descriptions of their overall answer to the research questions (alongside numeric results). These answers varied substantially for both datasets.

Looking at the effect of grass cover on the number of Eucalyptus seedlings, we had 63 responses. Eighteen described a negative effect (more grass means fewer seedlings), 31 described no effect, six teams described a positive effect (more grass means more seedlings), and eight described a mixed effect (some analyses found positive effects and some found negative effects).

For the effect of sibling competition on blue tit growth, we had 74 responses. Sixty-four teams described a negative effect (more competition means slower growth, though only 37 of these teams thought this negative effect was conclusive), five described no effect, and five described a mixed effect.

What the results mean

Perhaps unsurprisingly, we and our co-authors had a range of views on how these results should be interpreted.

We have asked three of our co-authors to comment on what struck them most.

Peter Vesk, who was the source of the Eucalyptus data, said, "Looking at the mean of all the analyses, it makes sense. Grass has essentially a negligible effect on [the number of] eucalypt tree seedlings, compared to the distance from the nearest mother tree. But the range of estimated effects is gobsmacking. It fits with my own experience that lots of small differences in the analysis workflow can add to large variation [in results]."

Simon Griffith collected the blue tit data more than 20 years ago, and it was not previously analyzed due to the complexity of decisions about the right analytical pathway. He said,

"This study demonstrates that there isn't one answer from any set of data. There are a wide range of different outcomes and understanding the underlying biology needs to account for that diversity."

Meta-researcher Fiona Fidler, who studies research itself, said, "The point of these studies isn't to scare people or to create a crisis. It is to help build our understanding of heterogeneity and what it means for the practice of science. Through metaresearch projects like this we can develop better intuitions about uncertainty and make better calibrated conclusions from our research."

What should we do about it?

In our view, the results suggest three courses of action for researchers, publishers, funders and the broader science community.

First, we should avoid treating published research as fact. A single scientific article is just one piece of evidence, existing in a broader context of limitations and biases.

The push for "novel" science means studying something that has already been investigated is discouraged, and consequently we inflate the value of individual studies. We need to take a step back and consider each article in context, rather than treating them as the final word on the matter.

Second, we should conduct more analyses per article and report all of them. If research depends on what analytic choices are made, it makes sense to present multiple analyses to build a fuller picture of the result.

And third, each study should include a description of how the results depend on data analysis decision. Research publications tend to focus on discussing the ecological implications of their findings, but they should also talk about how different analysis choices influenced the results, and what that means for interpreting the findings.

More information: Elliot Gould et al, Same data, different analysts: variation in effect sizes due to analytical decisions in ecology and evolutionary biology, BMC Biology (2023). DOI: 10.32942/X2GG62

Journal information: BMC Biology

Provided by The Conversation

This article is republished from The Conversation under a Creative Commons license. Read the original article.The Conversation

Citation: How small differences in data analysis make huge differences in results (2023, October 30) retrieved 27 April 2024 from https://phys.org/news/2023-10-small-differences-analysis-huge-results.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Does competition make us less moral? New research says yes, but only a little bit

1 shares

Feedback to editors