This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:


trusted source

written by researcher(s)


Scientific fraud is rising, and automated systems won't stop it. We need research detectives

Credit: CC0 Public Domain

Fraud in science is alarmingly common. Sometimes researchers lie about results and invent data to win funding and prestige. Other times, researchers might pay to stage and publish entirely bogus studies to win an undeserved pay rise—fuelling a "paper mill" industry worth an estimated €1 billion a year.

Some of this rubbish can be easily spotted by peer reviewers, but the peer review system has become badly stretched by ever-rising paper numbers. And there's a new threat, as more sophisticated AI is able to generate plausible scientific data.

The latest idea among academic publishers is to use automated tools to screen all papers submitted to for telltale signs. However, some of these tools are easy to fool.

I am part of a group of multidisciplinary scientists working to tackle and poor practice using metascience or the "science of science". Ours is a new field, but we already have our own society and our members have worked with funders and publishers to investigate improvements to research practice.

The limits of automated screening

The problems with automated screening are highlighted by a new screening tool publicized last month. The tool suggested around one in three neuroscience papers might be fraudulent.

However, this tool detects suspected simply by flagging authors with a non-institutional email (such as and with a hospital affiliation. While this could catch some fraud, it will also flag many honest researchers, and the flagged a whopping 44% of genuine papers as potentially fake.

One big problem with simple screening tools is that fraudsters will quickly find workarounds. For instance, telling their clients to use their institutional email address to submit the paper.

Given the amount of money to be made, fraudsters have the time and motivation to find workarounds to automated screening systems.

This is not to say automated tools have no place. They have been used successfully to check papers for faulty experiments, and to hunt for pilfered text reworked to avoid plagiarism checkers.

A project launched by the International Association of Scientific, Technical and Medical Publishers which aims to use screening tools to tackle fraud is also welcome. But automated tools cannot be the only line of defense.

A crowdfunded detective

There are remarkably few people who hunt through published research to detect scientific fraud. Perhaps the best known is the Dutch microbiologist Elisabeth Bik, who is an expert at catching manipulated images in scientific papers.

Bik has single-handedly caught multiple massive fraudsters, with the dodgy papers eventually being retracted from the scientific record.

Bik's work is a tremendous public service. However, she isn't paid by a university or a scientific publisher. Her detective work—which has seen her face harassment and court cases—is crowd funded.

With the billions of dollars in the publishing world, can't a few million be found for ? In the meantime, one of our best-known lines of defense relies on good will and passion.

In Australia, spending just 0.1% of the annual scientific research budget on quality control would be A$12 million per year. This would be enough to fund a whole office of detectives and also training for researchers in good scientific practice, increasing the return on investment for the remaining 99.9% of the annual budget.

Call the fraud police

A solution—or at least a partial one—seems obvious: somebody should employ lots of people like Bik to check quality. However, "somebody should" is a dangerous phrase, because it could easily mean nobody will.

Research funders wait for scientific publishers to take action. Publishers expect universities and other institutions to do something. Those institutions in turn look to government for a solution.

Meanwhile, paper mills are happily making a mint, and the world's pool of scientific evidence is becoming increasingly contaminated by rubbish.

Quality control systems need not be expensive, as we don't need to check every paper in detail. Random spot checks might be effective.

Say one in every 300 submissions gets checked by the "fraud police". That's a small probability, but people are notoriously bad at judging small probabilities, as proved by the popularity of lotteries.

There would also need to be consequences, such as notifying all the institutions and funders involved, and an expectation of a rapid response. If an institution were involved in multiple cases, publishers could flag all papers from that institution for extra checks.

Publicity would be a good start

Of course, this could disadvantage honest researchers from that institution—but personally I would like to know if my colleagues had been submitting fraud. And given institutions rarely publicize the wrongdoing of their own staff, it may be the first I hear about it.

If honest researchers pressure their institutions to act, it would be a tremendous change. Publishers can't be the only line of defense in tackling fraud.

Funding for stronger screening systems is a great start, but we also need to spend money on people. We need to turn the arms race with the fraudsters into a brains race, because we have the better brains.

Provided by The Conversation

This article is republished from The Conversation under a Creative Commons license. Read the original article.The Conversation

Citation: Scientific fraud is rising, and automated systems won't stop it. We need research detectives (2023, June 21) retrieved 14 July 2024 from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Accountants' tricks can help identify cheating scientists, says new study


Feedback to editors