Credit: Unsplash/CC0 Public Domain

Factual corrections published on Facebook's news feed can reduce a user's belief in misinformation, even across partisan lines, according to a new paper published this month in the Journal of Politics. Social media users were tested on their accuracy in recognizing misinformation through exposure to corrections on a simulated news feed that was made to look like Facebook's news feed. However, just like in the real world, people in the experiment were free to ignore the information in the feed that corrected false stories also posted on the news feed. Even when given the freedom to choose what to read in the experiment, users' accuracy improved when fact-checks were included with false stories.

The study's findings contradict previous research that suggests displaying corrections on was ineffective or could even backfire by increasing inaccuracy. Instead, even when users are not compelled to read fact-checks in a simulation of Facebook's feed, the new study found they nonetheless became more factually accurate despite exposure to misinformation. This finding was consistent for both liberal and conservative users with only some variation depending on the topic of the misinformation.

"Our study finds that posting fact-checks on Facebook increases users' accuracy when it comes to encountering misinformation," Ethan Porter, assistant professor of media and public affairs at the George Washington University and co-author of the study, said. "While other researchers have studied simulated news environments, ours is among the first that we are aware of to study corrections and misinformation on a detailed simulation of the news feed."  

Researchers administered two experiments on large, nationally representative population samples using a novel platform designed to mimic Facebook's news feed.  

Subjects in the first experiment were exposed to news feeds that contained, at random, multiple items of misinformation, factual corrections and non-political placebo content. Subjects were free to read, or avoid reading, any material that they wished. The fake stories used in the experiment were stories that had actually circulated on Facebook and included, for example, false claims relating to immigrants and measles, climate activist Greta Thunberg, and President Trump. Corrections directly contradicted the false claims and offered subjects the opportunity to access a more detailed fact-check on an external website. In the first experiment, the correction would appear above a facsimile of the original fake story, and test subjects were asked on a 1-5 scale whether they agreed with the fake story. 

The second experiment investigated whether results would change depending how the corrections were displayed in the news feed with the correction placed above the original fake story that had been blurred out. The test subjects were then asked to judge on a 1-5 scale the truthfulness of the fake .

In both versions tested, the fact-checks decreased false beliefs by larger amounts than misinformation shared without corrections. Additionally, the researchers tested the differences among participants who identified as either liberal or conservative and did not find one group to be more resistant to facts than the other.

"Our previous research has shown that individuals respond to corrections by becoming more accurate, and this study finds the same impact even when people weren't forced to read the shared fact-checks," Porter said. "Even on a platform that approximates Facebook's news feed, in which were presented with fact-checks that didn't align with their politics, sharing corrections increased their ability to recognize misinformation as such. Our results suggest that social media companies, policymakers and scholars need not resign themselves to the spread of on social media but can use corrections to rebut it."

The researchers did note limitations to their study, pointing out that while their simulation closely mimicked the look of a Facebook news , their research did not factor how social ties from friends and connections may led to different results in gaging the effects of sharing fact-checks and corrections. 

More information: Ethan Porter et al, Political Misinformation and Factual Corrections on the Facebook News Feed: Experimental Evidence, The Journal of Politics (2022). DOI: 10.1086/719271

Thomas Wood et al, The Elusive Backfire Effect: Mass Attitudes' Steadfast Factual Adherence, SSRN Electronic Journal (2017). DOI: 10.2139/ssrn.2819073

Journal information: Journal of Politics