Researchers find new way to amplify trustworthy news content on social media without shielding bias

Researchers find new way to amplify trustworthy news content on social media without shielding bias
Average audience partisanship versus variance (N = 11,793 domains). Audience size as number of individual visitors. Domains for which there are NewsGuard reliability scores 43 are shaded in blue (where darker shades equal lower scores). Domains with no available score are plotted in gray. Credit: University of South Florida

Social media sites continue to amplify misinformation and conspiracy theories. To address this concern, an interdisciplinary team of computer scientists, physicists and social scientists led by the University of South Florida (USF) has found a solution to ensure social media users are exposed to more reliable news sources.

In their study published in the journal Nature Human Behaviour, the researchers focused on the recommendation algorithm that is used by social media platforms to prioritize content displayed to users. Rather than measuring engagement based on the number of users and pageviews, the researchers looked at what content gets amplified on a newsfeed, focusing on a source's reliability score and the political diversity of their audience.

"Low-quality content is engaging because it conforms to what we already know and like, regardless of whether it is accurate or not," said Giovanni Luca Ciampaglia, assistant professor of computer science and engineering at USF. "As a result, misinformation and conspiracy theories often go viral within like-minded audiences. The algorithm ends up picking the wrong signal and keeps promoting it further. To break this cycle, one should look for content that is engaging, but for a diverse audience, not for a like-minded one."

In collaboration with researchers at Indiana University and Dartmouth College, the team created a new algorithm using data on the web traffic and self-reported partisanship of 6,890 individuals who reflect the diversity of the United States in sex, race and political affiliation. The data was provided by online polling company YouGov. They also reviewed the reliability scores of 3,765 news sources based on the NewGuard Reliability Index, which rates on several journalistic criteria, such as editorial responsibility, accountability and financial transparency.

They found that incorporating the partisan diversity of a news audience can increase the reliability of recommended sources while still providing users with relevant recommendations. Since the algorithm isn't exclusively based on engagement or popularity, it is still able to promote reliable sources, regardless of their partisanship.

"This is especially welcome news for social media platforms, especially since they have been reluctant of introducing changes to their algorithms for fear of criticism about partisan bias," said co-author Filippo Menczer, distinguished Luddy professor of informatics and computer science at Indiana University.

Researchers say that platforms would easily be able to include audience diversity into their own recommendation algorithms because diversity measures can be derived from engagement data, and platforms already log this type of data whenever users click "like" or share something on a newsfeed. Ciampaglia and his colleagues propose adopt this new strategy in order to help prevent the spread of misinformation.

More information: Giovanni Ciampaglia, Political audience diversity and news reliability in algorithmic ranking, Nature Human Behaviour (2022). DOI: 10.1038/s41562-021-01276-5.

Journal information: Nature Human Behaviour

Citation: Researchers find new way to amplify trustworthy news content on social media without shielding bias (2022, February 3) retrieved 13 April 2024 from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Political bias on social media emerges from users, not platform


Feedback to editors