Misinformation on social media—can technology save us?

November 28, 2016 by Filippo Menczer, The Conversation
Sharing election hashtags: Dots are Twitter accounts; lines show retweeting; larger dots are retweeted more. Red dots are likely bots; blue ones are likely humans. Credit: Clayton Davis, CC BY-ND

If you get your news from social media, as most Americans do, you are exposed to a daily dose of hoaxes, rumors, conspiracy theories and misleading news. When it's all mixed in with reliable information from honest sources, the truth can be very hard to discern.

In fact, my research team's analysis of data from Columbia University's Emergent rumor tracker suggests that this misinformation is just as likely to go viral as reliable information.

Many are asking whether this onslaught of digital misinformation affected the outcome of the 2016 U.S. election. The truth is we do not know, although there are reasons to believe it is entirely possible, based on past analysis and accounts from other countries. Each piece of misinformation contributes to the shaping of our opinions. Overall, the harm can be very real: If people can be conned into jeopardizing our children's lives, as they do when they opt out of immunizations, why not our democracy?

As a researcher on the spread of misinformation through social media, I know that limiting news fakers' ability to sell ads, as recently announced by Google and Facebook, is a step in the right direction. But it will not curb abuses driven by political motives.

Exploiting social media

About 10 years ago, my colleagues and I ran an experiment in which we learned 72 percent of college students trusted links that appeared to originate from friends – even to the point of entering personal login information on phishing sites. This widespread vulnerability suggested another form of malicious manipulation: People might also believe misinformation they receive when clicking on a link from a social contact.

To explore that idea, I created a fake web page with random, computer-generated gossip news – things like "Celebrity X caught in bed with Celebrity Y!" Visitors to the site who searched for a name would trigger the script to automatically fabricate a story about the person. I included on the site a disclaimer, saying the site contained meaningless text and made-up "facts." I also placed ads on the page. At the end of the month, I got a check in the mail with earnings from the ads. That was my proof: Fake news could make money by polluting the internet with falsehoods.

Sadly, I was not the only one with this idea. Ten years later, we have an industry of fake news and digital misinformation. Clickbait sites manufacture hoaxes to make money from ads, while so-called hyperpartisan sites publish and spread rumors and to influence public opinion.

This industry is bolstered by how easy it is to create social bots, fake accounts controlled by software that look like real people and therefore can have real influence. Research in my lab uncovered many examples of fake grassroots campaigns, also called political astroturfing.

In response, we developed the BotOrNot tool to detect social bots. It's not perfect, but accurate enough to uncover persuasion campaigns in the Brexit and antivax movements. Using BotOrNot, our colleagues found that a large portion of online chatter about the 2016 elections was generated by bots.

Creating information bubbles

We humans are vulnerable to manipulation by digital misinformation thanks to a complex set of social, cognitive, economic and algorithmic biases. Some of these have evolved for good reasons: Trusting signals from our social circles and rejecting information that contradicts our experience served us well when our species adapted to evade predators. But in today's shrinking online networks, a social network connection with a conspiracy theorist on the other side of the planet does not help inform my opinions.

Copying our friends and unfollowing those with different opinions give us echo chambers so polarized that researchers can tell with high accuracy whether you are liberal or conservative by just looking at your friends. The network structure is so dense that any misinformation spreads almost instantaneously within one group, and so segregated that it does not reach the other.

Inside our bubble, we are selectively exposed to information aligned with our beliefs. That is an ideal scenario to maximize engagement, but a detrimental one for developing healthy skepticism. Confirmation bias leads us to share a headline without even reading the article.

Our lab got a personal lesson in this when our own research project became the subject of a vicious misinformation campaign in the run-up to the 2014 U.S. midterm elections. When we investigated what was happening, we found stories about our research being predominantly shared by Twitter users within one partisan echo chamber, a large and homogeneous community of politically active users. These people were quick to retweet and impervious to debunking information.

In this graph of echo chambers in the Twittersphere, purple dots represent people spreading false claims about the Truthy research project; the two accounts that sought to debunk the false information are in orange on the far left. Credit: Giovanni Luca Ciampaglia, CC BY-ND
Viral inevitability

Our research shows that given the structure of our social networks and our limited attention, it is inevitable that some memes will go viral, irrespective of their quality. Even if individuals tend to share information of higher quality, the network as a whole is not effective at discriminating between reliable and fabricated information. This helps explain all the viral hoaxes we observe in the wild.

The attention economy takes care of the rest: If we pay attention to a certain topic, more information on that topic will be produced. It's cheaper to fabricate information and pass it off as fact than it is to report actual truth. And fabrication can be tailored to each group: Conservatives read that the pope endorsed Trump, liberals read that he endorsed Clinton. He did neither.

Beholden to algorithms

Since we cannot pay attention to all the posts in our feeds, algorithms determine what we see and what we don't. The algorithms used by today are designed to prioritize engaging posts – ones we're likely to click on, react to and share. But a recent analysis found intentionally misleading pages got at least as much online sharing and reaction as real news.

This algorithmic bias toward engagement over truth reinforces our social and cognitive biases. As a result, when we follow links shared on social media, we tend to visit a smaller, more homogeneous set of sources than when we conduct a search and visit the top results.

Existing research shows that being in an echo chamber can make people more gullible about accepting unverified rumors. But we need to know a lot more about how different people respond to a single hoax: Some share it right away, others fact-check it first.

We are simulating a social network to study this competition between sharing and fact-checking. We are hoping to help untangle conflicting evidence about when fact-checking helps stop hoaxes from spreading and when it doesn't. Our preliminary results suggest that the more segregated the community of hoax believers, the longer the hoax survives. Again, it's not just about the hoax itself but also about the network.

Many people are trying to figure out what to do about all this. According to Mark Zuckerberg's latest announcement, Facebook teams are testing potential options. And a group of college students has proposed a way to simply label shared links as "verified" or not.

Some solutions remain out of reach, at least for the moment. For example, we can't yet teach artificial intelligence systems how to discern between truth and falsehood. But we can tell ranking algorithms to give higher priority to more reliable sources.

Studying the spread of fake news

We can make our fight against fake news more efficient if we better understand how bad information spreads. If, for example, bots are responsible for many of the falsehoods, we can focus attention on detecting them. If, alternatively, the problem is with echo chambers, perhaps we could design recommendation systems that don't exclude differing views.

To that end, our lab is building a platform called Hoaxy to track and visualize the spread of unverified claims and corresponding fact-checking on social media. That will give us real-world data, with which we can inform our simulated social networks. Then we can test possible approaches to fighting fake news.

Hoaxy may also be able to show people how easy it is for their opinions to be manipulated by online information – and even how likely some of us are to share falsehoods online. Hoaxy will join a suite of tools in our Observatory on Social Media, which allows anyone to see how memes spread on Twitter. Linking tools like these to human fact-checkers and platforms could make it easier to minimize duplication of efforts and support each other.

It is imperative that we invest resources in the study of this phenomenon. We need all hands on deck: Computer scientists, social scientists, economists, journalists and industry partners must work together to stand firm against the spread of misinformation.

Explore further: Zuckerberg: 'Crazy' to say Facebook influenced election

Related Stories

3Qs: What a sham(e)—how to filter out fake news

November 24, 2016

The spread of fake online news has become a hot topic of conversation, particularly in the wake of the presidential election. According to a BuzzFeed news analysis, the top-performing fake election news stories posted on ...

Facebook in crosshairs as fake news battle heats up

November 22, 2016

Hillary Clinton was about to be indicted, Pope Francis endorsed Donald Trump: the battle over fake news is heating up after a White House campaign in which the misinformation industry may have swung the outcome of the vote.

Fake Tweets, real consequences for the election

November 4, 2016

Software robots masquerading as humans are influencing the political discourse on social media as never before and could threaten the very integrity of the 2016 U.S. presidential election, said Emilio Ferrara, a computer ...

Recommended for you

Understanding dynamic stall at high speeds

December 18, 2018

When a bird in flight lands, it performs a rapid pitch-up maneuver during the perching process to keep from overshooting the branch or telephone wire. In aerodynamics, that action produces a complex phenomenon known as dynamic ...

Pushing lithium ion batteries to the next performance level

December 13, 2018

Conventional lithium ion batteries, such as those widely used in smartphones and notebooks, have reached performance limits. Materials chemist Freddy Kleitz from the Faculty of Chemistry of the University of Vienna and international ...

Uber filed paperwork for IPO: report

December 8, 2018

Ride-share company Uber quietly filed paperwork this week for its initial public offering, the Wall Street Journal reported late Friday.


Adjust slider to filter visible comments by rank

Display comments: newest first

5 / 5 (1) Nov 30, 2016
This is an interesting topic. Politics has pushed it up, but its a thing in every field: For instance, I just spent about a day trying to figure out an extremely widely repeated refutation of the EMDrive being possible, only to find it based on a misrepresentation of what Eagleworks said. A thought:

Truth is secondary to Cultural Capital. "Cultural Capital" was an idea used in the music industry to model why some bands became so much more popular than others. It was about group identity (for youth) far more than liking the music. They bought the album to have the capital to buy into an identity, a social group. People without the knowledge of the right bands would be bankrupt in the group and not belong. In my case, a "skeptic" group identity was more important than the authenticity of the refutation.

Wrapping up that idea, maybe an economic framing can offer some insight, along with looking at the sense of isolation in our society that makes people seek to belong online.
1 / 5 (3) Dec 04, 2016
The recent election is truly the gift that keeps on giving. Liberals wring their hands trying to figure out who or what sabotaged them because clearly they have the only true and correct ideas. All the other ideas are just ignorant or racist or homophobic or xenophobic or transgender-bathroom-use-phobic or whatever the hippest term is that belittles others.

I have news for you. It may come as a shock but "fake" news has been around for a long time. It's just become a thing lately because liberals want to blame losing an election on something other than their ideas or their candidates. The natural instinct of liberals is to censor ideas that challenge theirs, so they look for solutions like "fact-checking" and tagging links as "verified" because people can't be trusted to do their own research. Misinformation has been around forever, but suddenly it seems important because liberals have run out of ideas.
1 / 5 (1) Dec 04, 2016
At least we have Wikileaks who curates the info they get before providing it to us. And they've done a good job as no one has shown they've published false info, though it seems the people whose emails have been published typically complain about it and try to discredit it. And what did they show - that the MSM is faking their impartiality.

As for other fake news, it seems liberals are pretending to be conservatives, and publish fake news so they can discredit conservatives. All they are really doing, is discrediting themselves and liberals who claim the fake news is from the right. http://www.npr.or...-suburbs
5 / 5 (1) Dec 04, 2016
Another ignorati rightwing echo chamber, courtesy of sackbag.

Ironic that sackbag appears to be unaware that it has been subject to exactly the same mode of manipulation as the liberal crybabies it so vehemently deplores.

Here then is objective proof of the flaw in the good professor's research, which turns out to be no more than a rhetorical question, since his hypothesized technology is no substitute for critical analysis.

But, then again, who has time to think critically or analytically these days?

Certainly not sackbag. Or sauce-for-the-goose ForF**kHeads.
5 / 5 (1) Dec 05, 2016
Fake new IS made for the right because they are so much more emotional--you know those guys who act before they think; all it takes is the mere suggestion that they not gonna get theirs and they lose it like a man-child.

That NPR report is sorta proof liberals are more thoughtful---they said fake liberal news was tried but their target audience saw through it.

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.