Study finds legacy media boosted fake Russian Twitter accounts in 2016

twitter
Credit: CC0 Public Domain

Many legacy media outlets played an unwitting role in the growth of the four most successful fake Twitter accounts hosted by the Russian Internet Research Agency (IRA) that were created to spread disinformation during the 2016 U.S. presidential campaign, according to a study led by a University at Buffalo communication researcher.

In roughly two years beginning in late 2015, these accounts went from obscurity to microcelebrity status, growing from about 100 to more than 100,000 followers. With its heavily populated follower base ready to spread the word—like all heavily engaged Twitter audiences—the IRA could strategically deploy messages and provide visible metrics, creating an illusion of authority and authenticity that often escaped the scrutiny of casual consumers and professional journalists.

The frantic retweets, by what the study showed to be extreme ideological enclaves, certainly fueled the accumulation of followers, but Yini Zhang, Ph.D., an assistant professor of communication at UB, says that mainstream and hyperpartisan news media also significantly amplified IRA messaging and contributed to that follower growth by unknowingly embedding IRA tweets in their content.

Zhang says there was an ideological asymmetry to the study's results. Of the four puppet accounts in the study, @TEN_GOP and @Pamela_Moore13 posed as conservative trolls, while @Crystal1Johnson and @glod_up imitated liberals.

"We did not observe the same effect on the liberal and conservative accounts," she says. "The two conservative accounts received a huge boost from and hyperconservative media quoting tweets in their , but we did not see mainstream media and hyperprogressive media doing the same thing for the two liberal accounts."

The findings, published in the Journal of Communication, reveal how large social media followings can often depend on a combination of the dynamics within a particular platform and the news media's treatment of the messages emerging from those platforms. The evidence revealed in the study provides insights into the ecology of the 21st century political communication environment, suggesting that people's tendency to seek confirmation and engage with pro-attitudinal information, as well as the media's drive for audience attention, can work in favor of successful political disinformation actors.

In this case, constructive attempts to provide new information by integrating digital and legacy content ironically resulted in the unintended spread of disinformation, which Zhang defines as fabricated information that's intended to cause harm in ways that benefit its agents.

"Examining how and why these accounts grew so quickly and to such astounding proportions allows us to understand the mechanisms of influence accrual in the digital era," says Zhang, the study's corresponding author and an expert in social media and political communication. "None of this was intentional. It's about operational realties.

"But with this knowledge, we can begin to address and curtail the problem of disinformation."

The research team started their work with 2,700 puppet accounts released in 2017 by the House Intelligence Committee, which received the information from Twitter. From that group, the researchers identified the four most retweeted English-speaking accounts: two conservative accounts and two liberal accounts. They collected data from Twitter about the tweets and retweets of the IRA accounts. They then searched more than 200 across the ideological spectrum to determine where the uptake of IRA tweets was occurring.

"Strong posts can validate content," says Zhang. "But in their effort to turn heads, these legacy outlets were contributing to the growth of Russian sock puppet accounts."

The processes of incorporating digital content into mainstream media makes sense, but requires careful consideration, according to Zhang.

"Social media content looks very attractive given the cost cutting realities in mainstream media and lost advertising revenue," says Zhang. "But it also demonstrates a vulnerability within the current economy.

"Turning heads might also mean unintentionally contributing to the growth of fake accounts, which should be subject to the same questions of credibility as any other news source: Is this in fact what it actually claims to be?"

More information: Yini Zhang et al, Assembling the Networks and Audiences of Disinformation: How Successful Russian IRA Twitter Accounts Built Their Followings, 2015–2017, Journal of Communication (2021). DOI: 10.1093/joc/jqaa042

Journal information: Journal of Communication

Citation: Study finds legacy media boosted fake Russian Twitter accounts in 2016 (2021, July 16) retrieved 24 April 2024 from https://phys.org/news/2021-07-legacy-media-boosted-fake-russian.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Artificial intelligence system could help counter the spread of disinformation

12 shares

Feedback to editors