Tech giants face questions on hate speech going into debates

Tech giants face questions on hate speech going into debates
Members of the media gather for a walk-through of the stage set-up for the first democratic debate, Wednesday, June 26, 2019. Ten presidential candidates, led by Sen. Elizabeth Warren, are set to converge on the debate stage on the first night of Democratic debates to offer their pitches to the American people and attempt a breakout moment for their campaigns. (AP Photo/Marta Lavandier)

Executives of Facebook, Google and Twitter faced questioning by a House panel Wednesday on their efforts to stanch terrorist content and viral misinformation on their social media platforms.

The scrutiny comes as the tech giants step up safety measures to prevent disinformation online targeting the Democratic presidential debates starting Wednesday night.

Lawmakers and tech industry executives are concerned that the debates could be targeted by Russian or other hostile parties to foment political conflict using social media, as happened in the 2016 election. U.S. intelligence officials have determined that Russia carried out a sweeping political disinformation campaign on social media to influence the election, and they have repeatedly warned about the threat of foreign meddling in American politics, especially ahead of elections.

"As the presidential debates begin, we are building on our efforts to protect the public conversation and enforce our policies against platform manipulation," Twitter said in a statement Wednesday. "It's always an election year on Twitter."

Facebook said it will have "a dedicated team proactively monitoring for threats as well as investigating any reports of abuse in real time in the lead up to, during and following the debates."

The hearing by the Homeland Security Committee was prompted by the mosque shootings in New Zealand in March that killed 50 people, attributed to a self-professed white supremacist who livestreamed the attacks on Facebook.

Rep. Bennie Thompson, D-Miss., the panel's chairman, noted that the livestreamed massacre occurred nearly two years after Facebook, Twitter, Google and other big tech companies established a global internet forum to fight the spread of online terrorist content.

"I want to know how you will prevent content like the New Zealand attack video from spreading on your platforms again," Thompson told the information policy executives from the three companies.

Thompson said he also wanted to know how the companies are working to keep and misinformation off their platforms.

Controversy over white nationalism and hate speech has dogged online platforms such as Facebook and Google's YouTube for years. In 2017, following the deadly violence in Charlottesville, Virginia, tech giants began banishing extremist groups and individuals espousing white supremacist views and support for violence. Facebook extended the ban to white nationalists.

But the big tech companies now are under closer scrutiny than ever in Congress, following a stream of scandals including Facebook's lapses in opening the personal data of millions of users to Donald Trump's 2016 campaign. Google's dominant search engine and hyper data collection have raised privacy concerns and accusations by Republicans of suppressing conservative viewpoints.

Trump on Wednesday renewed his criticism of the , insisting that their platforms censor conservative views. "They're doing it to me on Twitter," Trump said in an interview with Fox Business Network's "Mornings with Maria."

"You know, I have millions and millions of followers, but I will tell you they make it very hard for people to join me on Twitter, and they make it very much harder for me to get out the message," Trump said. "These people are all Democrats. It's totally biased toward Democrats."

Monika Bickert, Facebook's head of global policy management, said at the hearing that in response to the events in New Zealand, the company now prohibits livestreaming by people who have violated rules covering organizations and individuals deemed dangerous and potentially violent.

"We want to make sure we're doing everything to make sure it doesn't happen again," Bickert said.

The social network giant has improved its technology and techniques and is now able to more effectively detect terrorist content, including through tools now working in 19 languages, she said.

Twitter has suspended more than 1.5 million accounts for violations related to promoting terrorism from Aug. 1, 2015, to Dec. 31, 2018, said Nick Pickles, global senior strategist for public policy.

"We continue to invest in technology ... to ensure we can respond as quickly as possible to a potential incident," he said. "Twitter will take concrete steps to reduce the risk of livestreaming being abused by terrorists, while recognizing that during a crisis these tools are also used by news organizations, citizens and governments."

Google's policies for search, news and YouTube make clear the types of conduct that are prohibited, such as misrepresenting ownership or primary purpose, said Derek Slater, director of information policy.

"We want to do everything we can to ensure users are not exposed to content that promotes or glorifies acts of terrorism," Slater said.

Explore further

Facebook, Google face House hearing on white nationalism

© 2019 The Associated Press. All rights reserved.

Citation: Tech giants face questions on hate speech going into debates (2019, June 26) retrieved 18 May 2022 from
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Feedback to editors