Facebook to Remove Misinformation That Leads to Violence

Advertisement

Supported by

Facebook to Remove Misinformation That Leads to Violence

Image
A Rohingya Muslim woman at a displacement camp in Myanmar. Facebook has been accused of facilitating attacks on the Rohingya in the country by allowing anti-Muslim hate speech on its platform.CreditLauren Decicca/Getty Images
  • July 18, 2018

SAN FRANCISCO — Facebook, facing growing criticism for posts that have incited violence in some countries, said Wednesday that it would begin removing misinformation that could lead to people being physically harmed.

The policy expands Facebook’s rules about what type of false information it will remove, and comes largely in response to instances in Sri Lanka, Myanmar and India, in which rumors spread on Facebook led to real-world attacks.

The new rules do not apply to Facebook’s other big social media properties, Instagram and WhatsApp.

“We have identified that there is a type of misinformation that is shared in certain countries that can incite underlying tensions and lead to physical harm offline. We have a broader responsibility to not just reduce that type of content but remove it,” said Tessa Lyons, a Facebook product manager.

Facebook has been roundly criticized over the way its platform has been used to spread hate speech and false information that leads to violence. The company has struggled to balance its belief in free speech with those concerns, particularly in countries where access to the internet is relatively new and mainstream news sources that would counter social media rumors are limited.

In Myanmar, Facebook has been accused by United Nations investigators and human rights groups of facilitating violence against Rohingya Muslims, a minority ethnic group, by allowing anti-Muslim hate speech and false news.

In Sri Lanka, riots broke out after false news pitted the country’s majority Buddhist community against Muslims. Near-identical social media rumors have also led to attacks in India and Mexico. In many cases, the rumors included no call for violence, but amplified underlying tensions.

In an interview published Wednesday by the technology news site Recode, Mark Zuckerberg, Facebook’s chief executive, attempted to explain how the company is trying to differentiate between offensive speech — the example he used was people who deny the Holocaust — and false posts that could lead to physical harm.

“I think that there’s a terrible situation where there’s underlying sectarian violence and intention. It is clearly the responsibility of all of the players who were involved there,” Mr. Zuckerberg told Recode’s Kara Swisher, who will become an Opinion contributor with The New York Times later this summer.

While the social media company already has rules in place in which a direct threat of violence or hate speech is removed, it has been hesitant to remove rumors that don’t directly violate its content policies.

Under the new rules, Facebook has said it will create partnerships with local civil society groups to identify misinformation for removal. The new rules are already being put in effect in Sri Lanka, and Ms. Lyons said the company hopes to soon introduce them in Myanmar, before expanding elsewhere.

Follow Sheera Frenkel on Twitter: @sheeraf

Advertisement

Be the first to comment

Leave a Reply

Your email address will not be published.


*