Facebook hit with $150 billion lawsuit over Myanmar hate speech

The tech giant has been accused of failing to prevent disinformation from thriving on its platform

Meta, formerly known as Facebook, could be forced to pay even $150 billion (£113bn) for its alleged role in the 2017 Rohingya genocide.

Legal complaints filed in the UK and US on behalf of the Rohingya refugees claim that the tech giant failed to prevent anti-Rohingya hate speech and disinformation from thriving on its platform.

This included thousands of posts describing the community as animals and foreign invaders, falsely accusing them of crimes, and calling for them to be killed. 

Facebook’s algorithm had allegedly amplified such hateful posts on users’ news feeds, with the company failing to hire enough Burmese-speaking content moderators despite record-breaking profits that year.

The spread of anti-Rohingya propaganda ultimately resulted in real-life violence that cost the lives of 24,000 people and displaced up to a million, forcing them into “abject poverty”, according to the class-action complaint filed in California by law firms Edelson and Fields.

The US lawsuit is seeking damages “in excess of $150 billion”. The legal notice to Meta’s London offices has not been made publicly available.

Related Resource

The state of brand protection 2021

A new front opens up in the war for brand safety

A log-in screen with a red background - whitepaper from MimecastFree download

The lawsuit references claims made by a former Facebook employee, who said that the company’s executives “were fully aware that posts ordering hits by the Myanmar government on the minority Muslim Rohingya were spreading wildly on Facebook”, and that “the issue of the Rohingya being targeted on Facebook was well known inside the company for years”.

The claims echo testimonies made by another former-employee-turned-whistleblower, Frances Haugan, who in October told members of the US Congress that Facebook was “literally fanning” ethnic violence in developing countries.

Weeks later, Haugen told UK’s MPs that, due to shortages of moderators, Facebook had been unable to police harmful content in multiple languages around the world, leading to civil unrest in Myanmar in 2017 as well as Ethiopia in 2021.

However, the issue also impacts the UK, she added, due to the fact that Facebook's AI is unable to detect online abuse in British English.

Despite the widely-reported anti-Rohingya violence in Myanmar, the tech giant also failed to prevent the spread of anti-Muslim hate speech on its platform in the Assam region of northeast India, according to the lawsuit. 

Meta didn’t respond to IT Pro’s request for comment, yet had previously admitted to being “too slow to prevent misinformation and hate” in Myanmar.

Featured Resources

How virtual desktop infrastructure enables digital transformation

Challenges and benefits of VDI

Free download

The Okta digital trust index

Exploring the human edge of trust

Free download

Optimising workload placement in your hybrid cloud

Deliver increased IT agility with the cloud

Free Download

Modernise endpoint protection and leave your legacy challenges behind

The risk of keeping your legacy endpoint security tools

Download now

Recommended

Senator wants social media companies held liable for spreading anti-vax lies
social media

Senator wants social media companies held liable for spreading anti-vax lies

23 Jul 2021
Four in ten Americans would rather give up a pet than social media
social media

Four in ten Americans would rather give up a pet than social media

21 Jun 2021

Most Popular

Sony pulls out of MWC 2022
Business operations

Sony pulls out of MWC 2022

14 Jan 2022
How to boot Windows 11 in Safe Mode
Microsoft Windows

How to boot Windows 11 in Safe Mode

6 Jan 2022
Dell XPS 15 (2021) review: The best just got better
Laptops

Dell XPS 15 (2021) review: The best just got better

14 Jan 2022