Facebook hit with $150bn lawsuit over Myanmar hate speech

Meta Sign
Image: Dennis

The tech giant has been accused of failing to prevent disinformation from thriving on its platform

Print

PrintPrint
Life

Read More:

7 December 2021 | 0

Meta, formerly known as Facebook, could be forced to pay even $150 billion for its alleged role in the 2017 Rohingya genocide.

Legal complaints filed in the UK and US on behalf of the Rohingya refugees claim that the tech giant failed to prevent anti-Rohingya hate speech and disinformation from thriving on its platform.

This included thousands of posts describing the community as animals and foreign invaders, falsely accusing them of crimes, and calling for them to be killed.

 

advertisement



 

Facebook’s algorithm had allegedly amplified such hateful posts on users’ news feeds, with the company failing to hire enough Burmese-speaking content moderators despite record-breaking profits that year.

The spread of anti-Rohingya propaganda ultimately resulted in real-life violence that cost the lives of 24,000 people and displaced up to a million, forcing them into “abject poverty”, according to the class-action complaint filed in California by law firms Edelson & Fields.

The US lawsuit is seeking damages “in excess of $150 billion”. The legal notice to Meta’s London offices has not been made publicly available.

The lawsuit references claims made by a former Facebook employee, who said that the company’s executives “were fully aware that posts ordering hits by the Myanmar government on the minority Muslim Rohingya were spreading wildly on Facebook”, and that “the issue of the Rohingya being targeted on Facebook was well known inside the company for years”.

The claims echo testimonies made by another former-employee-turned-whistleblower, Frances Haugan, who in October told members of the US Congress that Facebook was “literally fanning” ethnic violence in developing countries.

Weeks later, Haugen told UK MPs that, due to shortages of moderators, Facebook had been unable to police harmful content in multiple languages around the world, leading to civil unrest in Myanmar in 2017 as well as Ethiopia in 2021.

However, the issue also impacts the UK, she added, due to the fact that Facebook’s AI is unable to detect online abuse in British English.

Despite the widely-reported anti-Rohingya violence in Myanmar, the tech giant also failed to prevent the spread of anti-Muslim hate speech on its platform in the Assam region of northeast India, according to the lawsuit.

© Dennis Publishing

Read More:



Comments are closed.

Back to Top ↑