Meta is changing who handles content moderation in East Africa

On Tuesday, Facebook confirmed that it would be ending its contract with Sama, the outsourcing firm responsible for moderating graphic content in East Africa. The news, first reported by Time Magazine, comes on the backs of a lawsuit by a former content moderator employed by Sama in Nairobi, Kenya, alleging severe mental health trauma as a result of their work…

On Tuesday, Facebook confirmed that it would be ending its contract with Sama, the outsourcing firm responsible for moderating graphic content in East Africa.

The news, first reported by Time Magazine, comes on the backs of a lawsuit by a former content moderator employed by Sama in Nairobi, Kenya, alleging severe mental health trauma as a result of their work alongside other labour violations.

According to Foxglove Legal, a legal nonprofit that investigates Big Tech companies, $2.2 billion European outsourcing firm Majorel will be taking over the contract. But Foxglove said Majorel is no better than Sama in its treatment of moderators — something also shown by Insider’s reporting from this past summer investigating the abhorrent treatment of Majorel’s content moderators working for TikTok’s parent-company ByteDance in Morocco.

It’s a sign of how troubled the content moderation industry remains. Despite several such lawsuits against social media companies like Facebook, Youtube, TikTok, and Reddit from around the world, workers are still often forced to work several hours long shifts moderating some of the most graphic content on the internet, from child abuse material to videos of gruesome accidents and beheadings, often with very few protections in place for their mental wellbeing.

In August, Insider investigated the labour conditions for TikTok’s content moderators working for Majorel in Morocco — the hub of ByteDance’s Middle East and North Africa content moderation operation. Workers told us they often work over 12-hour shifts, where they flag videos of animal abuse, sexual violence, and other forms of gruesome content. They had fewer breaks than their US counterparts and said the company’s “wellness counsellors” were of little help.

Social media companies claim to use sophisticated algorithms that help clean up people’s feeds, but this covers up the grim reality of how nearly every social media company works. Behind the scenes, a global workforce of tens of thousands filters out repugnant content so it doesn’t end up in front of your eyes.

In recent years, Facebook has settled lawsuits with moderators who reported PTSD as a result of their work for the company and promised to make changes to its labour conditions. But as Vittoria Elliott and I reported in 2020, the relatively meagre concessions rarely make it to workers in India, the Philippines, or Kenya.

Experts like Foxglove Legal have been calling on social media companies like Meta to bring their global content moderation workforce in-house. It may be the only way to make sure handling the worst elements of social media is shouldered by those who are truly accountable to the company and its users. Until then, contractors like those at Sama or Marjorel or dozens of other outsourcing firms will pay the price.

msn