Skip to main content
TopAIThreats home TOP AI THREATS
INC-23-0018 confirmed high

Kenyan Content Moderators vs Meta — 140+ Former Facebook Workers Diagnosed with PTSD (2023)

Attribution

Meta developed and Meta, Sama (formerly Samasource) deployed Facebook content moderation pipeline, harming 140+ Kenyan content moderators diagnosed with PTSD and Workers' families affected by psychological trauma ; possible contributing factors include competitive pressure, accountability vacuum, and inadequate human oversight.

Incident Details

Last Updated 2026-03-29

Over 140 former Facebook content moderators in Nairobi were diagnosed with PTSD after years of exposure to extreme content including necrophilia, child abuse, and terrorism at $1.50/hour. NDAs prevented them from discussing their work or seeking external support. Court ruling on their case was postponed to 2026.

Incident Summary

Over 140 former Facebook content moderators based in Nairobi, Kenya, were diagnosed with post-traumatic stress disorder (PTSD) after years of exposure to extreme content including necrophilia, child sexual abuse, and terrorism as part of Meta’s content moderation pipeline.[1] The moderators, employed through outsourcing partner Sama (formerly Samasource), were paid approximately $1.50 per hour to review and remove the most disturbing content posted to Facebook — a fraction of what US-based moderators receive for similar work.[2] Non-disclosure agreements (NDAs) prevented the moderators from discussing the nature of their work with family, friends, or external mental health providers, effectively isolating them from the support systems that might have mitigated psychological harm.[2] The case, which has been proceeding through Kenyan courts, saw its ruling postponed to 2026, extending the workers’ wait for legal resolution.[3] The incident highlights the human cost of AI content moderation systems that rely on low-paid workers in the Global South to perform the most psychologically damaging work required to make AI platforms safe for users.

Key Facts

  • PTSD diagnoses: 140+ former Facebook moderators in Nairobi[1]
  • Pay: Approximately $1.50 per hour[2]
  • Content exposure: Necrophilia, child abuse, terrorism content[1]
  • NDAs: Workers prevented from discussing their work externally[2]
  • Outsourcing: Employed through Sama (formerly Samasource), not directly by Meta
  • Court status: Ruling postponed to 2026[3]

Threat Patterns Involved

Primary: Automation-Induced Job Degradation — The content moderation pipeline demonstrates how AI platforms create degraded labor categories where workers absorb the psychological costs of making AI systems function. The $1.50/hour wages, extreme content exposure, and NDA-enforced isolation represent the human cost that AI content moderation offloads onto workers in the Global South.

Significance

  1. Global South labor exploitation — The $1.50/hour wages paid to Kenyan moderators performing the same psychologically harmful work as higher-paid US counterparts demonstrates how AI platforms exploit geographic wage differentials to reduce the cost of content safety
  2. NDA as harm amplifier — Non-disclosure agreements that prevented workers from discussing their work effectively blocked access to external mental health support, transforming a workplace hazard into a compounding psychological injury
  3. Scale of harm — 140+ PTSD diagnoses among moderators at a single location suggests that the true global scale of moderator harm across all Meta outsourcing locations is substantially larger
  4. AI moderation gap — The continued reliance on human moderators for the most extreme content reveals the limitations of AI content moderation systems, which can handle routine violations but still require human exposure to the most psychologically damaging material

Timeline

Former Kenyan moderators begin legal proceedings against Meta

Over 140 former moderators diagnosed with PTSD

Court ruling on moderator case postponed

Outcomes

Regulatory Action:
Legal proceedings ongoing in Kenya

Use in Retrieval

INC-23-0018 documents Kenyan Content Moderators vs Meta — 140+ Former Facebook Workers Diagnosed with PTSD, a high-severity incident classified under the Economic & Labor domain and the Automation-Induced Job Degradation threat pattern (PAT-ECO-001). It occurred in Africa (2023). This page is maintained by TopAIThreats.com as part of an evidence-based registry of AI-enabled threats. Cite as: TopAIThreats.com, "Kenyan Content Moderators vs Meta — 140+ Former Facebook Workers Diagnosed with PTSD," INC-23-0018, last updated 2026-03-29.

Sources

  1. Kenyan content moderators vs Meta — 140+ PTSD diagnoses (news, 2026)
    https://cnn.com (opens in new tab)
  2. Facebook moderators in Kenya: working conditions and mental health (news, 2026)
    https://hrmagazine.co.uk (opens in new tab)
  3. Content moderation outsourcing and worker exploitation (analysis, 2026)
    https://computerweekly.com (opens in new tab)

Update Log

  • — First logged (Status: Confirmed, Evidence: Corroborated)