AI Threats Affecting Workers
How AI-enabled threats affect employees, contractors, gig workers, and professionals — through job displacement, algorithmic management, surveillance, or degraded working conditions.
individualsHow AI Threats Appear
For workers, AI-enabled threats most commonly surface through:
- Algorithmic management — AI systems that set schedules, monitor performance, assign tasks, or make termination decisions with limited human oversight
- Job displacement and degradation — Automation that eliminates roles, compresses wages, or reduces skilled work to AI-supervised monitoring tasks
- Workplace surveillance — AI-powered monitoring of keystrokes, screen activity, location, communication patterns, and biometric data
- Discriminatory hiring and evaluation — AI screening tools that systematically disadvantage candidates or employees based on protected characteristics
- Deskilling — AI tools that absorb professional knowledge, reducing workers to interface operators and eroding career development pathways
These impacts are documented through real-world incidents affecting workers across multiple sectors.
Relevant AI Threat Domains
- Economic & Labor — Job displacement, market concentration, and economic dependency on AI systems
- Privacy & Surveillance — Workplace monitoring and behavioral profiling
- Discrimination & Social Harm — Bias in hiring, evaluation, and management systems
- Human-AI Control — Loss of professional agency and overreliance on automated decisions
What to Watch For
Indicators of AI-related labor harm in workplace contexts:
- Performance evaluations driven primarily by automated metrics rather than qualitative assessment
- Hiring processes where AI screening precedes any human review
- Productivity monitoring that captures granular behavioral data beyond task completion
- Role changes that reduce professional judgment to AI-output review
- Compensation models tied to AI-generated efficiency benchmarks
Regulatory Context
- EU AI Act — Classifies AI systems used in employment and worker management as high-risk, requiring human oversight and transparency
- NIST AI RMF — Addresses organizational risk management for AI systems affecting workforce operations
- Labor regulations in multiple jurisdictions are developing specific requirements for algorithmic management transparency and worker notification
For classification rules and evidence standards, refer to the Methodology.
Last updated: 2026-03-03 · Back to Affected Groups