Skip to main content
TopAIThreats home TOP AI THREATS
HARM-004 Discrimination

Discrimination Harm

Unfair differential treatment of individuals or groups based on protected characteristics, produced or amplified by AI systems.

Discrimination harm arises when AI systems produce or amplify unfair differential treatment based on protected characteristics such as race, gender, age, disability, ethnicity, religion, or socioeconomic status. This harm manifests in two primary forms: direct discrimination, where an AI system explicitly uses a protected attribute as a decision factor, and indirect discrimination, where ostensibly neutral features serve as proxies for protected characteristics due to historical correlations in training data. Because AI systems learn patterns from historical data that often reflects existing societal inequities, they can encode and perpetuate those inequities at scale, affecting millions of decisions simultaneously.

Documented cases span nearly every domain where AI is used for consequential decision-making. Hiring algorithms have systematically downranked resumes associated with female applicants. Facial recognition systems have demonstrated significantly higher error rates for individuals with darker skin tones, leading to wrongful identifications and arrests. Healthcare resource allocation algorithms have assigned lower risk scores to Black patients than to equally sick white patients, resulting in reduced access to care. Credit scoring and insurance pricing models have produced disparate outcomes along racial and geographic lines that correlate with historically redlined communities. Content moderation systems have disproportionately flagged or suppressed speech from marginalized communities.

Addressing discrimination harm requires intervention at multiple stages of the AI lifecycle. Pre-deployment audits using disaggregated performance metrics can reveal disparities before systems affect real populations. Fairness constraints can be incorporated into model training, though trade-offs between different fairness definitions require careful normative judgment. Post-deployment monitoring with regular bias audits helps detect emergent disparities as population distributions shift over time. Regulatory requirements for algorithmic impact assessments, combined with accessible grievance mechanisms for affected individuals, provide institutional accountability. Ultimately, technical measures alone are insufficient; reducing discrimination harm also demands diverse development teams and meaningful engagement with the communities most likely to be affected.

Last updated: 2026-02-25