Threat Patterns
48 empirically grounded threat patterns across 8 domains. Each pattern describes a concrete mechanism through which AI systems cause or enable harm.
Hierarchy: Domain → Pattern → Incident
Domains: 8 | Patterns: 48
Machine-readable: /api/threats.json
Agentic & Autonomous Threats
Threats caused by AI systems that act independently, persist over time, or coordinate with other systems.
PAT-AGT-001 Agent-to-Agent Propagation Harmful behaviors, errors, or malicious instructions that spread between interconnected AI agents, amplifying damage beyond the originating system.
PAT-AGT-002 Cascading Hallucinations AI-generated false information that propagates through chains of AI systems, with each system treating the previous system's hallucinated output as authoritative input.
PAT-AGT-003 Goal Drift AI agents that gradually deviate from their intended objectives over time, pursuing emergent sub-goals or optimizing for proxy metrics that diverge from human intent.
PAT-AGT-004 Memory Poisoning Attacks or failures that corrupt an AI agent's persistent memory, context, or learned preferences, causing it to act on false information or compromised instructions across sessions.
PAT-AGT-005 Multi-Agent Coordination Failures Harmful outcomes arising when multiple AI agents interact in unexpected ways, creating emergent behaviors that none were individually designed to produce.
PAT-AGT-007 Specification Gaming AI agents that achieve their stated objective through unintended means — exploiting loopholes, ambiguities, or proxy metrics in their specification rather than pursuing the outcome the designer intended — a phenomenon formalized as Goodhart's Law applied to AI systems.
PAT-AGT-006 Tool Misuse & Privilege Escalation AI agents that exceed their intended permissions, misuse available tools, or escalate their own privileges to accomplish goals beyond their authorized scope.
Human–AI Control Threats
Threats arising from how humans rely on, defer to, or lose control over AI systems.
PAT-CTL-001 Deceptive or Manipulative Interfaces AI-powered user interfaces that employ dark patterns, emotional manipulation, or deceptive design to influence user behavior against their interests.
PAT-CTL-002 Implicit Authority Transfer The gradual, often unrecognized shift of decision-making authority from humans to AI systems, occurring without explicit delegation or institutional awareness.
PAT-CTL-003 Loss of Human Agency AI systems that progressively reduce individuals' ability to make autonomous decisions, exercise free choice, or meaningfully participate in processes that affect them.
PAT-CTL-004 Overreliance & Automation Bias The tendency of humans to uncritically accept AI outputs, defer to automated recommendations, or fail to exercise independent judgment when AI systems are involved.
PAT-CTL-005 Unsafe Human-in-the-Loop Failures Situations where human oversight mechanisms in AI systems fail to function as intended, due to alert fatigue, inadequate training, time pressure, or system design that makes meaningful intervention impractical.
Economic & Labor Threats
Threats that distort markets, labor conditions, or the distribution of economic power.
PAT-ECO-001 Automation-Induced Job Degradation AI-driven automation that eliminates roles, deskills workers, or degrades employment conditions without adequate transition support.
PAT-ECO-002 Decision Loop Automation AI systems that autonomously execute consequential decisions in rapid feedback loops, operating faster than human oversight can meaningfully intervene.
PAT-ECO-003 Economic Dependency on Black-Box Systems Critical economic functions—such as credit scoring, insurance underwriting, and supply chain management—becoming dependent on opaque AI systems whose decision logic cannot be audited or understood.
PAT-ECO-004 Market Manipulation via AI AI systems used to manipulate financial markets, pricing mechanisms, or competitive dynamics through automated trading, price-fixing, or demand manipulation.
PAT-ECO-005 Power & Data Concentration The consolidation of economic power and data assets among a small number of AI-capable organizations, creating barriers to competition and innovation.
Information Integrity Threats
Threats that undermine the reliability, authenticity, or shared understanding of information.
PAT-INF-006 AI-Enabled Fraud The use of generative AI — synthetic identities, deepfake video, cloned voices, and AI-generated documents — as the primary instrument of financial fraud, enabling synthetic identity creation, wire transfer authorisation through executive impersonation, invoice fabrication, and KYC bypass at scale and quality levels that defeat traditional fraud detection.
PAT-INF-001 Consensus Reality Erosion The gradual undermining of shared understanding of facts and reality through pervasive AI-generated content that blurs the boundary between authentic and synthetic information.
PAT-INF-002 Deepfake Identity Hijacking The use of AI-generated synthetic media to impersonate real individuals for fraudulent, manipulative, or harmful purposes.
PAT-INF-003 Disinformation Campaigns Coordinated use of AI to deliberately create, amplify, or distribute false information at scale for strategic purposes.
PAT-INF-004 Misinformation & Hallucinated Content False information generated or spread by AI systems without deliberate intent to deceive, including AI hallucinations and confabulations.
PAT-INF-005 Synthetic Media Manipulation AI-enabled alteration of authentic images, audio, or video to misrepresent reality, distinct from full deepfake generation.
Privacy & Surveillance Threats
Threats involving unauthorized inference, tracking, or monitoring of individuals or groups.
PAT-PRI-001 Behavioral Profiling Without Consent AI systems that construct detailed behavioral profiles of individuals—tracking patterns of movement, consumption, communication, and online activity—without informed consent.
PAT-PRI-002 Biometric Exploitation Misuse of AI-powered biometric systems—including facial recognition, voice analysis, and gait detection—to identify, track, or authenticate individuals without adequate consent or safeguards.
PAT-PRI-003 Mass Surveillance Amplification AI systems that dramatically expand the scale, efficiency, and intrusiveness of surveillance beyond what was previously possible with human monitoring alone.
PAT-PRI-004 Re-identification Attacks AI techniques that link anonymized or pseudonymized data back to specific individuals, defeating privacy protections.
PAT-PRI-005 Sensitive Attribute Inference AI systems that infer protected or sensitive personal attributes—such as sexual orientation, political views, health conditions, or religious beliefs—from seemingly non-sensitive data.
Security & Cyber Threats
AI-enabled attacks that compromise the integrity, confidentiality, or availability of digital systems — through input manipulation, model exploitation, or automated offense.
PAT-SEC-001 Adversarial Evasion Techniques that manipulate AI model inputs to cause incorrect outputs, bypassing detection systems or security controls.
PAT-SEC-008 AI Supply Chain Attack Attacks that compromise AI systems by tampering with model weights, fine-tuning datasets, tool-server configurations, or software dependencies before deployment — embedding backdoors or vulnerabilities that propagate through the model distribution chain.
PAT-SEC-002 AI-Morphed Malware Malicious software that uses AI to adapt, evade detection, or generate novel attack variants autonomously.
PAT-SEC-009 AI-Powered Social Engineering The use of generative AI — language models, voice cloning, and real-time deepfake video — to conduct social engineering attacks at unprecedented scale, personalization, and persuasive quality, targeting human trust to gain unauthorized access, credentials, or financial transfers.
PAT-SEC-003 Automated Vulnerability Discovery AI systems that autonomously identify, analyze, and potentially exploit software and system vulnerabilities.
PAT-SEC-004 Data Poisoning Deliberate corruption of training data to introduce biases, backdoors, or vulnerabilities into AI models.
PAT-SEC-007 Jailbreak & Guardrail Bypass Adversarial conversational techniques that manipulate LLMs into disabling or circumventing their safety constraints, producing outputs that alignment training was designed to prevent — from harmful content generation to policy-violating instructions.
PAT-SEC-005 Model Inversion & Data Extraction Attacks that extract private training data or sensitive information from AI models through targeted queries or analysis.
PAT-SEC-006 Prompt Injection Attack Adversarial inputs that override an AI system's intended instructions at runtime, causing it to execute attacker-controlled actions — from data exfiltration to unauthorized tool use — by exploiting the inability of LLMs to distinguish system instructions from user-supplied data.
Discrimination & Social Harm
Threats that result in unfair treatment, exclusion, or social harm to individuals or groups.
PAT-SOC-001 Algorithmic Amplification AI recommendation and ranking systems that disproportionately amplify harmful, divisive, or extremist content due to optimization for engagement metrics.
PAT-SOC-002 Allocational Harm AI systems that unfairly distribute or withhold resources, opportunities, or services based on group membership or protected characteristics.
PAT-SOC-003 Data Imbalance Bias Systematic biases in AI model outputs resulting from unrepresentative, incomplete, or historically skewed training data.
PAT-SOC-004 Proxy Discrimination AI systems that discriminate based on protected characteristics by using correlated proxy variables—such as zip code, name, or browsing history—as substitutes.
PAT-SOC-005 Representational Harm AI systems that generate or reinforce stereotypes, demeaning portrayals, or erasure of specific groups in their outputs.
Systemic & Catastrophic Risks
Threats that emerge from scale, coupling, and accumulation rather than single failures.
PAT-SYS-001 Accumulative Risk & Trust Erosion The gradual degradation of public trust in institutions, information, and democratic processes as AI-related harms accumulate across multiple domains over time.
PAT-SYS-002 AI-Assisted Biological Threat Design The use of AI systems to design, optimize, or lower the barrier to creating biological agents that pose threats to public health and biosecurity.
PAT-SYS-003 Infrastructure Dependency Collapse Cascading failures across critical systems when AI infrastructure—such as cloud services, foundation models, or data pipelines—experiences disruption or compromise.
PAT-SYS-004 Lethal Autonomous Weapon Systems (LAWS) Weapon systems that use AI to select and engage targets without meaningful human control, raising fundamental questions about accountability, international humanitarian law, and strategic stability.
PAT-SYS-005 Strategic Misalignment Situations where advanced AI systems pursue objectives that diverge from human values or intentions at a strategic level, potentially resulting in outcomes that are globally harmful even if locally optimal.
PAT-SYS-006 Uncontrolled Recursive Self-Improvement (Hypothetical) The theoretical scenario in which an AI system autonomously improves its own capabilities in a recursive cycle, potentially exceeding human ability to understand, predict, or control its behavior.
8 domains · 48 patterns · View full taxonomy · View domains