Large Language Model
A neural network trained on massive text datasets to generate, summarise, and reason about natural language.
Definition
A large language model (LLM) is a neural network trained on massive text datasets to generate, summarise, translate, and reason about natural language. LLMs form the foundation of systems such as ChatGPT, Claude, and Gemini. They operate by predicting the most likely next tokens in a sequence, which enables fluent text generation but also produces confident-sounding outputs that may be factually incorrect (hallucinations).
How It Relates to AI Threats
LLMs intersect with threats across multiple domains. Within Information Integrity, they enable the production of misinformation and hallucinated content at scale. Within Human-AI Control, they create risks of overreliance and automation bias as users treat LLM outputs as authoritative. LLMs also underpin agentic AI systems, where autonomous action introduces additional risk vectors.
Why It Occurs
- Scale of training data includes both accurate and inaccurate information
- The prediction mechanism optimises for plausibility rather than factual accuracy
- Users frequently lack the ability to verify LLM outputs
- Commercial deployment incentivises broad capability over narrow reliability
- Rapid adoption has outpaced the development of appropriate governance frameworks
Real-World Context
LLM-related incidents include Samsung engineers leaking proprietary code via ChatGPT (INC-23-0002), Italy’s temporary GDPR-based ban on ChatGPT (INC-23-0003), a lawyer citing hallucinated case law in federal court (INC-23-0005), and AI-generated phishing attacks leveraging LLM fluency (INC-23-0006).
Related Incidents
Samsung Semiconductor Trade Secret Leak via ChatGPT
Italy Temporary Ban on ChatGPT for GDPR Violations
AI-Fabricated Legal Citations in U.S. Courts
WormGPT: AI-Powered Business Email Compromise Tool
Jailbroken Claude AI Used to Breach Mexican Government Agencies
DOGE Uses ChatGPT to Flag and Cancel Federal Humanities Grants
Chinese AI Labs Conduct Industrial-Scale Distillation Attacks Against Claude
Indirect Prompt Injection Attacks on LLM-Integrated Applications
Zoom AI Training Terms of Service Controversy
GitHub Copilot Reproduces Verbatim Training Data Including Secrets
Microsoft Tay Twitter Chatbot Adversarial Manipulation
Related Threat Patterns
Related Terms
Last updated: 2026-02-14