Skip to main content
TopAIThreats home TOP AI THREATS
AI Capability

Diffusion Model

A class of generative AI model that creates new data by learning to reverse a gradual noising process — starting from random noise and iteratively denoising it into coherent outputs such as images, video, or audio. Diffusion models power leading image generators including Stable Diffusion, DALL-E, and Midjourney.

Definition

A diffusion model is a generative AI architecture that produces new data by learning to reverse a diffusion (noising) process. During training, the model learns how noise is gradually added to real data samples until they become pure random noise. During generation, the model starts from random noise and iteratively removes noise, guided by learned patterns, to produce coherent outputs. Diffusion models have become the dominant architecture for high-quality image generation since 2022, surpassing GANs in output quality and controllability. Key implementations include Stable Diffusion (Stability AI), DALL-E (OpenAI), Midjourney, and Imagen (Google). The architecture has been extended to video generation (Sora, Runway), audio synthesis, and 3D object generation.

How It Relates to AI Threats

Diffusion models are the current state-of-the-art technology driving synthetic media threats within the Information Integrity Threats domain. They have dramatically lowered the barrier to creating photorealistic fake images, manipulated photographs, and increasingly convincing synthetic video. The ease of use of diffusion-based tools — text-to-image generation from natural language prompts — means that creating synthetic media no longer requires technical expertise. This has amplified threats including non-consensual intimate imagery generation, political disinformation through fabricated photographs, and deepfake-augmented fraud schemes.

Why It Occurs

  • Diffusion models produce higher-quality, more diverse outputs than previous generative architectures
  • Text-to-image interfaces make generation accessible to non-technical users through natural language prompts
  • Open-source diffusion models (Stable Diffusion, Flux) can be run locally, evading platform-level content policies
  • Fine-tuning techniques allow diffusion models to be customised to generate specific individuals’ likenesses from a few reference photos
  • The rapid pace of research continuously improves generation quality, speed, and controllability

Real-World Context

Since 2022, diffusion model-generated images have been used in documented disinformation campaigns, non-consensual intimate imagery targeting individuals, financial fraud, and fabricated evidence. The accessibility of open-source diffusion models has shifted the synthetic media threat landscape from requiring technical expertise (GAN-era) to requiring only a text prompt. Detection of diffusion model outputs is an active area of research, with techniques including metadata analysis, statistical artifact detection, and provenance standards such as C2PA. Regulatory responses including the EU AI Act require transparency labeling for AI-generated content.

Last updated: 2026-04-03