Diffusion Model
A type of generative AI model that creates data by starting with random noise and gradually removing it, step by step, until a coherent output (like an image) emerges. This process is called denoising.
Why It Matters
Diffusion models power the most impressive AI image generators — Midjourney, DALL-E 3, and Stable Diffusion — and are expanding to video and audio generation.
Example
Stable Diffusion starting with a noisy, static-like image and progressively refining it over 20-50 steps into a detailed photo-realistic image matching a text prompt.
Think of it like...
Like a sculptor starting with a rough block of marble and gradually chipping away until a beautiful statue emerges — each step removes a bit more randomness.
Related Terms
Generative AI
AI systems that can create new content — text, images, music, code, video — rather than just analyzing or classifying existing data. These models learn patterns from training data and generate novel outputs that resemble the original data.
Stable Diffusion
An open-source text-to-image diffusion model that generates detailed images from text descriptions. It works in a compressed latent space, making it more efficient than pixel-level diffusion.
DALL-E
A text-to-image AI model created by OpenAI that generates original images from text descriptions. DALL-E can create realistic images, art, and conceptual visualizations from natural language prompts.
Noise
Random variation or errors in data that do not represent true underlying patterns. In deep learning, noise can also refer to the random input used in generative models.
Denoising
The process of removing noise from data to recover the underlying clean signal. In generative AI, denoising is the core mechanism of diffusion models.