Stable Diffusion
An open-source text-to-image diffusion model that generates detailed images from text descriptions. It works in a compressed latent space, making it more efficient than pixel-level diffusion.
Why It Matters
Stable Diffusion democratized AI image generation by being open-source, enabling local execution, fine-tuning, and a massive ecosystem of community tools and models.
Example
Generating a photorealistic image from the prompt 'A cozy cabin in the snow with warm light glowing from the windows at dusk, cinematic lighting'.
Think of it like...
Like giving an artist a description of a scene and having them paint it — except this artist works in seconds and can produce infinite variations.
Related Terms
Diffusion Model
A type of generative AI model that creates data by starting with random noise and gradually removing it, step by step, until a coherent output (like an image) emerges. This process is called denoising.
Text-to-Image
AI models that generate visual images from natural language text descriptions (prompts). This technology converts written descriptions into original images, illustrations, or photorealistic visuals.
DALL-E
A text-to-image AI model created by OpenAI that generates original images from text descriptions. DALL-E can create realistic images, art, and conceptual visualizations from natural language prompts.
Latent Space
A compressed, lower-dimensional representation of data learned by a model. Points in latent space capture the essential features of the data, and nearby points represent similar data items.
Generative AI
AI systems that can create new content — text, images, music, code, video — rather than just analyzing or classifying existing data. These models learn patterns from training data and generate novel outputs that resemble the original data.