Diffusion, Score-Based Models & Flow Matching
Canonical Papers
Deep Unsupervised Learning using Nonequilibrium Thermodynamics
Read paper →Denoising Diffusion Probabilistic Models
Read paper →Score-Based Generative Modeling through SDEs
Read paper →Flow Matching for Generative Modeling
Read paper →Core Mathematics
Forward diffusion adds noise:
Model learns to predict noise via MSE:
Score-based SDE view: forward SDE . Reverse-time SDE uses score .
Flow matching: train vector field to match the "true" conditional field (often optimal transport / straight lines).
Key Equation
Interactive Visualization
Why It Matters for Modern Models
- Stable Diffusion: latent diffusion — DDPM in a VAE latent space
- Sora: diffusion transformer over 3D spacetime patches
- Flow-matching and rectified flows enable one-step or few-step generation
Missing Intuition
What is still poorly explained in textbooks and papers:
- Intuitive explanation that denoising is learning ∇ₓ log pₜ(x) (scores), and how reverse-time SDE sampling corresponds to "walking uphill in log-density space"
- Visual/interactive demonstrations of different probability paths (diffusion vs optimal transport)