Representation Learning & Embedding Geometry
Canonical Papers
Representation Learning: A Review and New Perspectives
Read paper →Core Mathematics
Learn a mapping such that inner products or distances reflect meaningful relations.
Contrastive objective (InfoNCE-style):
This pushes "positive" pairs together, "negatives" apart; at optimum, it maximizes a lower bound on mutual information between views.
Key Equation
Interactive Visualization
Why It Matters for Modern Models
- Word & token embeddings in LMs, vision embeddings in CLIP-like models, multimodal embeddings in Gemini and GPT-4V
- Latent spaces of Stable Diffusion designed so distances correspond to semantic similarity
Missing Intuition
What is still poorly explained in textbooks and papers:
- Geometric explanation of anisotropy (representations bunch along a few directions) and how normalization/whitening alter behavior
- Visuals showing how representations evolve across layers (local to global features)