This section covers the core families of generative models used in modern machine learning. Notes begin with the foundational variational autoencoder (VAE) and its extensions (importance-weighted objectives, contrastive priors), then move through continuous normalizing flows, neural ODEs, score-based and diffusion models, and flow matching. The thread connecting these works is the shared goal of learning to sample from complex distributions, and each set of notes tries to make the mathematical connections between approaches explicit rather than treating them as isolated methods.

YearPaperKey Idea
1994Mixture Density NetworksNeural nets predicting Gaussian mixture parameters for multimodal outputs
1997A Convexity Principle for Interacting GasesDisplacement interpolation via optimal transport for energy functionals
2011Score Matching and Denoising AutoencodersProves denoising autoencoders are equivalent to score matching
2013Auto-Encoding Variational BayesReparameterization trick enabling end-to-end VAE training
2016Importance Weighted AutoencodersTighter log-likelihood bounds via importance sampling in VAEs
2018Neural ODEsContinuous-depth networks via ODE solvers with constant-memory training
2021Contrastive Learning for VAE PriorsNoise contrastive priors to fix the VAE “prior hole” problem
2021D3PM: Discrete Diffusion ModelsDiffusion extended to discrete state-spaces with structured transitions
2021Score-Based Generative Modeling with SDEsUnified SDE framework for score-based models with Predictor-Corrector sampling
2022Latent Diffusion ModelsDiffusion in compressed latent space for high-res image synthesis
2023Consistency ModelsOne-step generation by mapping ODE trajectory points to their origin
2023Flow Matching for Generative ModelingSimulation-free CNF training with optimal transport paths
2023Rectified FlowODE-based generation with iterative trajectory straightening (reflow)
2023Stochastic InterpolantsCNFs between arbitrary densities via quadratic velocity field objective