This group covers models that learn continuous latent representations of molecules and use that space for generation and optimization. The seminal SMILES VAE (Gomez-Bombarelli et al., 2018) established the paradigm: encode molecules into a smooth latent space, then search it via gradient-based or Bayesian optimization.

PaperYearArchitectureKey Idea
Grammar VAE2017VAE + CFGContext-free grammar decoder ensuring syntactically valid SMILES
Automatic Chemical Design2018VAESeminal SMILES VAE enabling Bayesian optimization in latent space
LatentGAN2019WGAN + heteroencoderWasserstein GAN in heteroencoder latent space, bypassing SMILES syntax
CogMol2020VAE + CLaSSControlled latent sampling for target-specific design without retraining
PASITHEA2021Gradient dreamingDeep dreaming on SELFIES via property network inversion
LIMO2022VAE + property predictorStacked property predictor enabling gradient-based search for high-affinity molecules

All Notes