This section covers the application of language model architectures to chemistry. Most notes here focus on models that treat molecular strings (SMILES, IUPAC names, InChI) as sequences, including encoder-only transformers like ChemBERTa and sequence-to-sequence models like ChemFormer and STOUT for structure-to-name translation. A growing set of notes also covers multimodal approaches: models like ChemVLM and InstructMol that connect 2D molecular images or graphs with natural language, enabling tasks like captioning, question answering, and structure retrieval.

Struct2IUPAC: Translating SMILES to IUPAC via Transformers
This paper proposes a Transformer-based approach (Struct2IUPAC) to convert chemical structures to IUPAC names, challenging the dominance of rule-based systems. Trained on ~47M PubChem examples, it achieves near-perfect accuracy using a round-trip verification step with OPSIN.

