
ChemBERTa-3: Open Source Training Framework
An open-source framework integrating DeepChem and Ray for training and benchmarking chemical foundation models like …

An open-source framework integrating DeepChem and Ray for training and benchmarking chemical foundation models like …

Optimizing transformer pretraining for molecules using MLM vs MTR objectives, scaling to 77M compounds from PubChem for …

A 46.8M parameter transformer for molecular generation trained on 1.1B SMILES, introducing pair-tuning for efficient …

BART-based Transformer pre-trained on 100M molecules using self-supervision to accelerate convergence on chemical …

Multimodal chemical model integrating 5 modalities (2D graphs, 3D conformations, images, MS2/IR spectra) trained on 7.6M …

A multimodal search engine that integrates text passages, molecular diagrams, and reaction data to enable passage-level …

A Transformer-based model for translating SMILES to IUPAC names, trained on ~1 billion molecules, achieving ~99% …

A deep-learning neural machine translation approach to translate between SMILES strings and IUPAC names using the STOUT …

A Transformer-based model for translating between SMILES strings and IUPAC names, trained on 47M PubChem examples, …

Sequence-to-sequence Transformer translating InChI identifiers to IUPAC names with 91% accuracy on organic compounds.

A 26B parameter multimodal LLM for chemistry, combining InternViT-6B and ChemLLM-20B for molecular structure …

Deep learning model using improved SwinTransformer encoder and attention-based feature fusion to convert molecular …