<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/"><channel><title>Position Papers: Perspectives, Critiques, and Meta-Science on Hunter Heidenreich | ML Research Scientist</title><link>https://hunterheidenreich.com/paper-types/position/</link><description>Recent content in Position Papers: Perspectives, Critiques, and Meta-Science on Hunter Heidenreich | ML Research Scientist</description><image><title>Hunter Heidenreich | ML Research Scientist</title><url>https://hunterheidenreich.com/img/avatar.webp</url><link>https://hunterheidenreich.com/img/avatar.webp</link></image><generator>Hugo -- 0.147.7</generator><language>en-US</language><copyright>2026 Hunter Heidenreich</copyright><lastBuildDate>Sun, 05 Apr 2026 00:00:00 +0000</lastBuildDate><atom:link href="https://hunterheidenreich.com/paper-types/position/index.xml" rel="self" type="application/rss+xml"/><item><title>NLP Models That Automate Programming for Chemistry</title><link>https://hunterheidenreich.com/notes/chemistry/llm-applications/nlp-models-transform-chemistry/</link><pubDate>Thu, 26 Mar 2026 00:00:00 +0000</pubDate><guid>https://hunterheidenreich.com/notes/chemistry/llm-applications/nlp-models-transform-chemistry/</guid><description>A perspective on how code-generating LLMs like OpenAI Codex and GPT-3 will reshape computational chemistry research workflows and education.</description><content:encoded><![CDATA[<h2 id="a-perspective-on-code-generating-llms-for-chemistry">A Perspective on Code-Generating LLMs for Chemistry</h2>
<p>This is a <strong>Position</strong> paper that argues large language models (LLMs) capable of generating code from natural language prompts, specifically OpenAI&rsquo;s Codex and GPT-3, are poised to transform both chemistry research and chemistry education. Published in the inaugural volume of Digital Discovery (RSC), the paper combines a brief history of NLP developments with concrete demonstrations of code generation for computational chemistry tasks, then offers a forward-looking perspective on challenges and opportunities.</p>
<h2 id="bridging-the-gap-between-natural-language-and-scientific-software">Bridging the Gap Between Natural Language and Scientific Software</h2>
<p>The authors identify a core friction in modern computational chemistry: while the number of available software packages has grown dramatically, researchers spend a large fraction of their time learning interfaces to these packages rather than doing science. Tasks like searching documentation, following tutorials, and trial-and-error experimentation with APIs consume effort that could be directed at research itself.</p>
<p>At the same time, programming assignments in chemistry courses serve dual pedagogical purposes (reinforcing physical intuition and teaching marketable skills), but are constrained by students&rsquo; median programming experience. The emergence of code-generating NLP models opens the possibility of reducing both barriers simultaneously.</p>
<h2 id="code-generation-as-a-chemistry-interface">Code Generation as a Chemistry Interface</h2>
<p>The paper&rsquo;s core thesis is that NLP models trained on code can serve as a natural language interface to the entire ecosystem of scientific computing tools. The authors demonstrate this with several concrete examples using OpenAI Codex:</p>
<ol>
<li>
<p><strong>Quantum chemistry</strong>: Prompting Codex to &ldquo;compute the dissociation curve of H2 using pyscf&rdquo; produced correct, runnable code that selected <a href="https://en.wikipedia.org/wiki/Hartree%E2%80%93Fock_method">Hartree-Fock</a> with <a href="https://en.wikipedia.org/wiki/STO-nG_basis_sets">STO-3G</a>. A follow-up prompt requesting &ldquo;the most accurate method&rdquo; caused it to switch to <a href="https://en.wikipedia.org/wiki/Coupled_cluster">CCSD</a> in a large basis set.</p>
</li>
<li>
<p><strong>Chemical entity recognition</strong>: Using GPT-3 with only three training examples, the authors demonstrated extraction of chemical entity names from published text, a task that previously required thousands of labeled examples.</p>
</li>
<li>
<p><strong>Molecular visualization</strong>: Drawing caffeine from its <a href="/notes/chemistry/molecular-representations/notations/smiles/">SMILES</a> string, generating Gaussian input files from SMILES, implementing random walks, and downloading and analyzing <a href="https://en.wikipedia.org/wiki/Protein_Data_Bank">PDB structures</a> with MDTraj.</p>
</li>
<li>
<p><strong>Voice-controlled molecular dynamics</strong>: The authors previously built MARVIS, a voice-controlled <a href="/notes/chemistry/molecular-simulation/">molecular dynamics</a> analysis tool that uses GPT-3 to convert natural language into <a href="https://en.wikipedia.org/wiki/Visual_Molecular_Dynamics">VMD</a> commands. Only about a dozen examples were needed to teach GPT-3 to render proteins, change representations, and select atoms.</p>
</li>
</ol>
<p>An important caveat: the authors emphasize that all chemistry &ldquo;knowledge&rdquo; (including the SMILES string for caffeine) is entirely contained in the model&rsquo;s learned floating-point weights. The model has no access to databases or curated lists of chemical concepts.</p>
<h2 id="demonstrations-and-practical-evaluation">Demonstrations and Practical Evaluation</h2>
<p>Rather than a formal experimental evaluation with benchmarks and metrics, this perspective paper relies on qualitative demonstrations. The key examples, with full details provided in the ESI, include:</p>
<table>
  <thead>
      <tr>
          <th>Task</th>
          <th>Input</th>
          <th>Result</th>
      </tr>
  </thead>
  <tbody>
      <tr>
          <td>H2 dissociation curve</td>
          <td>Natural language prompt</td>
          <td>Correct PySCF code (HF/STO-3G)</td>
      </tr>
      <tr>
          <td>Upgrade method accuracy</td>
          <td>Follow-up prompt</td>
          <td>Switched to CCSD with large basis</td>
      </tr>
      <tr>
          <td>Chemical NER</td>
          <td>3 examples + new text</td>
          <td>Extracted compound names (with some gaps)</td>
      </tr>
      <tr>
          <td>Molecule drawing</td>
          <td>&ldquo;Load caffeine from SMILES, draw it&rdquo;</td>
          <td>Correct RDKit rendering</td>
      </tr>
      <tr>
          <td>Gaussian input file</td>
          <td>Function with docstring</td>
          <td>Complete file writer with B3LYP/6-31G(d)</td>
      </tr>
      <tr>
          <td>PDB analysis</td>
          <td>Natural language description</td>
          <td>Downloaded structure and computed <a href="https://en.wikipedia.org/wiki/Radius_of_gyration">radius of gyration</a></td>
      </tr>
  </tbody>
</table>
<p>The authors note that Codex generates correct code at about a 30% rate on a single attempt for standard problems, improving to above 50% when multiple solutions are tried. Mistakes tend to occur when complex algorithms are requested with little specificity, and the code rarely has syntax errors but may fail in obvious ways (missing imports, wrong data types).</p>
<h2 id="challenges-access-correctness-and-bias">Challenges: Access, Correctness, and Bias</h2>
<p>The paper identifies three ongoing challenges:</p>
<p><strong>Access and price.</strong> Advanced models from OpenAI were, at the time of writing, limited to early testers. Per-query costs (1-3 cents for GPT-3) would become prohibitive at the scale needed for parsing academic literature or supporting medium-sized courses. The authors advocate for open-source models and equitable deployment by researchers with computational resources.</p>
<p><strong>Correctness.</strong> Code generation does not guarantee correctness. The authors raise a subtle point: Codex may produce code that executes successfully but does not follow best scientific practice for a particular computational task. Over-reliance on AI-generated code without verification could erode trust in scientific software. However, they argue that strategies for assessing code correctness apply equally to human-written and AI-generated code.</p>
<p><strong>Fairness and bias.</strong> The authors flag several concerns: AI-generated code trained on its own outputs could narrow the range of packages, methods, or programming languages used in chemistry. They observed Codex&rsquo;s preference for Python and for specific popular libraries (e.g., defaulting to <a href="https://en.wikipedia.org/wiki/PSI_(computational_chemistry)">Psi4</a> for single-point energy calculations). GPT-3 has also been shown to reflect racism, sexism, and other biases present in its training data.</p>
<h2 id="implications-for-research-and-education">Implications for Research and Education</h2>
<p>The authors conclude with an optimistic but measured outlook:</p>
<ul>
<li><strong>For research</strong>: NLP code generation will increase accessibility of software tools and expand what a single research group can accomplish. Better tools have historically not reduced the need for scientists but expanded the complexity of problems that can be tackled.</li>
<li><strong>For programming skills</strong>: Using Codex will make chemists better programmers, not worse. The process of crafting prompts, mentally checking outputs, testing on sample inputs, and iterating develops algorithmic thinking. The authors report discovering chemistry software libraries they would not have found otherwise through iterative prompt creation.</li>
<li><strong>For education</strong>: Instructors should rethink programming assignments. The authors suggest moving toward more difficult compound assignments, treating code exercises as laboratory explorations of scientific concepts rather than syntax drills, and aligning coursework with the tools students will have access to in their careers.</li>
<li><strong>For accessibility</strong>: NLP models can reduce barriers for non-native English speakers (though accuracy with non-English prompts was not fully explored) and for users who have difficulty with keyboard-and-mouse interfaces (via voice control).</li>
</ul>
<p>The paper acknowledges that these capabilities were, in early 2022, just beginning, with Codex being the first capable code-generation model. Already at the time of writing, models surpassing GPT-3 in language tasks had appeared, and models matching GPT-3 with 1/20th the parameters had been demonstrated.</p>
<hr>
<h2 id="reproducibility-details">Reproducibility Details</h2>
<p>This is a perspective paper with qualitative demonstrations rather than a reproducible experimental study. The authors provide all prompts and multiple responses in the ESI.</p>
<h3 id="data">Data</h3>
<p>All prompts and code outputs are provided in the Electronic Supplementary Information (ESI) available from the RSC.</p>
<h3 id="algorithms">Algorithms</h3>
<p>The paper does not introduce new algorithms. It evaluates existing models (GPT-3, Codex) on chemistry-related code generation tasks.</p>
<h3 id="models">Models</h3>
<table>
  <thead>
      <tr>
          <th>Model</th>
          <th>Provider</th>
          <th>Access</th>
      </tr>
  </thead>
  <tbody>
      <tr>
          <td>GPT-3</td>
          <td>OpenAI</td>
          <td>API access (commercial)</td>
      </tr>
      <tr>
          <td>Codex</td>
          <td>OpenAI</td>
          <td>Early tester program (2021)</td>
      </tr>
      <tr>
          <td>GPT-Neo</td>
          <td>EleutherAI</td>
          <td>Open source</td>
      </tr>
  </tbody>
</table>
<h3 id="evaluation">Evaluation</h3>
<p>No formal metrics are reported for the chemistry demonstrations. The authors cite the Codex paper&rsquo;s reported ~30% pass rate on single attempts and &gt;50% with multiple attempts on standard programming problems.</p>
<h3 id="hardware">Hardware</h3>
<p>No hardware requirements are specified for the demonstrations (API-based inference).</p>
<h3 id="artifacts">Artifacts</h3>
<table>
  <thead>
      <tr>
          <th>Artifact</th>
          <th>Type</th>
          <th>License</th>
          <th>Notes</th>
      </tr>
  </thead>
  <tbody>
      <tr>
          <td><a href="https://github.com/whitead/marvis">MARVIS</a></td>
          <td>Code</td>
          <td>MIT</td>
          <td>Voice-controlled MD analysis using GPT-3</td>
      </tr>
  </tbody>
</table>
<hr>
<h2 id="paper-information">Paper Information</h2>
<p><strong>Citation</strong>: Hocky, G. M., &amp; White, A. D. (2022). Natural language processing models that automate programming will transform chemistry research and teaching. <em>Digital Discovery</em>, 1(2), 79-83. <a href="https://doi.org/10.1039/d1dd00009h">https://doi.org/10.1039/d1dd00009h</a></p>
<div class="highlight"><pre tabindex="0" style="color:#f8f8f2;background-color:#272822;-moz-tab-size:4;-o-tab-size:4;tab-size:4;"><code class="language-bibtex" data-lang="bibtex"><span style="display:flex;"><span><span style="color:#a6e22e">@article</span>{hocky2022natural,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">title</span>=<span style="color:#e6db74">{Natural language processing models that automate programming will transform chemistry research and teaching}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">author</span>=<span style="color:#e6db74">{Hocky, Glen M. and White, Andrew D.}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">journal</span>=<span style="color:#e6db74">{Digital Discovery}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">volume</span>=<span style="color:#e6db74">{1}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">number</span>=<span style="color:#e6db74">{2}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">pages</span>=<span style="color:#e6db74">{79--83}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">year</span>=<span style="color:#e6db74">{2022}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">publisher</span>=<span style="color:#e6db74">{Royal Society of Chemistry}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">doi</span>=<span style="color:#e6db74">{10.1039/d1dd00009h}</span>
</span></span><span style="display:flex;"><span>}
</span></span></code></pre></div>]]></content:encoded></item><item><title>Genetic Algorithms as Baselines for Molecule Generation</title><link>https://hunterheidenreich.com/notes/chemistry/molecular-design/generation/search-based/genetic-algorithms-molecule-generation-baselines/</link><pubDate>Mon, 23 Mar 2026 00:00:00 +0000</pubDate><guid>https://hunterheidenreich.com/notes/chemistry/molecular-design/generation/search-based/genetic-algorithms-molecule-generation-baselines/</guid><description>Genetic algorithms outperform many deep learning methods for molecule generation. Tripp and Hernández-Lobato propose the GA criterion.</description><content:encoded><![CDATA[<h2 id="a-position-paper-on-molecular-generation-baselines">A Position Paper on Molecular Generation Baselines</h2>
<p>This is a <strong>Position</strong> paper that argues genetic algorithms (GAs) are underused and underappreciated as baselines in the molecular generation community. The primary contribution is empirical evidence that a simple GA implementation (MOL_GA) matches or outperforms many sophisticated deep learning methods on standard benchmarks. The authors propose the &ldquo;GA criterion&rdquo; as a minimum bar for evaluating new molecular generation algorithms.</p>
<h2 id="why-molecular-generation-may-be-easier-than-assumed">Why Molecular Generation May Be Easier Than Assumed</h2>
<p>Drug discovery is fundamentally a molecular generation task, and many machine learning methods have been proposed for it (Du et al., 2022). The problem has many variants, from unconditional generation of novel molecules to directed optimization of specific molecular properties.</p>
<p>The authors observe that generating valid molecules is, in some respects, straightforward. The rules governing molecular validity are well-defined bond constraints that can be checked using standard cheminformatics software like <a href="https://en.wikipedia.org/wiki/RDKit">RDKit</a>. This means new molecules can be generated simply by adding, removing, or substituting fragments of known molecules. When applied iteratively, this is exactly what a genetic algorithm does. Despite this, many papers in the field propose complex deep learning methods without adequately comparing to simple GA baselines.</p>
<h2 id="the-ga-criterion-for-evaluating-new-methods">The GA Criterion for Evaluating New Methods</h2>
<p>The core proposal is the <strong>GA criterion</strong>: new methods in molecular generation should offer some clear advantage over genetic algorithms. This advantage can be:</p>
<ul>
<li><strong>Empirical</strong>: outperforming GAs on relevant benchmarks</li>
<li><strong>Conceptual</strong>: identifying and overcoming a specific limitation of randomly modifying known molecules</li>
</ul>
<p>The authors argue that the current state of molecular generation research reflects poor empirical practices, where comprehensive baseline evaluation is treated as optional rather than essential.</p>
<h2 id="genetic-algorithm-framework-and-benchmark-experiments">Genetic Algorithm Framework and Benchmark Experiments</h2>
<h3 id="how-genetic-algorithms-work-for-molecules">How Genetic Algorithms Work for Molecules</h3>
<p>GAs operate through the following iterative procedure:</p>
<ol>
<li>Start with an initial population $P$ of molecules</li>
<li>Sample a subset $S \subseteq P$ from the population (possibly biased toward better molecules)</li>
<li>Generate new molecules $N$ from $S$ via mutation and crossover operations</li>
<li>Select a new population $P&rsquo;$ from $P \cup N$ (e.g., keep the highest-scoring molecules)</li>
<li>Set $P \leftarrow P&rsquo;$ and repeat from step 2</li>
</ol>
<p>The MOL_GA implementation uses:</p>
<ul>
<li><strong>Quantile-based sampling</strong> (step 2): molecules are sampled from the top quantiles of the population using a log-uniform distribution over quantile thresholds:</li>
</ul>
<p>$$
u \sim \mathcal{U}[-3, 0], \quad \epsilon = 10^{u}
$$</p>
<p>A molecule is drawn uniformly from the top $\epsilon$ fraction of the population.</p>
<ul>
<li><strong>Mutation and crossover</strong> (step 3): graph-based operations from <a href="/notes/chemistry/molecular-design/generation/search-based/graph-based-genetic-algorithm-chemical-space/">Jensen (2019)</a>, as implemented in the <a href="/notes/chemistry/molecular-design/generation/evaluation/guacamol-benchmarking-de-novo-molecular-design/">GuacaMol benchmark (Brown et al., 2019)</a></li>
<li><strong>Greedy population selection</strong> (step 4): molecules with the highest scores are retained</li>
</ul>
<h3 id="unconditional-generation-on-zinc-250k">Unconditional Generation on ZINC 250K</h3>
<p>The first experiment evaluates unconditional molecule generation, where the task is to produce novel, valid, and unique molecules distinct from a reference set (ZINC 250K). Success is measured by validity, novelty (at 10,000 generated molecules), and uniqueness.</p>
<table>
  <thead>
      <tr>
          <th>Method</th>
          <th>Paper</th>
          <th>Validity</th>
          <th>Novelty@10k</th>
          <th>Uniqueness</th>
      </tr>
  </thead>
  <tbody>
      <tr>
          <td>JT-VAE</td>
          <td>Jin et al. (2018)</td>
          <td>99.8%</td>
          <td>100%</td>
          <td>100%</td>
      </tr>
      <tr>
          <td>GCPN</td>
          <td>You et al. (2018)</td>
          <td>100%</td>
          <td>100%</td>
          <td>99.97%</td>
      </tr>
      <tr>
          <td><a href="/notes/chemistry/molecular-design/generation/rl-tuned/molecularrnn-graph-generation-optimized-properties/">MolecularRNN</a></td>
          <td>Popova et al. (2019)</td>
          <td>100%</td>
          <td>100%</td>
          <td>99.89%</td>
      </tr>
      <tr>
          <td>Graph NVP</td>
          <td>Madhawa et al. (2019)</td>
          <td>100%</td>
          <td>100%</td>
          <td>94.80%</td>
      </tr>
      <tr>
          <td>Graph AF</td>
          <td>Shi et al. (2020)</td>
          <td>100%</td>
          <td>100%</td>
          <td>99.10%</td>
      </tr>
      <tr>
          <td>MoFlow</td>
          <td>Zang and Wang (2020)</td>
          <td>100%</td>
          <td>100%</td>
          <td>99.99%</td>
      </tr>
      <tr>
          <td>GraphCNF</td>
          <td>Lippe and Gavves (2020)</td>
          <td>96.35%</td>
          <td>99.98%</td>
          <td>99.98%</td>
      </tr>
      <tr>
          <td>Graph DF</td>
          <td>Luo et al. (2021)</td>
          <td>100%</td>
          <td>100%</td>
          <td>99.16%</td>
      </tr>
      <tr>
          <td>ModFlow</td>
          <td>Verma et al. (2022)</td>
          <td>98.1%</td>
          <td>100%</td>
          <td>99.3%</td>
      </tr>
      <tr>
          <td>GraphEBM</td>
          <td>Liu et al. (2021)</td>
          <td>99.96%</td>
          <td>100%</td>
          <td>98.79%</td>
      </tr>
      <tr>
          <td>AddCarbon</td>
          <td>Renz et al. (2019)</td>
          <td>100%</td>
          <td>99.94%</td>
          <td>99.86%</td>
      </tr>
      <tr>
          <td>MOL_GA</td>
          <td>(this paper)</td>
          <td>99.76%</td>
          <td>99.94%</td>
          <td>98.60%</td>
      </tr>
  </tbody>
</table>
<p>All methods perform near 100% on all metrics, demonstrating that unconditional molecule generation is not a particularly discriminative benchmark. The authors note that generation speed (molecules per second) is an important missing dimension from these comparisons, where simple methods like GAs have a clear advantage.</p>
<h3 id="molecule-optimization-on-the-pmo-benchmark">Molecule Optimization on the PMO Benchmark</h3>
<p>The second experiment evaluates directed molecule optimization on the <a href="/notes/chemistry/molecular-design/generation/evaluation/pmo-sample-efficient-molecular-optimization/">Practical Molecular Optimization (PMO) benchmark (Gao et al., 2022)</a>, which measures the ability to find molecules optimizing a scalar objective function $f: \mathcal{M} \mapsto \mathbb{R}$ with a budget of 10,000 evaluations.</p>
<p>A key insight is that previous GA implementations in PMO used large generation sizes ($\approx 100$), which limits the number of improvement iterations. The authors set the generation size to 5, allowing approximately 2,000 iterations of improvement within the same evaluation budget.</p>
<table>
  <thead>
      <tr>
          <th>Task</th>
          <th><a href="/notes/chemistry/molecular-design/generation/rl-tuned/reinvent-deep-rl-molecular-design/">REINVENT</a></th>
          <th>Graph GA</th>
          <th>MOL_GA</th>
      </tr>
  </thead>
  <tbody>
      <tr>
          <td>albuterol_similarity</td>
          <td>0.882 +/- 0.006</td>
          <td>0.838 +/- 0.016</td>
          <td><strong>0.896 +/- 0.035</strong></td>
      </tr>
      <tr>
          <td>amlodipine_mpo</td>
          <td>0.635 +/- 0.035</td>
          <td>0.661 +/- 0.020</td>
          <td><strong>0.688 +/- 0.039</strong></td>
      </tr>
      <tr>
          <td>celecoxib_rediscovery</td>
          <td><strong>0.713 +/- 0.067</strong></td>
          <td>0.630 +/- 0.097</td>
          <td>0.567 +/- 0.083</td>
      </tr>
      <tr>
          <td>drd2</td>
          <td>0.945 +/- 0.007</td>
          <td><strong>0.964 +/- 0.012</strong></td>
          <td>0.936 +/- 0.016</td>
      </tr>
      <tr>
          <td>fexofenadine_mpo</td>
          <td>0.784 +/- 0.006</td>
          <td>0.760 +/- 0.011</td>
          <td><strong>0.825 +/- 0.019</strong></td>
      </tr>
      <tr>
          <td>isomers_c9h10n2o2pf2cl</td>
          <td>0.642 +/- 0.054</td>
          <td>0.719 +/- 0.047</td>
          <td><strong>0.865 +/- 0.012</strong></td>
      </tr>
      <tr>
          <td>sitagliptin_mpo</td>
          <td>0.021 +/- 0.003</td>
          <td>0.433 +/- 0.075</td>
          <td><strong>0.582 +/- 0.040</strong></td>
      </tr>
      <tr>
          <td>zaleplon_mpo</td>
          <td>0.358 +/- 0.062</td>
          <td>0.346 +/- 0.032</td>
          <td><strong>0.519 +/- 0.029</strong></td>
      </tr>
      <tr>
          <td><strong>Sum (23 tasks)</strong></td>
          <td>14.196</td>
          <td>13.751</td>
          <td><strong>14.708</strong></td>
      </tr>
      <tr>
          <td><strong>Rank</strong></td>
          <td>2</td>
          <td>3</td>
          <td><strong>1</strong></td>
      </tr>
  </tbody>
</table>
<p>MOL_GA achieves the highest aggregate score across all 23 PMO tasks, outperforming both the previous best GA (Graph GA) and the previous best overall method (REINVENT). The authors attribute this partly to the tuning of the baselines in PMO rather than MOL_GA being an especially strong method, since MOL_GA is essentially the same algorithm as Graph GA with different hyperparameters.</p>
<h2 id="implications-for-molecular-generation-research">Implications for Molecular Generation Research</h2>
<p>The key findings and arguments are:</p>
<ol>
<li>
<p><strong>GAs match or outperform deep learning methods</strong> on standard molecular generation benchmarks, both for unconditional generation and directed optimization.</p>
</li>
<li>
<p><strong>Hyperparameter choices matter significantly</strong>: MOL_GA&rsquo;s strong performance on PMO comes partly from using a smaller generation size (5 vs. ~100), which allows more iterations of refinement within the same evaluation budget.</p>
</li>
<li>
<p><strong>The GA criterion should be enforced in peer review</strong>: new molecular generation methods should demonstrate a clear advantage over GAs, whether empirical or conceptual.</p>
</li>
<li>
<p><strong>Deep learning methods may implicitly do what GAs do explicitly</strong>: many generative models are trained on datasets of known molecules, so the novel molecules they produce may simply be variants of their training data. The authors consider this an important direction for future investigation.</p>
</li>
<li>
<p><strong>Poor empirical practices are widespread</strong>: the paper argues that many experiments in molecule generation are conducted with an explicit desired outcome (that the novel algorithm is the best), leading to inadequate baseline comparisons.</p>
</li>
</ol>
<p>The authors are careful to note that this result should not be interpreted as GAs being exceptional algorithms. Rather, it is an indication that more complex methods have made surprisingly little progress beyond what simple heuristic search can achieve.</p>
<hr>
<h2 id="reproducibility-details">Reproducibility Details</h2>
<h3 id="data">Data</h3>
<table>
  <thead>
      <tr>
          <th>Purpose</th>
          <th>Dataset</th>
          <th>Size</th>
          <th>Notes</th>
      </tr>
  </thead>
  <tbody>
      <tr>
          <td>Unconditional generation</td>
          <td>ZINC 250K</td>
          <td>250,000 molecules</td>
          <td>Reference set for novelty evaluation</td>
      </tr>
      <tr>
          <td>Directed optimization</td>
          <td>PMO benchmark</td>
          <td>23 tasks</td>
          <td>10,000 evaluation budget per task</td>
      </tr>
  </tbody>
</table>
<h3 id="algorithms">Algorithms</h3>
<ul>
<li><strong>GA implementation</strong>: MOL_GA package, using graph-based mutation and crossover from Jensen (2019) via the GuacaMol implementation</li>
<li><strong>Generation size</strong>: 5 molecules per iteration (allowing ~2,000 iterations with 10,000 evaluations)</li>
<li><strong>Population selection</strong>: Greedy (highest-scoring molecules retained)</li>
<li><strong>Sampling</strong>: Quantile-based with log-uniform distribution over quantile thresholds</li>
</ul>
<h3 id="evaluation">Evaluation</h3>
<table>
  <thead>
      <tr>
          <th>Metric</th>
          <th>Benchmark</th>
          <th>Notes</th>
      </tr>
  </thead>
  <tbody>
      <tr>
          <td>Validity, Novelty@10k, Uniqueness</td>
          <td>ZINC 250K unconditional</td>
          <td>Calculated using <a href="/notes/chemistry/molecular-design/generation/evaluation/molecular-sets-moses/">MOSES package</a></td>
      </tr>
      <tr>
          <td>AUC top-10 scores</td>
          <td>PMO benchmark</td>
          <td>23 optimization tasks with 10,000 evaluation budget</td>
      </tr>
  </tbody>
</table>
<h3 id="hardware">Hardware</h3>
<p>The paper does not specify hardware requirements. Given that GAs are computationally lightweight compared to deep learning methods, standard CPU hardware is likely sufficient.</p>
<h3 id="artifacts">Artifacts</h3>
<table>
  <thead>
      <tr>
          <th>Artifact</th>
          <th>Type</th>
          <th>License</th>
          <th>Notes</th>
      </tr>
  </thead>
  <tbody>
      <tr>
          <td><a href="https://github.com/AustinT/mol_ga">MOL_GA</a></td>
          <td>Code</td>
          <td>MIT</td>
          <td>Python package for molecular genetic algorithms</td>
      </tr>
      <tr>
          <td><a href="https://pypi.org/project/mol-ga/">MOL_GA on PyPI</a></td>
          <td>Code</td>
          <td>MIT</td>
          <td>pip-installable package</td>
      </tr>
  </tbody>
</table>
<hr>
<h2 id="paper-information">Paper Information</h2>
<p><strong>Citation</strong>: Tripp, A., &amp; Hernández-Lobato, J. M. (2023). Genetic algorithms are strong baselines for molecule generation. <em>arXiv preprint arXiv:2310.09267</em>. <a href="https://arxiv.org/abs/2310.09267">https://arxiv.org/abs/2310.09267</a></p>
<p><strong>Publication</strong>: arXiv preprint, 2023</p>
<p><strong>Additional Resources</strong>:</p>
<ul>
<li><a href="https://github.com/AustinT/mol_ga">MOL_GA Python Package (GitHub)</a></li>
<li><a href="https://pypi.org/project/mol-ga/">MOL_GA on PyPI</a></li>
</ul>
<h2 id="citation">Citation</h2>
<div class="highlight"><pre tabindex="0" style="color:#f8f8f2;background-color:#272822;-moz-tab-size:4;-o-tab-size:4;tab-size:4;"><code class="language-bibtex" data-lang="bibtex"><span style="display:flex;"><span><span style="color:#a6e22e">@article</span>{tripp2023genetic,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">title</span>=<span style="color:#e6db74">{Genetic algorithms are strong baselines for molecule generation}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">author</span>=<span style="color:#e6db74">{Tripp, Austin and Hern{\&#39;a}ndez-Lobato, Jos{\&#39;e} Miguel}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">journal</span>=<span style="color:#e6db74">{arXiv preprint arXiv:2310.09267}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">year</span>=<span style="color:#e6db74">{2023}</span>
</span></span><span style="display:flex;"><span>}
</span></span></code></pre></div>]]></content:encoded></item><item><title>SELFIES and the Future of Molecular String Representations</title><link>https://hunterheidenreich.com/notes/chemistry/molecular-representations/notations/selfies-2022/</link><pubDate>Tue, 02 Dec 2025 00:00:00 +0000</pubDate><guid>https://hunterheidenreich.com/notes/chemistry/molecular-representations/notations/selfies-2022/</guid><description>Perspective on SELFIES as a 100% robust SMILES alternative, with 16 future research directions for molecular AI.</description><content:encoded><![CDATA[<h2 id="position-a-roadmap-for-robust-chemical-languages">Position: A Roadmap for Robust Chemical Languages</h2>
<p>This is a <strong>Position</strong> paper (perspective) that proposes a research agenda for molecular representations in AI. It reviews the evolution of chemical notation over 250 years and argues for extending SELFIES-style robust representations beyond traditional organic chemistry into polymers, crystals, reactions, and other complex chemical systems.</p>
<h2 id="the-generative-bottleneck-in-traditional-representations">The Generative Bottleneck in Traditional Representations</h2>
<p>While SMILES has been the standard molecular representation since 1988, its fundamental weakness for machine learning is well-established: randomly generated SMILES strings are often invalid. The motivation is twofold:</p>
<ol>
<li><strong>Current problem</strong>: Traditional representations (SMILES, <a href="/notes/chemistry/molecular-representations/notations/inchi-2013/">InChI</a>, DeepSMILES) lack 100% robustness; random mutations or generations can produce invalid strings, limiting their use in generative AI models.</li>
<li><strong>Future opportunity</strong>: SELFIES solved this for small organic molecules, but many important chemical domains (polymers, crystals, reactions) still lack robust representations, creating a bottleneck for AI-driven discovery in these areas.</li>
</ol>
<h2 id="16-concrete-research-directions-for-selfies">16 Concrete Research Directions for SELFIES</h2>
<p>The novelty is in the comprehensive research roadmap. The authors propose 16 concrete research projects organized around key themes:</p>
<ul>
<li><strong>Domain extension</strong>: Includes metaSELFIES for learning graph rules directly from data, BigSELFIES for stochastic polymers, and crystal structures via labeled quotient graphs.</li>
<li><strong>Chemical reactions</strong>: Robust reaction representations that enforce conservation laws.</li>
<li><strong>Programming perspective</strong>: Treating molecular representations as programming languages, potentially achieving Turing-completeness.</li>
<li><strong>Benchmarking</strong>: Systematic comparisons across representation formats.</li>
<li><strong>Interpretability</strong>: Understanding how humans and machines actually learn from different representations.</li>
</ul>
<h2 id="evidence-from-generative-case-studies">Evidence from Generative Case Studies</h2>
<p>This perspective paper includes case studies:</p>
<ol>
<li>
<p><strong>Pasithea (Deep Molecular Dreaming)</strong>: A generative model that first learns to predict a chemical property from a one-hot encoded SELFIES, then freezes the network weights and uses gradient descent on the one-hot input encoding to optimize molecular properties (logP). The target property increases or decreases nearly monotonically, demonstrating that the model has learned meaningful structure-property relationships from the SELFIES representation.</p>
</li>
<li>
<p><strong>DECIMER and STOUT</strong>: DECIMER (Deep lEarning for Chemical ImagE Recognition) is an image-to-structure tool, and STOUT (SMILES-TO-IUPAC-name Translator) translates between IUPAC names and molecular string representations. Both show improved performance when using SELFIES as an intermediate representation. STOUT internally converts SMILES to SELFIES before processing and decodes predicted SELFIES back to SMILES. These results suggest SELFIES provides a more learnable internal representation for sequence-to-sequence models.</p>
</li>
</ol>
<h2 id="strategic-outcomes-and-future-vision">Strategic Outcomes and Future Vision</h2>
<p>The paper establishes robust representations as a fundamental bottleneck in computational chemistry and proposes a clear path forward:</p>
<p><strong>Key outcomes</strong>:</p>
<ul>
<li>Identification of 16 concrete research projects spanning domain extension, benchmarking, and interpretability</li>
<li>Evidence that SELFIES enables capabilities (like smooth property optimization) impossible with traditional formats</li>
<li>Framework for thinking about molecular representations as programming languages</li>
</ul>
<p><strong>Strategic impact</strong>: The proposed extensions could enable new applications across drug discovery (efficient exploration beyond small molecules), materials design (systematic crystal structure discovery), synthesis planning (better reaction representations), and fundamental research (new ways to understand chemical behavior).</p>
<p><strong>Future vision</strong>: The authors emphasize that robust representations could become a bridge for bidirectional learning between humans and machines, enabling humans to learn new chemical concepts from AI systems.</p>
<h2 id="the-mechanism-of-robustness">The Mechanism of Robustness</h2>
<p>The key difference between SELFIES and other representations lies in how they handle syntax:</p>
<ul>
<li><strong>SMILES/DeepSMILES</strong>: Rely on non-local markers (opening/closing parentheses or ring numbers) that must be balanced. A mutation or random generation can easily break this balance, producing invalid strings.</li>
<li><strong>SELFIES</strong>: Uses a formal grammar (automaton) where derivation rules are entirely local. The critical innovation is <strong>overloading</strong>: a state-modifying symbol like <code>[Branch1]</code> starts a branch and changes the interpretation of the <em>next</em> symbol to represent a numerical parameter (the branch length).</li>
</ul>
<p>This overloading mechanism ensures that any arbitrary sequence of SELFIES tokens can be parsed into a valid molecular graph. The derivation can never fail because every symbol either adds an atom or modifies how subsequent symbols are interpreted.</p>
<h2 id="the-16-research-projects-technical-details">The 16 Research Projects: Technical Details</h2>
<p>This section provides technical details on the proposed research directions:</p>
<h3 id="extending-to-new-domains">Extending to New Domains</h3>
<p><strong>metaSELFIES (Project 1)</strong>: The authors propose learning graph construction rules automatically from data. This could enable robust representations for any graph-based system, from quantum optics to biological networks, without needing domain-specific expertise.</p>
<p><strong>Token Optimization (Project 2)</strong>: SELFIES uses &ldquo;overloading&rdquo; where a symbol&rsquo;s meaning changes based on context. This project would investigate how this affects machine learning performance and whether the approach can be optimized.</p>
<h3 id="handling-complex-molecular-systems">Handling Complex Molecular Systems</h3>
<p><strong>BigSELFIES (Project 3)</strong>: Current representations struggle with large, often random structures like polymers and biomolecules. BigSELFIES would combine hierarchical notation with stochastic building blocks to handle these complex systems where traditional small-molecule representations break down.</p>
<p><strong>Crystal Structures (Projects 4-5)</strong>: Crystals present unique challenges due to their infinite, periodic arrangements. An infinite net cannot be represented by a finite string directly. The proposed approach uses <strong>labeled quotient graphs (LQGs)</strong>, which are finite graphs that uniquely determine a periodic net. However, current SELFIES cannot represent LQGs because they lack symbols for edge directions and edge labels (vector shifts encoding periodicity). Extending SELFIES to handle these structures could enable AI-driven materials design without relying on predefined crystal structures, opening up systematic exploration of theoretical materials space.</p>
<p><strong>Beyond Organic Chemistry (Project 6)</strong>: Transition metals and main-group compounds feature complex bonding that breaks the simple two-center, two-electron model. The solution: use machine learning on large structural databases to automatically learn these complex bonding rules.</p>
<h3 id="chemical-reactions-and-programming-concepts">Chemical Reactions and Programming Concepts</h3>
<p><strong>Reaction Representations (Project 7)</strong>: Moving beyond static molecules to represent chemical transformations. A robust reaction format would enforce conservation laws and could learn reactivity patterns from large reaction datasets, improving synthesis planning.</p>
<h3 id="developing-a-100-robust-programming-language">Developing a 100% Robust Programming Language</h3>
<p><strong>Programming Language Perspective (Projects 8-9)</strong>: An intriguing reframing views molecular representations as programming languages executed by chemical parsers. This opens possibilities for adding loops, logic, and other programming concepts to efficiently describe complex structures. The ambitious goal is a Turing-complete programming language that is also 100% robust. While fascinating, it is worth critically noting that enforcing 100% syntactical robustness inherently restricts grammar flexibility. Can a purely robust string representation realistically describe highly fuzzy, delocalized electron bonds (like in Project 6) without becoming impractically long or collapsing into specialized sub-languages?</p>
<p><strong>Empirical Comparisons (Projects 10-11)</strong>: With multiple representation options (strings, matrices, images), we need systematic comparisons. The proposed benchmarks would go beyond simple validity metrics to focus on real-world design objectives in drug discovery, catalysis, and materials science.</p>
<p><strong>Human Readability (Project 12)</strong>: While SMILES is often called &ldquo;human-readable,&rdquo; this claim lacks scientific validation. The proposed study would test how well humans actually understand different molecular representations.</p>
<p><strong>Machine Learning Perspectives (Projects 13-16)</strong>: These projects explore how machines interpret molecular representations:</p>
<ul>
<li>Training networks to translate between formats to find universal representations</li>
<li>Comparing learning efficiency across different formats</li>
<li>Investigating latent space smoothness in generative models</li>
<li>Visualizing what models actually learn about molecular structure</li>
</ul>
<h2 id="reproducibility-details">Reproducibility Details</h2>
<p>Since this is a position paper outlining future research directions, standard empirical reproducibility metrics do not apply. However, the foundational tools required to pursue the proposed roadmap are open-source.</p>
<table>
  <thead>
      <tr>
          <th>Artifact</th>
          <th>Type</th>
          <th>License</th>
          <th>Notes</th>
      </tr>
  </thead>
  <tbody>
      <tr>
          <td><a href="https://github.com/aspuru-guzik-group/selfies">aspuru-guzik-group/selfies</a></td>
          <td>Code</td>
          <td>Apache-2.0</td>
          <td>Core SELFIES Python library, installable via <code>pip install selfies</code></td>
      </tr>
      <tr>
          <td><a href="https://arxiv.org/abs/2204.00056">arXiv:2204.00056</a></td>
          <td>Paper</td>
          <td>N/A</td>
          <td>Open-access preprint of the published Patterns article</td>
      </tr>
  </tbody>
</table>
<h2 id="paper-information">Paper Information</h2>
<p><strong>Citation</strong>: Krenn, M., Ai, Q., Barthel, S., Carson, N., Frei, A., Frey, N. C., Friederich, P., Gaudin, T., Gayle, A. A., Jablonka, K. M., Lameiro, R. F., Lemm, D., Lo, A., Moosavi, S. M., Nápoles-Duarte, J. M., Nigam, A., Pollice, R., Rajan, K., Schatzschneider, U., &hellip; Aspuru-Guzik, A. (2022). SELFIES and the future of molecular string representations. <em>Patterns</em>, <em>3</em>(10). <a href="https://doi.org/10.1016/j.patter.2022.100588">https://doi.org/10.1016/j.patter.2022.100588</a></p>
<p><strong>Publication</strong>: Patterns 2022</p>
<div class="highlight"><pre tabindex="0" style="color:#f8f8f2;background-color:#272822;-moz-tab-size:4;-o-tab-size:4;tab-size:4;"><code class="language-bibtex" data-lang="bibtex"><span style="display:flex;"><span><span style="color:#a6e22e">@article</span>{Krenn2022,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">title</span> = <span style="color:#e6db74">{SELFIES and the future of molecular string representations}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">volume</span> = <span style="color:#e6db74">{3}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">ISSN</span> = <span style="color:#e6db74">{2666-3899}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">url</span> = <span style="color:#e6db74">{http://dx.doi.org/10.1016/j.patter.2022.100588}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">DOI</span> = <span style="color:#e6db74">{10.1016/j.patter.2022.100588}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">number</span> = <span style="color:#e6db74">{10}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">journal</span> = <span style="color:#e6db74">{Patterns}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">publisher</span> = <span style="color:#e6db74">{Elsevier BV}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">author</span> = <span style="color:#e6db74">{Krenn, Mario and Ai, Qianxiang and Barthel, Senja and Carson, Nessa and Frei, Angelo and Frey, Nathan C. and Friederich, Pascal and Gaudin, Théophile and Gayle, Alberto Alexander and Jablonka, Kevin Maik and Lameiro, Rafael F. and Lemm, Dominik and Lo, Alston and Moosavi, Seyed Mohamad and Nápoles-Duarte, José Manuel and Nigam, AkshatKumar and Pollice, Robert and Rajan, Kohulan and Schatzschneider, Ulrich and Schwaller, Philippe and Skreta, Marta and Smit, Berend and Strieth-Kalthoff, Felix and Sun, Chong and Tom, Gary and von Rudorff, Guido Falk and Wang, Andrew and White, Andrew and Young, Adamo and Yu, Rose and Aspuru-Guzik, Alán}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">year</span> = <span style="color:#e6db74">{2022}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">month</span> = oct,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">pages</span> = <span style="color:#e6db74">{100588}</span>
</span></span><span style="display:flex;"><span>}
</span></span></code></pre></div><p><strong>Additional Resources</strong>:</p>
<ul>
<li><a href="/notes/chemistry/molecular-representations/notations/selfies-original-paper/">Original SELFIES Paper</a></li>
<li><a href="/notes/chemistry/molecular-representations/notations/selfies/">SELFIES Overview</a></li>
</ul>
]]></content:encoded></item><item><title>How to Fold Graciously: Levinthal's Paradox (1969)</title><link>https://hunterheidenreich.com/notes/biology/computational-biology/fold-graciously/</link><pubDate>Mon, 08 Sep 2025 00:00:00 +0000</pubDate><guid>https://hunterheidenreich.com/notes/biology/computational-biology/fold-graciously/</guid><description>A perspective paper defining the Grand Challenge of protein folding: distinguishing kinetic pathways from thermodynamic endpoints.</description><content:encoded><![CDATA[<h2 id="what-kind-of-paper-is-this">What kind of paper is this?</h2>
<p>This is technically a transcription of a conference talk, not a paper Levinthal wrote himself. The proceedings page credits &ldquo;Notes by: A. Rawitch, Retranscribed: B. Krantz&rdquo;, meaning what we have is a third-party record of an oral presentation Levinthal gave at the 1969 Mössbauer Spectroscopy in Biological Systems meeting at Allerton House, Illinois. This explains the informal, conversational register and the attached Q&amp;A discussion.</p>
<p>In terms of contribution type, it functions as a <strong>Position</strong> paper (with Theory and Discovery elements):</p>
<ul>
<li><strong>Position</strong>: Defines a &ldquo;Grand Challenge&rdquo; and argues for a conceptual shift in how we view biomolecular assembly</li>
<li><strong>Theory</strong>: Uses formal combinatorial arguments to establish the bounds of the search space ($10^{300}$ configurations)</li>
<li><strong>Discovery</strong>: Uses experimental data on alkaline phosphatase to validate the kinetic hypothesis</li>
</ul>
<h2 id="what-is-the-motivation">What is the motivation?</h2>
<p><strong>The Central Question</strong>: How does a protein choose one unique structure out of a hyper-astronomical number of possibilities in a biological timeframe (seconds)?</p>
<p>Levinthal provides a &ldquo;back-of-the-envelope&rdquo; derivation to define the problem scope:</p>
<ol>
<li><strong>Degrees of Freedom:</strong> A generic, unrestricted protein with 2,000 atoms would possess ~6,000 degrees of freedom. However, physical constraints (specifically the planar peptide bond) reduce this significantly. For a 150-amino acid protein, these constraints lower the complexity to ~450 degrees of freedom (300 rotations, 150 bond angles).</li>
<li><strong>The Combinatorial Explosion:</strong> Even with conservative estimates, this results in $10^{300}$ possible conformations.</li>
<li><strong>The Time Constraint:</strong> Since proteins fold in seconds, Levinthal argues they can sample at most <strong>$10^8$ conformations</strong> (&ldquo;postulating a minimum time from one conformation to another&rdquo;) before stabilizing. Against $10^{300}$ possibilities, this search effectively covers 0% of the space, proving the impossibility of random search.</li>
</ol>
<blockquote>
<p><strong>The Insight:</strong> The existence of folded proteins proves the <strong>impossibility of random global search</strong>. The system <em>must</em> be guided.</p></blockquote>
<h2 id="what-is-the-novelty-here">What is the novelty here?</h2>
<p><strong>Core Contribution</strong>: Levinthal reframes folding from a thermodynamic problem (seeking the absolute global minimum) to a <strong>Kinetic Control</strong> problem. He argues the native state is a &ldquo;metastable&rdquo; energy well found quickly by a specific pathway, which can differ from the system&rsquo;s lowest possible energy state.</p>
<h3 id="the-pathway-dependence-hypothesis">The Pathway Dependence Hypothesis</h3>
<p>The key insights of kinetic control:</p>
<ul>
<li><strong>Nucleation:</strong> The process is &ldquo;speeded and guided by the rapid formation of local interactions&rdquo;</li>
<li><strong>Pathway Constraints:</strong> Local amino acid sequences form stable interactions and serve as nucleation points in the folding process, restricting the conformational search space</li>
<li><strong>The &ldquo;Metastable&rdquo; State:</strong> The final structure represents a &ldquo;metastable state&rdquo; in a sufficiently deep energy well that is <em>kinetically accessible</em> via the folding pathway, independent of the global energy minimum. Think of a ball that rolls into a valley on the side of a hill and stays there: it is not in the lowest valley on the entire landscape, but it is stable enough that it never escapes.</li>
</ul>















<figure class="post-figure center ">
    <img src="/img/notes/folding-funnel.webp"
         alt="The protein folding energy landscape funnel, showing many unfolded states at high energy converging through multiple pathways to the native folded state at the bottom of the funnel"
         title="The protein folding energy landscape funnel, showing many unfolded states at high energy converging through multiple pathways to the native folded state at the bottom of the funnel"
         
         
         loading="lazy"
         class="post-image">
    
    <figcaption class="post-caption">The Energy Landscape Funnel: The modern resolution to Levinthal&rsquo;s Paradox. While Levinthal envisioned a single guided pathway, the &lsquo;funnel&rsquo; model (Wolynes, Dill) shows that many different pathways can lead to the same native state basin. The roughness of the funnel surface represents local energy minima (kinetic traps) that can slow folding.</figcaption>
    
</figure>

<h2 id="what-experiments-were-performed">What experiments were performed?</h2>
<p>To support the pathway hypothesis, Levinthal cites work on <strong>Alkaline Phosphatase</strong> (MW ~40,000), utilizing its property as a dimer of two identical subunits:</p>
<ul>
<li><strong>Renaturation Window:</strong> The wild-type enzyme refolds optimally at 37°C. However, mutants were isolated that only produce active enzyme (and renature) at temperatures <em>below</em> 37°C.</li>
<li><strong>Stability vs. Formation:</strong> Crucially, once folded, both the wild-type and mutant enzymes are stable up to 90°C.</li>
<li><strong>The Rate-Limiting Step:</strong> Levinthal notes that the rate-limiting step for activity is the <strong>formation of the dimer</strong> from monomers. This proves that the <em>order of assembly</em> (kinetic pathway) dictates the final structure, distinct from the final structure&rsquo;s thermodynamic stability.</li>
</ul>
<p>The talk concluded with a short motion picture Levinthal showed live, illustrating polypeptide synthesis and &ldquo;the process of then forming a desired interaction via the most favored energy path as displayed on the computer controlled oscilloscope.&rdquo;</p>
<p>The Q&amp;A discussion following the talk includes one exchange directly relevant to the folding argument: when asked whether a protein is ever truly unfolded (devoid of all secondary and tertiary structure), Levinthal answered that both physical measurements and synthetic polypeptide work suggest yes. The other exchanges concerned the tangent formula for x-ray crystallographic phase refinement and whether computed structures had been tested for thermal perturbations.</p>
<h2 id="what-outcomesconclusions">What outcomes/conclusions?</h2>
<h3 id="key-finding">Key Finding</h3>
<p>The mutant experiments serve as the &ldquo;smoking gun&rdquo;: a protein seeking a global thermodynamic minimum would fold spontaneously at any temperature where the final state is stable (up to 90°C). The fact that mutants require specific lower temperatures for <em>formation</em> (while remaining stable at high temperatures once formed) proves that the <strong>kinetic pathway</strong> determines the outcome alongside the thermodynamic endpoint.</p>
<h3 id="broader-implications">Broader Implications</h3>
<p>Levinthal explicitly asks: &ldquo;Is a unique folding necessary for any random 150-amino acid sequence?&rdquo; and answers &ldquo;Probably not.&rdquo; He supports this by noting the difficulty many researchers face in attempting to crystallize proteins, suggesting that not all sequences produce stably folded structures.</p>
<p>He concludes by connecting these computational models to <strong>Mössbauer spectroscopy</strong>, suggesting that these computational studies may help in understanding how small perturbations of polypeptide structures affect the Mössbauer nucleus (a reminder of the specific conference context where this perspective was delivered).</p>
<h3 id="connection-to-modern-work">Connection to Modern Work</h3>
<p>Levinthal&rsquo;s arguments remain relevant context for modern computational protein folding:</p>
<ul>
<li><strong>Early computational visualization:</strong> Levinthal used computer-controlled oscilloscopes and vector matrix multiplications to build and display 3D polypeptide structures, and showed a motion picture of forming a desired interaction via the most favored energy path. This was an early instance of computational molecular visualization.</li>
<li><strong>Local interactions and folding pathways:</strong> The hypothesis that &ldquo;local interactions&rdquo; serve as nucleation points that guide folding remains central to how modern structure prediction methods (e.g., AlphaFold) model residue-residue interactions.</li>
<li><strong>The paradox&rsquo;s lasting influence:</strong> The impossibility of random conformational search that Levinthal articulated continues to motivate approaches that exploit the structure of the energy landscape rather than exhaustive enumeration.</li>
<li><strong>Sequence-structure relationship:</strong> Levinthal&rsquo;s suggestion that not every random amino acid sequence would fold uniquely foreshadows the modern challenge of inverse folding (protein design), where the goal is to find sequences within the subset that does fold to a target structure.</li>
</ul>
<h2 id="paper-information">Paper Information</h2>
<p><strong>Citation</strong>: Levinthal, C. (1969). How to Fold Graciously. In <em>Mössbauer Spectroscopy in Biological Systems: Proceedings of a meeting held at Allerton House, Monticello, Illinois</em> (pp. 22-24). University of Illinois Press.</p>
<p><strong>Publication</strong>: Mössbauer Spectroscopy in Biological Systems Proceedings, 1969</p>
<div class="highlight"><pre tabindex="0" style="color:#f8f8f2;background-color:#272822;-moz-tab-size:4;-o-tab-size:4;tab-size:4;"><code class="language-bibtex" data-lang="bibtex"><span style="display:flex;"><span><span style="color:#a6e22e">@inproceedings</span>{levinthal1969fold,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">title</span>=<span style="color:#e6db74">{How to fold graciously}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">author</span>=<span style="color:#e6db74">{Levinthal, Cyrus}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">booktitle</span>=<span style="color:#e6db74">{M{\&#34;o}ssbauer spectroscopy in biological systems}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">pages</span>=<span style="color:#e6db74">{22--24}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">year</span>=<span style="color:#e6db74">{1969}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">publisher</span>=<span style="color:#e6db74">{University of Illinois Press}</span>,
</span></span><span style="display:flex;"><span>  <span style="color:#a6e22e">url</span>=<span style="color:#e6db74">{https://faculty.cc.gatech.edu/~turk/bio_sim/articles/proteins_levinthal_1969.pdf}</span>
</span></span><span style="display:flex;"><span>}
</span></span></code></pre></div><p><strong>Additional Resources</strong>:</p>
<ul>
<li><a href="https://en.wikipedia.org/wiki/Levinthal%27s_paradox">Levinthal&rsquo;s Paradox (Wikipedia)</a></li>
</ul>
]]></content:encoded></item></channel></rss>